-  This bill is not active in this session.
 
     
  •  Summary 
  •  
  •  Actions 
  •  
  •  Committee Votes 
  •  
  •  Floor Votes 
  •  
  •  Memo 
  •  
  •  Text 
  •  
  •  LFIN 
  •  
  •  Chamber Video/Transcript 

A06453 Summary:

BILL NOA06453B
 
SAME ASSAME AS S06953-B
 
SPONSORBores
 
COSPNSRLasher, Seawright, Paulin, Tapia, Raga, Shimsky, Reyes, Epstein, Burke, Hevesi, Carroll P, Zaccaro, Hyndman, Lupardo, Kassay, Lee, Davila, Schiavoni, Lunsford, Brown K, Tannousis, Torres, Hooks, Gibbs, Romero, Colton, Conrad, Meeks, Glick, Cruz, Cunningham, Forrest, Chandler-Waterman, Stirpe, Wright, Simon, Dais, Jensen, Rozic, Gonzalez-Rojas, Solages, Gallagher, Otis, Kelles, Weprin, Wieder, Griffin
 
MLTSPNSR
 
Add Art 44-B §§1420 - 1425, Gen Bus L
 
Relates to the training and use of artificial intelligence frontier models; defines terms; establishes remedies for violations.
Go to top    

A06453 Actions:

BILL NOA06453B
 
03/05/2025referred to science and technology
05/16/2025amend and recommit to science and technology
05/16/2025print number 6453a
05/29/2025reported referred to codes
06/09/2025amend and recommit to codes
06/09/2025print number 6453b
06/11/2025reference changed to ways and means
06/12/2025reported referred to rules
06/12/2025reported
06/12/2025rules report cal.656
06/12/2025ordered to third reading rules cal.656
06/12/2025substituted by s6953b
 S06953 AMEND=B GOUNARDES
 03/27/2025REFERRED TO INTERNET AND TECHNOLOGY
 06/03/2025AMEND AND RECOMMIT TO INTERNET AND TECHNOLOGY
 06/03/2025PRINT NUMBER 6953A
 06/09/2025AMEND AND RECOMMIT TO INTERNET AND TECHNOLOGY
 06/09/2025PRINT NUMBER 6953B
 06/12/2025COMMITTEE DISCHARGED AND COMMITTED TO RULES
 06/12/2025ORDERED TO THIRD READING CAL.1889
 06/12/2025PASSED SENATE
 06/12/2025DELIVERED TO ASSEMBLY
 06/12/2025referred to ways and means
 06/12/2025substituted for a6453b
 06/12/2025ordered to third reading rules cal.656
 06/12/2025passed assembly
 06/12/2025returned to senate
 12/09/2025DELIVERED TO GOVERNOR
 12/19/2025SIGNED CHAP.699
 12/19/2025APPROVAL MEMO.76
 03/27/2025REFERRED TO INTERNET AND TECHNOLOGY
 06/03/2025AMEND AND RECOMMIT TO INTERNET AND TECHNOLOGY
 06/03/2025PRINT NUMBER 6953A
 06/09/2025AMEND AND RECOMMIT TO INTERNET AND TECHNOLOGY
 06/09/2025PRINT NUMBER 6953B
 06/12/2025COMMITTEE DISCHARGED AND COMMITTED TO RULES
 06/12/2025ORDERED TO THIRD READING CAL.1889
 06/12/2025PASSED SENATE
 06/12/2025DELIVERED TO ASSEMBLY
 06/12/2025referred to ways and means
 06/12/2025substituted for a6453b
 06/12/2025ordered to third reading rules cal.656
 06/12/2025passed assembly
 06/12/2025returned to senate
 12/09/2025DELIVERED TO GOVERNOR
 12/19/2025SIGNED CHAP.699
 12/19/2025APPROVAL MEMO.76
Go to top

A06453 Memo:

NEW YORK STATE ASSEMBLY
MEMORANDUM IN SUPPORT OF LEGISLATION
submitted in accordance with Assembly Rule III, Sec 1(f)
 
BILL NUMBER: A6453BREVISED 6/12/25
 
SPONSOR: Bores
  TITLE OF BILL: An act to amend the general business law, in relation to the training and use of artificial intelligence frontier models   PURPOSE OR GENERAL IDEA OF BILL: To allow for safety reports for powerful frontier artificial intelli- gence models to limit critical harm.   SUMMARY OF PROVISIONS: Section 1421 details transparency requirements regarding frontier model training and deployment, including requiring safety plans, disclosing safety incidents, and reporting safety incidents to the attorney general and division of homeland security and emergency services. Section 1422 defines penalties. Section 1423 clarifies the duties and obligations. Section 1424 requires that this article shall only apply to frontier models developed, deployed, or operating in New York state. Section 1425 establishes severability.   JUSTIFICATION: Artificial intelligence is evolving faster than any technology in human history. It is driving groundbreaking scientific advances leading to life-changing medicines, unlocking new creative potential, and automat- ing mundane tasks. At the same time, experts and practitioners in the field readily acknowledge the potential for serious risks. Al companies, leading scientists, and international bodies are preparing for a world in which Al can be used to conduct devastating cyberattacks, aid in the production of bioweapons, and even circumvent controls imposed by devel- opers. In March 2023, more than 1,000 tech leaders signed a letter calling for a "pause for at least 6 months" on training frontier models until inter- national safety standards could be established. The signatories from figures like Rachel Bronson, Steve Wozniak, Andrew Yang, and Elon Musk span the political spectrum and underscore the urgent need for caution (1). Since then, Ai models have gotten exponentially more powerful. We are only a few years away from when Al models will code themselves; already, 25% of the new code from Google's parent company Alphabet is written by Ai (2). In December 2024, Apollo Research tested large Ai models by making them believe they would be shut down; models from every lab test- ed (OpenAi, Anthropic, Google, and Meta) tried to make copies of them- selves on new servers and then lied to humans about their actions (3). Current models are already showing the potential to aid nefarious actors in inflicting harms. China's government has employed Meta's Ai model for both broad military uses (4) and citizen surveillance (5). The Interna- tional Al Safety Report, written by over 100 experts from 30 countries (including the US) and led by the "godfather of Ai" Yoshua Bengio, iden- tified several emerging risks including that an existing model produced plans for biological weapons "rated superior to plans generated by experts with a PhD 72% of the time and provides details that expert evaluators could not find online (6)." Developers of this technology continue to sound the alarm. In reviewing the safety of its latest model, OpenAi stated, "Several of our biology evaluations indicate our models are on the cusp of being able to mean- ingfully help novices create known biological threats, which would cross our high-risk threshold. We expect current trends of rapidly increasing capability to continue, and for models to cross this threshold in the near future (7)." Another leading lab, Anthropic, warned that "the window for proactive risk prevention is closing fast" and that governments must put in place regulation of frontier models by April 2026 at the latest (8). Given New York's legislative calendar, that requires urgent action in our 2025 session. Anthropic also said that while they would prefer action at the federal level, they admitted that "the federal legislative process will not be fast enough to address risks on the timescale about which we're concerned" and "urgency may demand it is instead developed by individual states (9)." Our laws have not kept up. We do not let people do things as mundane as open a daycare center without a safety plan. This bill simply says that companies spending hundreds of millions of dollars to train the most advanced Al models need to take the following common-sense steps: 1. Have a safety plan to prevent severe risks (as most of them already do); 2. Disclose major security incidents, so that no one has to make the same mistake twice; 3. Not release a model that they know causes an unreasonable risk of a catastrophic harm The risks noted above are more than sufficient to justify the measures taken in this bill, but we should be mindful that experts have repeated- ly warned of even more severe threats. In 2023, more than a thousand experts, including the CEOs of Google DeepMind, Anthropic, and OpenAi and many world leading academics signed a letter stating that "mitigat- ing the risk of extinction from Ai should be a global priority alongside other societal-scale risks such as pandemics and nuclear war (10)." In the face of these dangers, we must keep a clear focus on the immense promises of Al. Regulation needs to be targeted and surgical to reduce risks while promoting Ai's benefits. The RAISE Act is designed to do exactly this. Limited to only a small set of very severe risks, the RAISE act will not require many changes from what the vast majority of Ai companies are currently doing; instead, it will simply ensure that no company has an economic incentive to cut corners or abandon its safety plan. Notably, this bill does not attempt to tackle emery issue with Ai: important questions about bias, authenticity, workforce impacts, and other concerns can and should be handled with additional legislation. The RAISE Act focuses on severe risks that could cause over $1 billion in damage or hundreds of deaths or injuries. For these kinds of risks, this bill is the bare minimum that New Yorkers expect.   PRIOR LEGISLATIVE HISTORY: This is a new bill.   FISCAL IMPLICATIONS FOR STATE AND LOCAL GOVERNMENTS: None   EFFECTIVE DATE: This act shall take effect ninety days after it becomes law (1) https://futureoflife.org/open-letter /pause-giant-ai-experiments/ (2) https://fortune.com/2024/10/30 /googles-code-ai-sundar-pichai/ (3) https://www.apolloresearch.ai/research/scherning- reasoning-evalua- tions (4)https://www.reuters.com/technology/artificial- intelligence/chineseresearchersdevelop-ai-model- military-use-back-me- tas-Ilama-2024-11-01/ (5) https://www.nytimes.com/2025/02/21/technology /openai-chinese-sur- veillance.html (6) https://arxiv.org/pC1f/2501.17805 (7) https://cdn.openai.com /deep-research-system-card.pdf (8) https://www.anthropic.com/news/the- case-for-targeted-regulation (9) Ibid (10) https://www.safe.ai/work /statement-on-al-risk
Go to top

A06453 Text:



 
                STATE OF NEW YORK
        ________________________________________________________________________
 
                                         6453--B
 
                               2025-2026 Regular Sessions
 
                   IN ASSEMBLY
 
                                      March 5, 2025
                                       ___________
 
        Introduced  by  M.  of A. BORES, LASHER, SEAWRIGHT, PAULIN, TAPIA, RAGA,
          SHIMSKY, REYES, EPSTEIN, BURKE, HEVESI, P. CARROLL, ZACCARO,  HYNDMAN,
          LUPARDO,  KASSAY, LEE, DAVILA, SCHIAVONI, LUNSFORD, K. BROWN, TANNOUS-
          IS, TORRES, HOOKS, GIBBS, ROMERO, COLTON, CONRAD, MEEKS, GLICK,  CRUZ,
          CUNNINGHAM,  FORREST,  CHANDLER-WATERMAN, STIRPE, WRIGHT, SIMON, DAIS,
          JENSEN, ROZIC, GONZALEZ-ROJAS -- read once and referred to the Commit-
          tee on Science and Technology -- committee discharged,  bill  amended,
          ordered  reprinted  as  amended  and  recommitted to said committee --
          reported  and  referred  to  the  Committee  on  Codes  --   committee
          discharged, bill amended, ordered reprinted as amended and recommitted
          to said committee
 
        AN  ACT  to  amend the general business law, in relation to the training
          and use of artificial intelligence frontier models
 
          The People of the State of New York, represented in Senate and  Assem-
        bly, do enact as follows:
 
     1    Section  1.  Short  title. This act shall be known and may be cited as
     2  the "Responsible AI safety and education act" or "RAISE act".
     3    § 2. The general business law is amended by adding a new article  44-B
     4  to read as follows:
     5                                 ARTICLE 44-B
     6               RESPONSIBLE AI SAFETY AND EDUCATION (RAISE) ACT
     7  Section 1420. Definitions.
     8          1421. Transparency  requirements regarding frontier model train-
     9                  ing and use.
    10          1422. Violations.
    11          1423. Duties and obligations.
    12          1424. Scope.
    13          1425. Severability.
    14    § 1420. Definitions. As used in  this  article,  the  following  terms
    15  shall have the following meanings:
 
         EXPLANATION--Matter in italics (underscored) is new; matter in brackets
                              [ ] is old law to be omitted.
                                                                   LBD00047-14-5

        A. 6453--B                          2
 
     1    1.  "Appropriate redactions" means redactions to a safety and security
     2  protocol that a developer may make when necessary to:
     3    (a)  protect  public safety to the extent the developer can reasonably
     4  predict such risks;
     5    (b) protect trade secrets;
     6    (c) prevent the release of confidential  information  as  required  by
     7  state or federal law;
     8    (d) protect employee or customer privacy; or
     9    (e)  prevent  the release of information otherwise controlled by state
    10  or federal law.
    11    2. "Artificial intelligence" means a machine-based  system  that  can,
    12  for a given set of human-defined objectives, make predictions, recommen-
    13  dations, or decisions influencing real or virtual environments, and that
    14  uses  machine- and human-based inputs to perceive real and virtual envi-
    15  ronments, abstract such perceptions into models through analysis  in  an
    16  automated  manner,  and  use  model  inference  to formulate options for
    17  information or action.
    18    3. "Artificial intelligence model"  means  an  information  system  or
    19  component  of  an information system that implements artificial intelli-
    20  gence technology and uses computational, statistical, or  machine-learn-
    21  ing techniques to produce outputs from a given set of inputs.
    22    4.  "Compute  cost" means the cost incurred to pay for compute used in
    23  the final training run of a model  when  calculated  using  the  average
    24  published  market  prices  of  cloud compute in the United States at the
    25  start of training such model as reasonably assessed by the person  doing
    26  the training.
    27    5.  "Deploy" means to use a frontier model or to make a frontier model
    28  foreseeably available to one or more third parties  for  use,  modifica-
    29  tion,  copying, or a combination thereof with other software, except for
    30  training or developing the frontier model, evaluating the frontier model
    31  or other frontier models, or complying with federal or state laws.
    32    6. "Frontier model" means either of the following:
    33    (a) an artificial intelligence model trained using greater than  10§26
    34  computational  operations  (e.g., integer or floating-point operations),
    35  the compute cost of which exceeds one hundred million dollars; or
    36    (b) an artificial intelligence model produced  by  applying  knowledge
    37  distillation  to  a  frontier  model as defined in paragraph (a) of this
    38  subdivision, provided that the compute cost for such model  produced  by
    39  applying knowledge distillation exceeds five million dollars.
    40    7. "Critical harm" means the death or serious injury of one hundred or
    41  more  people  or  at  least  one billion dollars of damages to rights in
    42  money or property caused or materially enabled by  a  large  developer's
    43  use,  storage,  or  release  of  a frontier model, through either of the
    44  following:
    45    (a) The creation or use of a chemical,  biological,  radiological,  or
    46  nuclear weapon; or
    47    (b)  An  artificial  intelligence  model engaging in conduct that does
    48  both of the following:
    49    (i) Acts with no meaningful human intervention; and
    50    (ii) Would, if committed by a human, constitute a crime  specified  in
    51  the  penal  law that requires intent, recklessness, or gross negligence,
    52  or the solicitation or aiding and abetting of such a crime.
    53    A harm inflicted by an intervening human actor shall not be deemed  to
    54  result  from  a  developer's  activities  unless  such activities were a
    55  substantial factor in bringing about the  harm,  the  intervening  human
    56  actor's  conduct was reasonably foreseeable as a probable consequence of

        A. 6453--B                          3

     1  the developer's activities, and could have been reasonably prevented  or
     2  mitigated  through  alternative  design, or security measures, or safety
     3  protocols.
     4    8.  "Knowledge  distillation"  means any supervised learning technique
     5  that uses a larger artificial intelligence model  or  the  output  of  a
     6  larger  artificial  intelligence  model  to  train  a smaller artificial
     7  intelligence model with similar or equivalent capabilities as the larger
     8  artificial intelligence model.
     9    9. "Large developer" means a person that  has  trained  at  least  one
    10  frontier model and has spent over one hundred million dollars in compute
    11  costs in aggregate in training frontier models.  Accredited colleges and
    12  universities shall not be considered large developers under this article
    13  to  the  extent  that  such  colleges  and  universities are engaging in
    14  academic research.  If a person subsequently transfers full intellectual
    15  property rights of the frontier model to another person  (including  the
    16  right  to  resell  the model) and retains none of those rights for them-
    17  self, then the receiving person shall be considered the large  developer
    18  and  shall  be  subject to the responsibilities and requirements of this
    19  article after such transfer.
    20    10. "Model weight" means a numerical parameter in an artificial intel-
    21  ligence model that is adjusted through training and that helps determine
    22  how inputs are transformed into outputs.
    23    11. "Person" means an individual, proprietorship,  firm,  partnership,
    24  joint  venture, syndicate, business trust, company, corporation, limited
    25  liability company, association, committee, or any other  nongovernmental
    26  organization or group of persons acting in concert.
    27    12.  "Safety  and  security  protocol"  means documented technical and
    28  organizational protocols that:
    29    (a) Describe reasonable protections and procedures that,  if  success-
    30  fully implemented would appropriately reduce the risk of critical harm;
    31    (b)   Describe  reasonable  administrative,  technical,  and  physical
    32  cybersecurity protections for frontier models within the large  develop-
    33  er's control that, if successfully implemented, appropriately reduce the
    34  risk  of unauthorized access to, or misuse of, the frontier models lead-
    35  ing to critical harm, including by sophisticated actors;
    36    (c) Describe in detail the testing procedure to evaluate if the  fron-
    37  tier  model  poses an unreasonable risk of critical harm and whether the
    38  frontier model could be misused, be modified, be executed with increased
    39  computational resources, evade the control of  its  large  developer  or
    40  user, be combined with other software or be used to create another fron-
    41  tier model in a manner that would increase the risk of critical harm;
    42    (d)  Enable  the  large  developer  or  third party to comply with the
    43  requirements of this article; and
    44    (e) Designate senior personnel to be responsible for ensuring  compli-
    45  ance.
    46    13.  "Safety  incident" means a known incidence of critical harm or an
    47  incident of the following kinds that  occurs  in  such  a  way  that  it
    48  provides demonstrable evidence of an increased risk of critical harm:
    49    (a)  A  frontier model autonomously engaging in behavior other than at
    50  the request of a user;
    51    (b) Theft, misappropriation, malicious use, inadvertent release, unau-
    52  thorized access, or escape of the model weights of a frontier model;
    53    (c) The critical failure of any technical or administrative  controls,
    54  including controls limiting the ability to modify a frontier model; or
    55    (d) Unauthorized use of a frontier model.

        A. 6453--B                          4

     1    14.  "Trade  secret"  means  any form and type of financial, business,
     2  scientific, technical, economic, or engineering information, including a
     3  pattern, plan, compilation, program device, formula, design,  prototype,
     4  method,  technique, process, procedure, program, or code, whether tangi-
     5  ble  or intangible, and whether or how stored, compiled, or memorialized
     6  physically, electronically, graphically, photographically or in writing,
     7  that:
     8    (a) Derives independent economic value, actual or potential, from  not
     9  being  generally known to, and not being readily ascertainable by proper
    10  means by, other persons who can obtain economic value from  its  disclo-
    11  sure or use; and
    12    (b)  Is  the  subject of efforts that are reasonable under the circum-
    13  stances to maintain its secrecy.
    14    § 1421. Transparency requirements regarding  frontier  model  training
    15  and  use.  1.  Before deploying a frontier model, the large developer of
    16  such frontier model shall do all of the following:
    17    (a) Implement a written safety and security protocol;
    18    (b) Retain an unredacted copy of the  safety  and  security  protocol,
    19  including records and dates of any updates or revisions. Such unredacted
    20  copy of the safety and security protocol, including records and dates of
    21  any  updates  or  revisions, shall be retained for as long as a frontier
    22  model is deployed plus five years;
    23    (c) (i) Conspicuously publish a copy of the safety and security proto-
    24  col with appropriate redactions and transmit a  copy  of  such  redacted
    25  safety  and  security  protocol  to the attorney general and division of
    26  homeland security and emergency services;
    27    (ii) Grant the attorney general and division of homeland security  and
    28  emergency  services  or  the  attorney  general access to the safety and
    29  security protocol, with redactions only to the extent required by feder-
    30  al law, upon request;
    31    (d) Record, as and when reasonably possible, and retain for as long as
    32  the frontier model is  deployed  plus  five  years  information  on  the
    33  specific  tests  and test results used in any assessment of the frontier
    34  model required by this section or the developer's  safety  and  security
    35  protocol  that provides sufficient detail for third parties to replicate
    36  the testing procedure; and
    37    (e) Implement appropriate safeguards to prevent unreasonable  risk  of
    38  critical harm.
    39    2.  A  large  developer  shall not deploy a frontier model if doing so
    40  would create an unreasonable risk of critical harm.
    41    3. A large developer shall conduct an annual review of any safety  and
    42  security    protocol required by this section to account for any changes
    43  to the capabilities of their frontier models and industry best practices
    44  and, if necessary, make modifications to such safety and security proto-
    45  col.  If any material modifications are made, the large developer  shall
    46  publish  the safety and security protocol in the same manner as required
    47  pursuant to paragraph (c) of subdivision one of this section.
    48    4. A large developer shall disclose each safety incident affecting the
    49  frontier model to the attorney general and division of homeland security
    50  and emergency services within seventy-two hours of the  large  developer
    51  learning of the safety incident or within seventy-two hours of the large
    52  developer  learning  facts  sufficient  to establish a reasonable belief
    53  that a safety incident has occurred.  Such disclosure shall include: (a)
    54  the date of the safety incident; (b) the reasons the incident  qualifies
    55  as a safety incident as defined in subdivision thirteen of section four-

        A. 6453--B                          5
 
     1  teen hundred twenty of this article; and (c) a short and plain statement
     2  describing the safety incident.
     3    5.  A  large  developer  shall  not knowingly make false or materially
     4  misleading statements or omissions in or  regarding  documents  produced
     5  pursuant to this section.
     6    §  1422.  Violations. 1. The attorney general may bring a civil action
     7  for a violation of this article and to recover  all  of  the  following,
     8  determined based on severity of the violation:
     9    (a)  For  a violation   of section fourteen hundred twenty-one of this
    10  article, a civil penalty in an amount not exceeding ten million  dollars
    11  for  a  first  violation  and  in an amount not exceeding thirty million
    12  dollars for any subsequent violation.
    13    (b) For a violation of section fourteen  hundred  twenty-one  of  this
    14  article, injunctive or declaratory relief.
    15    2.  Nothing  in this article shall be construed to establish a private
    16  right of action associated with violations of this article.
    17    3. Nothing in this subdivision shall be construed to prevent  a  large
    18  developer  from  asserting that another person, entity, or factor may be
    19  responsible for any alleged harm, injury, or  damage  resulting  from  a
    20  critical harm or a violation of this article.
    21    4. This section does not limit the application of other laws.
    22    §  1423. Duties and obligations. The duties and obligations imposed by
    23  this article are cumulative with any other duties or obligations imposed
    24  under other law and shall not be construed to relieve any party from any
    25  duties or obligations imposed under other  law  and  do  not  limit  any
    26  rights or remedies under existing law.
    27    §  1424.  Scope. This article shall only apply to frontier models that
    28  are developed, deployed, or operating in whole or in part  in  New  York
    29  state.
    30    § 1425. Severability. If any clause, sentence, paragraph, subdivision,
    31  section or part of this article shall be adjudged by any court of compe-
    32  tent jurisdiction to be invalid, such judgment shall not affect, impair,
    33  or invalidate the remainder thereof, but shall be confined in its opera-
    34  tion  to  the clause, sentence, paragraph, subdivision, section, or part
    35  thereof directly involved in the  controversy  in  which  such  judgment
    36  shall have been made.
    37    §  3.  This  act shall take effect on the ninetieth day after it shall
    38  have become a law.
Go to top