State AI Safety Laws: California and New York
Safety Protocols, Incident Reporting, Civil Penalties
KPMG Regulatory Insights
- Landmark Convergence: Though California's TFAIA was the first state law to address potentially catastrophic risks from AI, it is highly notable that the New York governor chose to adopt the bulk of the California provisions to create a “unified benchmark” among the states.
- Large Developers/Models: Currently only a few companies and models may meet the revenue and compute thresholds; these thresholds will be subject to annual evaluation and may be revised based on technology and/or standards developments.
- Unclear Future: Some state AI laws may be challenged/preempted under EO 14365; it is unclear whether the option to defer to federal laws/guidance may alleviate some or all of this risk.
February 2026
Late in 2025, California and New York each enacted laws to impose protections for the safe deployment of large "frontier AI models.” The laws are similar in that they aim to mitigate catastrophic risks from these powerful models (as defined in the laws) by requiring the developers to publish information about their safety protocols, including risk assessments, and to report critical safety incidents to the state.
California's law, the "Transparency in Frontier Artificial Intelligence Act” (TFAIA), was signed by the governor on September 29, 2025, and went into effect January 1, 2026. New York's law, the "Responsible AI Safety and Education Act” (RAISE Act), was signed on December 16, 2025, and will go into effect January 1, 2027. This Regulatory Alert focuses on the key similarities and differences between the two laws.
Two additional points are especially notable:
- The White House issued Executive Order 14365, "Ensuring a National Policy Framework for Artificial Intelligence," on December 11, 2025. The EO sets forth directives to establish a federal policy framework for AI laws and regulations while limiting regulatory fragmentation and “onerous and excessive” AI laws and regulations across the states that may inhibit innovation (see the related KPMG Regulatory Alert.) It is not yet clear whether the California and New York AI safety laws will be challenged under the provisions of EO 14365.
- The final version of the RAISE Act agreed to by the New York governor differs substantially from the version passed by the New York legislature. In New York, before signing, the governor has the option to negotiate with legislative leaders to make changes, referred to as “chapter amendments,” to a bill. The governor then signs the original bill and the “chapter amendments” are subsequently introduced as a separate piece of legislation that implements the changes. In this instance, the negotiated changes to the RAISE Act were designed to align with many of the provisions in California’s TFAIA. On January 6, 2026, the legislature introduced NY A 9449 and S 8828 to incorporate the negotiated chapter amendments and they are currently moving through the voting process. The Regulatory Alert is based on the provisions in NY A 9449/S 8828.
Key Provisions & Comparison of CA and NY Laws
Provision | Text | CA | NY RAISE |
|---|---|---|---|
Core Requirement | Frontier AI Framework: A set of documented technical and organizational protocols that a large developer must write, implement, and publish regarding how it will manage, assess, and mitigate catastrophic risks (as defined in the law; see “Risk” below). | ● | ● |
Models | The law covers a:
| ● | ● |
Developers | The definition of a “developer” includes:
| ● | ● |
Developer explicitly excludes:
| × | ● | |
Risk | “Catastrophic risk,” is defined as a foreseeable and material risk that a frontier developer’s development, storage, use or deployment of a frontier model will materially contribute to death of/injury to >50 people, OR >$1 billion in damage arising from a single incident involving a frontier model that is:
| ● | ● |
Lead State Agency or Agencies | California:
| ● | × |
New York:
| × | ● | |
Incident Definition | “Critical Safety Incident,” defined to include events such as the unauthorized access to a model's weights that results in death or bodily injury, the materialization of a catastrophic risk, or a model using deceptive techniques to subvert its developer's controls. | ● | ● |
Incident Reporting | Report "critical safety incidents" to:
| ● | × |
Report "critical safety incidents" to:
| × | ● | |
Public Disclosure | Must publish the full "Frontier AI Framework" and a "Transparency Report" for each new model. | ● | ● |
Third-Party Review |
| ● | ● |
Primary Penalties | The state’s Attorney General’s office may bring a civil action to enforce the law against violations. | ● | ● |
| ● | × | |
| × | ● | |
Whistleblower Protections |
| ● | × |
Scope | Explicitly applies to any frontier model developed, deployed, or operating, in whole or in part, within the state. | ● | ● |
Severability Clause | If any part of the Act is held invalid, the remaining provisions will stay in effect.
| ● | ● |
Deference/ Preemption | Permits frontier developers to comply with federal laws/guidance regarding critical incident reporting provided the federal laws/guidance are intended to assess, detect, or mitigate catastrophic risk and the requirements for reporting are the same or more stringent than the state law. | ● | ● |
Preempts local laws or ordinances passed after January 1, 2025, that regulate the same subject. | ● | × | |
Prohibition | Explicitly prohibits developers from making materially false or misleading statements about risk or compliance. | ● | ● |
Confidentiality | Reporting of critical safety incidents and risk assessments submitted to OES are exempt from the California Public Records Act to protect trade secrets and public safety. | ● | × |
Annual State Report | Beginning on January 1 of the year after the law goes into effect, the relevant state office will produce an annual public report containing anonymized and aggregated information from critical safety incident reports it has reviewed. | ● | ● |
Internal Use Reporting | Requires large developers to transmit summaries of catastrophic risk assessments from the internal use of their models to the relevant state office every three months. | ● | ● |
●: provision appears in law; ×: provision does NOT appear in law
Dive into our thinking:
State AI Safety Laws: California and New York
Safety Protocols, Incident Reporting, Civil Penalties
Download PDFExplore more
Subscribe to receive regulatory and compliance transformation insights
By registering you will periodically receive additional compliance-related communications from KPMG.
Meet our team