No open cycle

This cause does not have an open cycle right now. Your grant status is still available in the dashboard.

National AI safety rules
Lobbying

Agency Rulemaking and Enforcement

Empower federal agencies to regulate AI under existing laws (or new authority) and aggressively enforce against AI harms.

Push federal agencies to set and enforce clear expectations for AI harms using existing authority (and any new authority Congress grants). This strategy focuses on targeted rules, guidance, investigations, and enforcement that require fixes in practice, not just promises. It’s especially valuable because agencies can act while Congress debates a broader framework.

Why this works

  • Agencies like the FTC, FDA, EEOC, etc.
  • can use their domain expertise to issue targeted rules – for instance, treating biased AI decisions as an “unfair or deceptive practice” under FTC Act authority.
  • This path can be faster than waiting for Congress; indeed, agencies are already investigating AI misuse (e.g.
  • the FTC’s investigation into OpenAI).
  • It also allows flexibility: agencies can update guidelines as AI evolves.

Public Citizen

Advocacy
citizen.org

Championing consumer rights and accountable government

Public Citizen is a nonprofit consumer advocacy organization founded in 1971 by Ralph Nader. It works to ensure that all citizens are represented in the halls of power by promoting public health and safety, government transparency, and corporate accountability. Public Citizen fights for campaign finance reform and ethics (it helped create the Office of Congressional Ethics:), advocates for safe pharmaceuticals and medical devices, pushes for strong consumer protections in trade deals, and litigates to enforce health, safety, and environmental laws.

How Public Citizen uses funding

  1. Identify high-impact AI uses where errors, bias, or privacy harms are most likely to affect people.
  2. Develop grounded legal theories under existing statutes for targeted agency action.
  3. Engage agencies with petitions, briefings, and evidence that supports guidance, rules, and enforcement priorities.
  4. Support investigations and enforcement that treat harmful AI practices as actionable, not just reputational issues.
  5. Coordinate across agencies so oversight is consistent and gaps are reduced.
  6. Update guidance as AI evolves so enforcement stays relevant over time.

Milestones

Checkpoints and the expected timing for each step

  1. 1

    Agency priority map defined

    Early

    Priority harms, high-impact contexts, and responsible agencies are mapped with a clear engagement plan.

  2. 2

    Petitions and briefings delivered

    As engagement begins

    Agencies receive a concrete package of evidence, recommendations, and draft guidance concepts.

  3. 3

    Rulemaking or guidance launched

    During agency action

    One or more agencies initiate formal guidance or rulemaking tied to AI harms.

  4. 4

    Enforcement outcomes established

    As cases proceed

    Early investigations or actions result in remedies that demonstrate real accountability.

  5. 5

    Iteration loop in place

    Ongoing

    Guidance is updated as needed and enforcement remains active as AI use expands.

Risks, trade-offs & sources

Updates

No updates yet.

Updates will appear here as the strategy progresses.

0votes left
Using bonus
0