A new coalition including academics, faith leaders, and public figures has launched a Pro‑Human AI Declaration, advocating for stricter safety standards and accountability as public concern grows over rapid AI development and governance.
An eclectic group of academics, business figures, faith leaders and politicians has publicly endorsed a new Pro‑Human AI Declaration that urges tougher safety measures and greater accountability for companies developing advanced artificial intelligence. According to the announcement from the Future of Life Institute, signatories include Richard Branson, economist Daron Acemoglu and former Trump adviser Steve Bannon, while backers range from the American Federation of Teachers to the Congress of Christian Leaders and the Progressive Democrats of America.
The declaration opens with a stark assertion: “Artificial intelligence should serve humanity, not the reverse.” It frames a vision in which “trustworthy and controllable AI tools amplify rather than diminish human potential, empower people, enhance human dignity, protect individual liberty, strengthen families and communities, preserve self‑governance and hep create unprecedented health and prosperity.” The document sets out core principles including human oversight of AI, measures to prevent dominant AI monopolies, protections for children, preservation of individual agency and legal accountability for unsafe systems.
Organisers deliberately excluded representatives from the tech industry when assembling the signatories, a departure from some earlier safety petitions that involved company figures. The move appears intended to present civil society, labour and moral authorities as the primary voice pressing for limits on corporate discretion in AI deployment. A parallel poll released with the declaration found strong public support for the approach, with about four in five US voters saying humans should remain in charge of AI and that companies should face greater accountability.
This latest campaign sits alongside a string of philanthropic and advocacy efforts aimed at counterbalancing commercially driven AI development. A coalition of foundations led by MacArthur and Omidyar this year announced a multi‑hundred‑million‑dollar Humanity AI initiative to fund projects that steer AI toward democratic, educational and civic goals, while grassroots campaigns such as Protect What’s Human have pressed for commonsense regulatory safeguards, transparency and independent oversight. Together these initiatives reflect a growing ecosystem of actors seeking to shape AI policy outside the industry’s boardrooms.
The Pro‑Human declaration follows earlier interventions from the Future of Life Institute, which in 2023 coordinated a widely publicised call for a six‑month pause on the training of systems more capable than GPT‑4 and later mobilised signatories around a ban on superintelligent AI until safety can be demonstrated. Those prior campaigns drew thousands of signatures from researchers, business leaders and public figures, but did not deter rapid corporate development; some past signatories subsequently launched AI startups of their own. The new declaration therefore appears to adopt a different tactic by emphasising a broad alliance of non‑industry institutions and public opinion.
Polling and survey data released by advocacy groups underscore widespread public unease with the pace and governance of advanced AI. A national survey of U.S. adults conducted for the Future of Life Institute found that roughly three‑quarters favour strong regulatory oversight, comparable to pharmaceutical standards, and that a substantial majority would delay development of superhuman AI until it can be shown safe and controllable. Advocates point to this gap between corporate momentum and public appetite as justification for urgent legislative and regulatory remedies.
Despite unanimity among many civil society actors about the need for stronger rules, observers note tensions within the broader coalition about means and ends. Some signatories press for outright moratoria on certain classes of research, while others favour phased regulation, public funding for alternatives that prioritise social goods, or legal liability frameworks to hold companies to account. According to reporting on earlier open letters, the movement also encompasses diverse political perspectives, making the question of durable policy consensus one of the movement’s immediate challenges.
Source Reference Map
Inspired by headline at: [1]
Sources by paragraph:
Source: Noah Wire Services
Noah Fact Check Pro
The draft above was created using the information available at the time the story first
emerged. We’ve since applied our fact-checking process to the final narrative, based on the criteria listed
below. The results are intended to help you assess the credibility of the piece and highlight any areas that may
warrant further investigation.
Freshness check
Score:
8
Notes:
The article was published on March 5, 2026, and reports on a recent event, indicating high freshness. However, similar narratives have appeared in other reputable outlets, such as The Independent, published on the same day, which may suggest some overlap in reporting. ([independent.co.uk](https://www.independent.co.uk/tech/ai-safety-declaration-steve-bannon-b2932570.html?utm_source=openai))
Quotes check
Score:
7
Notes:
The article includes direct quotes from the Pro-Human AI Declaration. While these quotes are consistent with the declaration’s publicly available text, they cannot be independently verified as originating from the specific signatories mentioned. ([humanstatement.org](https://humanstatement.org/?utm_source=openai))
Source reliability
Score:
9
Notes:
The article is published on Yahoo Finance, a major news organisation, which generally indicates high reliability. However, the presence of similar reports in other reputable outlets suggests that the content may be summarised or aggregated from a common source, potentially affecting its originality. ([independent.co.uk](https://www.independent.co.uk/tech/ai-safety-declaration-steve-bannon-b2932570.html?utm_source=openai))
Plausibility check
Score:
8
Notes:
The claims about the Pro-Human AI Declaration and its signatories are plausible and align with information from other reputable sources. However, the article’s reliance on a single source for direct quotes raises concerns about the independence of the information presented. ([humanstatement.org](https://humanstatement.org/?utm_source=openai))
Overall assessment
Verdict (FAIL, OPEN, PASS): PASS
Confidence (LOW, MEDIUM, HIGH): MEDIUM
Summary:
The article provides a timely and plausible account of the Pro-Human AI Declaration and its signatories. However, the reliance on a single source for direct quotes and the presence of similar reports in other reputable outlets suggest potential issues with originality and source independence. These factors warrant a medium level of confidence in the article’s accuracy and reliability.

