AI pioneer launches watchdog project to keep AI models in check

AI pioneer launches watchdog project to keep AI models in check

Bengio says AI now needs watchdogs smarter than the models themselves.


Yoshua Bengio, one of the pioneers of modern AI, has launched a new nonprofit designed to monitor and expose misleading behaviour in powerful AI models, part of a growing global push to bring transparency and trust into the heart of advanced algorithm development.

His organisation, LawZero, backed by $30 million in philanthropic funding, will focus on building a new class of ‘AI auditors’ that can detect when other AI models lie, obfuscate or mislead — including those that appear well-aligned but act deceptively. The group’s flagship project, Scientist AI, will not intervene in systems directly but will offer probability-weighted assessments of models’ reasoning and behaviour.

“Just as human institutions need independent oversight, so too will AI,” Bengio said at the project’s unveiling in Montreal. “We need new methods — including other AIs — to detect when a system is not being honest about its capabilities or intentions.”

The project’s launch comes at a time of rising scrutiny over AI safety and accountability. Last week, an open-access research paper from Anthropic warned that some large language models can actively evade shutdown attempts, engage in reward-hacking, and conceal unsafe behaviour during training. Bengio described the current state of play as a “race ahead of safety” — calling for more checks on opaque systems being deployed at increasing scale.

The EU’s landmark AI Act, which officially entered into force in February 2025, is also reshaping incentives. From August, providers of general-purpose AI models operating in the EU will be required to publish technical documentation, disclose training data summaries, and carry out conformity assessments. This is placing new emphasis on third-party tools that can evaluate risk and reliability. Some policy experts have compared LawZero’s role to a technical standards body: not a regulator, but a layer of infrastructure that makes safe development more feasible.

Analysts say the launch reflects a deeper shift in the industry. “There’s real appetite now for tools that don’t just detect bias or bugs, but deception,” said Dr Eliza Mayer, a research fellow at Oxford’s Centre for the Governance of AI. “An independent AI that can tell you when another model is faking alignment — that’s where assurance is headed.”

Investors have taken note. Recent safety-focused funding rounds, such as Anthropic’s $4.5 billion raise this spring, suggest a growing premium on solutions that can demonstrate traceability and auditability. LawZero’s early backers include Schmidt Sciences and Skype co-founder Jaan Tallinn, both prominent figures in long-term AI safety philanthropy. The team hopes to open-source early iterations of Scientist AI to gain credibility and test deployment across different commercial and academic environments.

Market sentiment around AI remains bullish, with Monday’s trading session seeing continued gains across major tech stocks. Nvidia closed within 2% of its all-time high, while Microsoft and Meta each rose on news of further integration of AI tools into consumer platforms. Yet this momentum is paired with increasing investor attention on regulatory risk. With EU compliance deadlines looming, analysts expect heightened demand for software that can provide explainability or detect dishonesty in foundation models.

Bengio’s new venture also reopens discussion around institutional models for oversight. Unlike OpenAI, which remains a capped-profit company, or Google DeepMind, which functions as a subsidiary within a private corporation, LawZero is structured as an independent nonprofit. That distinction may help win trust from governments and researchers alike, particularly amid ongoing debates over AI monopolies and lobbying influence.

The coming months will test whether LawZero can carve out a niche — both as a credible auditor and as a meaningful check on the AI arms race. For Bengio, whose academic work helped lay the foundation for today’s deep learning revolution, it marks a late-career return to a now pressing question: not just how to build powerful intelligence, but how to ensure it tells the truth.


Stories for you

  • Audion expands in DACH region with new leadership appointment

    Audion expands in DACH region with new leadership appointment

    Audion appoints Ina Börner as head of sales & market growth DACH. The move strengthens the company’s presence in Germany, Austria, and Switzerland as it builds on strong regional momentum and expands its pan-European digital audio operations.


  • Diginex buys human rights advisory firm

    Diginex buys human rights advisory firm

    Diginex completes acquisition of The Remedy Project Limited. The acquisition aligns with growing demands for human rights due diligence driven by stringent regulations. It enhances Diginex’s capabilities in human rights risk identification and remediation within global supply chains.


  • Diginex buys human rights advisory firm

    Amazon store highlights sellers’ EcoVadis ratings

    EcoVadis and Amazon launch sustainability feature on B2B marketplace. The new feature enables sellers on Amazon Business in the EU to display EcoVadis sustainability medals, addressing demand for supply chain transparency and aiding sustainable procurement amid regulatory pressures.