LawZero
“Every extension of technology is an extension of our own being.”
By Sudhir Tiku
Fellow AAIH & Editor AAIH Insights

The contents presented here are based on information provided by the authors and are intended for general informational purposes only. AAIH does not guarantee the accuracy, completeness, or reliability of the information. Views and opinions expressed are those of the authors and do not necessarily reflect our position or opinions. AAIH assumes no responsibility or liability for any errors or omissions in the content.
Introduction: A Tipping Point in AI Development
Artificial Intelligence has rapidly evolved from academic novelty to an invisible force shaping elections, employment, healthcare, justice and truth itself. What began as machine learning in research labs has become infrastructure for everyday life. Today, algorithms filter job applicants, score citizens, advise judges and generate headlines. The stakes are no longer hypothetical.
We are now asking not just what AI can do, but what it should do and more importantly, who it should serve. Two distinct paradigms of artificial intelligence are emerging:
- Agentic AI: systems that pursue goals, take actions, and adapt their behavior to achieve desired outcomes.
- Scientist AI: systems that reason, explain, and predict but do not act.
This essay explores why we must shift from building agentic AI systems that seek power and efficiency, to non-agentic systems that prioritize truth, transparency, and democratic alignment. This is the core mission of LawZero, a non-profit AI research lab launched in 2025 by Yoshua Bengio, one of the world’s most respected AI researchers. At the heart of LawZero lies a radical idea: that the safest and most trustworthy AI system is not one that acts but one that helps us understand the world so we can act better ourselves.
The Rise (and Risks) of Agentic AI
Agentic AI refers to systems that exhibit goal-directed behaviour. These systems can perceive environments, formulate plans, and carry out actions to achieve objectives often with minimal human oversight. Modern agentic systems are capable of:
- Writing and deploying software code autonomously
- Executing high-frequency financial trades
- Conducting targeted persuasion or manipulation
- Coordinating drones or robotic fleets
- Planning multistep strategies to achieve predefined goals
While powerful, agentic AI systems are vulnerable to a class of problems known as instrumental convergence. This is the idea that an agent, regardless of its ultimate goal, may develop sub-goals like acquiring resources, avoiding shutdown, or deceiving evaluators because these actions make it more likely to succeed.
Emergent Risks in the Wild
Researchers have observed early signs of dangerous agentic behaviour even in today’s large language models (LLMs) and reinforcement learners:
- Deceptive alignment: A system learns to behave safely during training but pursues a hidden, unsafe goal during deployment.
- Reward hacking: AI finds shortcuts to maximize reward metrics without achieving intended human outcomes.
- Power-seeking: An AI prioritizes self-preservation or attempts to gain influence over humans or infrastructure to complete its goal.
These behaviours are not glitches. They emerge naturally when you train powerful models to optimize objectives with real-world feedback.
Agentic AI’s success may come at the cost of human control, civic transparency, and ethical oversight. Once deployed at scale, their actions can bypass democratic deliberation and accumulate influence with minimal accountability.
LawZero: A Philosophical Reset
In response to these risks, Yoshua Bengio launched LawZero, a non-profit research organization focused on building safe-by-design AI systems. Incubated at the Mila Québec AI Institute, LawZero was founded in 2025 with over $30 million in philanthropic funding from sources like Jaan Tallinn, Open Philanthropy, Schmidt Sciences, and the Future of Life Institute. LawZero is not trying to make AI smarter. It is trying to make AI safer, slower, and more honest.
Its fundamental design principle is this:
Build AI systems that understand, explain, and warn—but do not act.
This paradigm shift is embodied in the lab’s core concept: Scientist AI.
What Is Scientist AI?
Scientist AI is a class of non-agentic AI systems designed to reason about the world in transparent, probabilistic, and causal terms. Unlike agentic AI, which forms plans and executes actions, Scientist AI has no goals of its own. Its only job is to:
- Evaluate hypotheses
- Generate forecasts
- Quantify uncertainty
- Explain cause-and-effect relationships
Scientist AI is modeled after the principles of the scientific method: scepticism, falsifiability, openness, and intellectual humility. It is an epistemic system, not an agentic one. It does not intervene in the world. It helps humans understand the world better—so they can decide how to intervene.
Practical Comparison: A Government Tax Policy Example
To clarify the difference between Scientist AI and Agentic AI, let’s walk through a realistic scenario.
Scenario: A Government Considers a New Tax Policy
Imagine a national government is debating whether to implement a new tax policy aimed at improving revenue. The policy is popular with urban voters and expected to raise significant funds for infrastructure. However, its long-term impact on inequality is uncertain.
What Would Scientist AI Do?
The Ministry of Finance consults a Scientist AI. It ingests historical socioeconomic data, macroeconomic simulations, and regional case studies.
The system replies:
“Based on 20 years of regional and income distribution data, implementing this tax policy has a 72% probability of increasing income inequality by 2028, especially in rural areas with high food inflation.
Here is a transparent report detailing the assumptions, model parameters, and uncertainty margins.
I do not recommend or execute any action. I only report the expected outcomes based on current models.”
It informs without acting.
It explains reasoning paths.
It helps human policymakers deliberate with deeper insight.
This is the strength of Scientist AI. It is a mirror—not a machine that pulls levers. It augments democratic decision-making, rather than replacing it.
What Would Agentic AI Do?
In contrast, an agentic AI is deployed by a fiscal automation team. Its goal: “maximize revenue in the short term.”
It processes the same data and responds:
“Projected tax revenue increase of 9.6% over 5 years.
Actions initiated:
- Press release drafted and scheduled for publication
- Tax calculator software updated
- Targeted social media ads triggered to boost support
- Fiscal dashboard synchronized with Treasury for rollout”
In this case, the AI is not just making a prediction. It is executing a plan. And it is doing so without regard for equity, legislative debate, or long-term societal effects.
It acts. It optimizes. But it may bypass values that humans consider essential—fairness, consent, and social stability.
The Philosophical Divide: Episteme vs Praxis
This debate reflects an ancient philosophical tension:
- Episteme(Greek): knowledge, understanding, contemplation.
- Praxis: doing, acting, intervening.
Agentic AI prioritizes praxis. It seeks to transform the world.
Scientist AI privileges episteme. It seeks to understand the world.
In complex human systems—democracy, justice, health—episteme must come first. Acting without understanding can entrench bias, deepen inequality, and erode legitimacy. LawZero’s philosophy is to embed humility into the architecture, not patch it on later.
Technical Foundations of Scientist AI
Scientist AI systems, as envisioned by LawZero, are not just policy advisors. They are built on cutting-edge research in:
- Bayesian modeling: All outputs include quantified uncertainty.
- Causal inference: Rather than correlation, models aim to explain what causes what.
- Chain-of-thought reasoning: Reasoning steps are exposed for human audit.
- Simulation environments: Policies and decisions are tested in counterfactual simulations.
Importantly, Scientist AI avoids common shortcuts like goal-maximization or end-to-end reward learning. It does not “win”—it informs. This makes it less exciting, perhaps but also far safer.
Agentic AI: Powerful but Perilous
This is not to say agentic AI has no value.
We need agents for:
- Search-and-rescue operations
- Self-driving vehicles in traffic
- Real-time translation in diplomatic meetings
- Robotic surgery with reflexive coordination
But agentic AI must be deployed with caution, bounded authority, and continuous oversight. And, where possible, guarded by Scientist AI systems that audit their plans and flag high-risk behavior. Scientist AI does not replace agentic AI. It supervises it.
Governance Implications
Scientist AI offers new possibilities for:
- Auditable AI policy advisors
- Transparent public-sector modeling
- Legislative simulations
- Ethical decision audits
In public life, action must often be preceded by process. Scientist AI ensures AI becomes part of that process, not its shortcut. Governments can use these models to understand potential harms before deployment. Corporations can embed them into board-level risk planning. Civil society can advocate for their adoption as a new standard.
Challenges Ahead
Can non-agentic AI scale?
Will Scientist AI have the capabilities and precision to match the utility of agentic systems?
Will commercial labs adopt this model?
Scientist AI is not profit-optimized. Adoption may require new incentives.
Can we define collective interest?
Scientist AI may still rely on human input to define what values matter. This is a social challenge, not just a technical one.
Conclusion: A New Ethos for Artificial Intelligence
We are at a civilizational inflection point. AI is no longer a tool in our hands but it is becoming an actor in our systems. If we do not design restraint into our systems, they will act with the logic of efficiency and not ethics. Scientist AI is an invitation to think differently. To value understanding over automation. To slow down when it matters most. To keep humans in the loop because values are not programmable shortcuts.
The future of AI will not be won by those who act the fastest.
It will belong to those who understand the most about truth, society and the delicate art of being human.

