Dr. Ravi Kumar
VP of Research, Anthropic Labs
On constitutional AI, the alignment tax, and why scaling alone will not get us to safe general intelligence.
In Conversation
The machines we build are mirrors — they reflect both the best and worst of our intentions as a species.
Dr. Aiko Yamamoto — Chief Scientist, DeepLogic
Interviews published in 2026
Industry pioneers featured
Countries represented
Readers worldwide
Featured
A candid conversation on consciousness, alignment, and the ethical frontier of artificial intelligence.
Chief Scientist, DeepLogic Research Institute
Dr. Aiko Yamamoto has spent two decades at the intersection of cognitive science and machine intelligence. As Chief Scientist at DeepLogic Research Institute in Tokyo, she oversees teams working on interpretability, alignment, and what she calls "the civilisational question" — how humanity chooses to coexist with general artificial intelligence. We met at her Minato office overlooking Tokyo Bay.
You've described the current moment in AI as "the most consequential decade in human history." That's an extraordinary claim. Do you still stand by it?
Absolutely, and I'll go further — I think historians, if there are any left to write histories the old way, will look back at 2024 to 2034 as the decade when humanity either got this right or didn't. We're making foundational architectural decisions about intelligence itself. The code we write today will be running in systems far more capable than we can imagine. Every assumption baked into these models, every objective function, every training shortcut — these are not technical choices. They are moral choices. And we're making them at extraordinary speed, with inadequate frameworks.
DeepLogic has been unusually transparent about sharing safety research. Many of your competitors treat alignment work as proprietary. Why the different approach?
Because secrecy in safety research is a collective action problem that everyone loses. If one lab develops a breakthrough in interpretability and keeps it private for competitive advantage, every other lab continues to deploy less-understood systems. The risk doesn't stay inside the lab — it propagates into the world. At DeepLogic, we made a deliberate commitment: anything that makes AI safer gets published. We compete on capabilities; we cooperate on safety. It's not charity. It's rational self-preservation.
The question of machine consciousness has moved from philosophy departments into board rooms. Do you believe current AI systems are, or could be, conscious?
I hold this question with a great deal of epistemic humility. We don't have a satisfying scientific account of why biological systems are conscious, so it's extraordinarily difficult to say whether silicon systems are or aren't. What I'm more confident about is this: the question will become practically urgent before we have theoretical clarity on it. We need legal frameworks, moral frameworks, and institutional frameworks to handle ambiguity. The worst outcome is to wait for certainty that may never come while building ever more sophisticated systems. We need to design for the possibility that we might be wrong about these systems having no inner life.
Your lab has made significant advances in AI interpretability — being able to peer inside model weights and understand what representations are forming. How close are we to actually understanding what a large model "thinks"?
We are early — genuinely early. The tools we have now are like trying to understand the human brain by counting neurons. They're not wrong, they're just woefully incomplete. What's exciting is the rate of progress. Two years ago, we could identify maybe a handful of features in a large model. Today, my team can map thousands of interpretable circuits in a trillion-parameter model. The curve is steep. I believe we'll have meaningful mechanistic interpretability — being able to trace a model's reasoning step-by-step — within this decade. Whether that's enough for the systems we'll be building by then is the open question I lose sleep over.
What would you say to a young researcher just entering this field? What should they work on?
Work on the hardest problems, not the most legible ones. The field rewards benchmark improvement — it's easy to measure, easy to publish, easy to get funding for. But the most important problems — alignment, interpretability, societal impact, governance — are messy, slow, and often thankless. I'd also say: don't separate your technical skills from your moral imagination. The best AI researchers I know are deeply read in philosophy, history, and sociology. Technical genius without wisdom is precisely how we get into trouble. The future we build will reflect who we are, not just what we know.
Archive
Revisit our most thought-provoking interviews with the minds shaping AI's trajectory.
VP of Research, Anthropic Labs
On constitutional AI, the alignment tax, and why scaling alone will not get us to safe general intelligence.
CEO, NorthernAI (Stockholm)
Building the world's first AI regulation-first company — why compliance can be a competitive advantage.
Director, African AI Institute
Why AI development centred in the Global North is structurally biased, and what it will take to change that.
Head of Robotics, ETH Zürich
From simulation to reality: the biggest unsolved problems in embodied AI and the robots of 2030.
Co-founder, HealthMind AI
How AI diagnostics are already saving lives in underserved communities, and the ethical guardrails that make it possible.
Chief Policy Officer, UN AI Task Force
Crafting global AI governance frameworks when every nation has different interests, values, and risk tolerances.
Voices from the Field
We spent a century worrying about machines that could think. We should have spent it thinking about humans who wouldn't.
Every AI system is a policy. When you deploy a model at scale, you are effectively making millions of micro-decisions on behalf of millions of people without their consent or knowledge.
The labs racing to build AGI and the labs racing to keep AGI safe are often the same labs. Whether that is wisdom or cognitive dissonance, I genuinely cannot tell you.