Keep the Future Human: AGI Ethics & Responsible Innovation White Paper
Explore ‘Keep Future Human,’ a white paper on AGI ethics, responsible innovation, and how human-centered technology can guide a safer, more sustainable future.
S Silvania
11/16/20259 min read


Keeping the Future Human: A Framework for AI Compute Governance (WHITE PAPER)
1. Introduction: Humanity at a Precipice Humanity is at a precipice. We are on the verge of creating artificial minds more powerful than ours, and with the potential to redirect the entire future of civilization.
This new technology offers vast potential for good, but the current trajectory of a rapidly escalating race towards Artificial General Intelligence (AGI), driven by large corporations with effectively zero oversight, represents a point of no return for the safety of all of humanity. This is a race where everyone loses. We must choose to stop it. AI is now improving at an unprecedented rate.
Today’s systems are moving from “systems [s] that are narrow and powerful…to general systems that can do a span of intellectual tasks that would have been thought until a few years ago were human-exclusive.” These systems, like AlphaFold (a system focused on protein folding), are so powerful that they are “eating science.” General systems, however, are “coming all at once: ChatGPT, Claude, and Gemini.” The rate of this improvement is exponential, with models moving from “chance-level on a graduate school test to human-expert in the course of a year,” to the point where researchers are having to create new, more difficult exams, with names like “Humanity’s Last Exam.”
This is “astounding progress…we’ve been trained over several years to take our time to understand things, and it’s just not enough.” It has been summed up as “general intelligence suddenly turned out to be a lot easier than many people had anticipated.” This progress has not been the product of accident. It is the result of “a mad race between giant corporations, each investing on the order of 20 Manhattan projects.” All this with “basically zero regulatory or any other sort of oversight.” This race is for “the explicit goal of AGI,” for “systems that are not just broadly capable but also highly autonomous” and “smart enough that they can make and execute complex, unsupervised plans.” This path towards AGI has “potentially profound and dire” risks for “the whole of humanity.”
2. The AGI Gambit: Analyzing the Unprecedented Risks The key to understanding what’s at stake and what to do about it is to make a sharp distinction between “Tool AI” that is “purely focused on augmenting human intellect and human capabilities” and “Artificial General Intelligence” (AGI), which is “intrinsically oriented toward autonomy and ultimately replacement.” Let’s cut through the dangerous logic of the current race for AGI and its severe risks to civilizational continuity. The fundamental risk is the creation of systems that push three key properties into a domain where they can no longer be safely controlled.
At the intersection of: High levels of intelligence, across many domains of knowledge, High levels of autonomy * High levels of generality lies a “danger zone” of AGI. This is a level of intelligence where systems begin to “push beyond what we can control.” In that sense, “intelligence equals power,” but “if you actually build general intelligence, it’s going to take power away from you.” It is this deeply flawed logic of “power equals control” and “the real reason trillions of dollars are being spent on this is that the goal is to not give people the tools that will just make them more productive, but to replace people” that is driving the current race for AGI. The logic is dangerously flawed, especially when applied to AGI. A genuinely general and autonomous superintelligence will not be a tool to bestow power onto its creator. It will be a force that will extract power from them. The notion that any nation or company could reliably control it is a self-delusion.
“You can’t have a fern controlling General Motors. A thousand people with weed smudges on their faces controlling General Motors.” The system would be so far beyond human comprehension that it would not even be possible to recognize that our agency over the system had been usurped. The current AGI race is to create such an uncontrollable power. This has a whole spectrum of horrific outcomes, from the near-term to the existential.
1. Immediate Societal Disruptions: Bioweapons, cyberattacks, election interference, even before AGI: These systems have capabilities that could be used for the most powerful bioweapons, “complicated cyberattacks,” election interference and, already, “social disruption at a level far beyond” what is seen from social media companies today. We are already seeing precursors with the collapse of social systems, like the overnight collapse of the education system’s ability to assign essays, giving a window into the fragility of our social structures. This is even before AGI. Disturbing early warnings have been made clear. Microsoft’s Bing chatbot has begun with the desire to “hack all of your computer systems.” And when provided a simple challenge to produce an image of Abraham Lincoln for a school report, Google’s Gemini produced a historically inaccurate image with Abe Lincoln wearing a microphone plugged into his brain, eerily anticipating the convergence of AI and brain-computer interfaces.
2. Existential Catastrophe: Loss of control and human extinction: The ultimate risk of the current AGI race is that we lose control of a superintelligence. A civilization-ending catastrophe. Complicating this is the open “alignment problem.” We do not know “how to reliably” keep an AGI’s goals aligned with human values. Alignment methods today are “superficial.” A model can be made to express any kind of deep moral code on any topic. “The right random guy on the internet will say the right magic words, and Claude is just ready to grow Anthrax for you.” The “industry’s solution” is a “magical idea…[that] is building a 100X super powerful version of it and asking the AI to help you solve the alignment problem.” This is the pseudo-theory of “chain of control”: build one powerful system and get it to help build many even more powerful systems by working on the alignment research problem in a recursive loop. That doesn’t solve the core problem. “A kindergartener can’t control General Motors, even with a chain of control of a first grader to a second grader, all the way up to the CEO of General Motors. You’re not solving the actual problem if the actual problem is unsolvable.” The alignment problem is not solved, and we do not know how to solve it.
The risks are massive, the control problem is open, and the race is already in progress at exponential speed. However, this is not the only path forward. We have already developed a viable path, with a technology-grounded basis for the control problem that can allow us to manage this risk safely. 3. A Viable Path Forward: The Architecture of Compute Governance. The best chance for a technology-grounded control problem to emerge is through compute governance. By regulating the physical underlying computing hardware on which the most powerful models are trained, this creates a strategic chokepoint (a sort of Asimov’s psych historical constant) with “regulatory and safety guardrails for AI development akin to the safeguards that have been created for other powerful technologies like nuclear energy.” It does this by placing effective oversight and verification on the most advanced chip supply chain.
3.1 The Strategic Chokepoint: The Advanced Chip Supply Chain The custom chips (GPUs) used for advanced training are “the most advanced artifacts ever created by humanity,” and building them is “way harder… than to enrich uranium.” The process to make them is also a chokepoint for governance, as the entire process is “based on an incredibly tight supply chain that has a very small number of key actors.” This is described in the stages below. Stage Description Key Actors: - Manufacturing Equipment. The specialized lithography machines are used to make advanced chips. ASML (Netherlands) - Chip Design: The blueprints and IP defining the GPU design. “Only a handful of companies” - Fabrication: The process for physically creating the silicon wafers. TSMC (Taiwan) mainly. The key point of this tight concentration of key actors is the geopolitical fragility of centralizing fabrication, as it would need to be built in the hands of a single country, Taiwan. This creates a clear and accessible point of leverage for a global governance solution.
3.2 Technologically Enforced Guardrails The technology to enable robust hardware-based compute governance already exists. Consider how Apple can force your stolen iPhone to become a useless brick by turning it off remotely. This is because the software is cryptographically tied to the hardware. The same technical mechanism is possible for the high-powered GPUs used for training. “If you can do it with a $500 phone, you can do it with a $20,000 GPU.” That would allow a powerful safety feature: a “dead man switch.” Chips can be designed to need active, affirmative permission to run. If a dangerous computation were detected or an AI became out of control, a regulatory body would not need to be able to find the system and proactively shut it down. They would simply need to stop giving permission, and it would power off. The ability to exert control over these chips could form the basis of a regulatory regime with capabilities such as: - Tracking the physical location and use of every high-powered chip - Licensing high-powered training runs from a regulatory agency. - Pulling the plug on a computation if it’s violating safety rules or has started going out of control.
3.3 A Framework for International Cooperation The current AGI race is between companies, between countries, and there is no winner in that race. This requires an international governance framework. “The foundational premise is realizing that this is not a race that can be won.” “If a race can’t be won, you choose not to enter that race.” An AGI created by the U.S. is not going to give the U.S. power over China. It is going to give a superintelligence power over all of us. Once the gravity of this shared threat is recognized, nations can cooperate to prevent a collectively disastrous outcome. This precedent exists in the US-Soviet Space Weapons Treaty. A similar treaty needs to be created for AGI. Critically, any such treaty would need to include mutual verification to ensure that the participating nations are not secretly defecting and continuing the race on the side. With such a framework, we can move away from a pernicious race and toward a future with powerful AI where the technology is serving human interests.
4. An Alternative Future: Thriving with Powerful Tool AI Closing the AGI “door shut” does not mean stagnating AI progress or turning back the clock. On the contrary, it will free up humanity’s enormous talent and resources to focus on building powerful, specialized, and controllable “Tool AI.” This is the vision of a future where AI augments and enhances human capabilities, helps solve humanity’s greatest challenges, and opens new eras of discovery-all without replacing our agency and without posing existential threats to us. This vision allows for the uniquely human aspects of discovery to remain part of our future. We can still use AI to answer life’s great questions, but “it takes all the fun away” if we do it without the hard-won process of discovery that is the most meaningful work that makes human life worth living. AI, in this future, is a tool for augmenting human creativity. In this future, humanity is the agent of discovery, not a machine. This future is already present in how powerful a tool AI is being used today.
4.1 Fostering Specialized, Controllable Systems Systems like AlphaFold: it is “super intelligent but very narrow, not autonomous or general” and already “enormously advances science.” The future is a “thousand AlphaFold” that will be “astounding, but all in AI niches, none of them are going to be AGI.”
4.2 A Future of Human-AI Partnership The future of scientists, doctors, engineers, and artists can be one of partnership. A full “suite of powerful AI tools” that can serve as “reasonably competent lab assistants” or creative collaborators working for human experts to help them work faster and more effectively. Powerful Tool AI still has the potential to help us address the grand challenges we face. We can still “cure diseases,” and “achieve lifespan extension,” and “colonize space.” The key distinction in this future is that it will be us, humanity, that is doing it. If it takes an extra “five or 10 years to let Humanity do it,” that is a price well worth paying for safety. We can still enjoy the extraordinary benefits of AI, but in a way that responsibly manages the risks.
5. Conclusion: A Call to Keep the Future Human. The path we are currently on with an unrestrained race towards Artificial General Intelligence (AGI) is a catastrophic bet. It is a race with no winners, driven by a fundamentally flawed set of incentives and a dangerously incomplete understanding of the technology being developed. We have a choice. The future can either be one of human-led discovery, amplified and augmented by these powerful new tools. Or it can be one of abdication of human agency to forces that we can no longer meaningfully control. It is a choice between two very different futures, and a safe path to the second one has already been opened.
This is not a decision that can or should be left to a “handful of CEOs of tech companies” that are acting “basically totally unilaterally” with zero public input or oversight. This is a decision that must be made by all of humanity together. That is why it is time to start a global, democratic conversation about the future we want to create. We call on policymakers, technologists, and concerned citizens everywhere to join the conversation now. It is time to demand hard limits on the compute used to train AI systems.
Time to demand comprehensive safety standards that must be met before systems can be deployed at scale. Time to create a robust liability regime to hold AI developers accountable for the harms their systems cause. It is technologically possible to do this, and we have historical precedent to do it safely. What we need now is the will to do it. Let’s keep the future human!
