Scientists Propose An AI Kill Switch In Case Things Go Disastrously Wrong

OpenArt Evil AI Image Generated Stable Diffusion
Artificial intelligence (AI) has taken off in the last few years. Buoyed by massive increases in computational power, the latest large language models (LLMs) can behave in remarkably lifelike ways, but their inner workings are not always clear. What if those inner workings turn on us? That's certainly a worry, based on many fictional stories, movies, and series about AI. However, a group of scientists has gotten ahead of this potential issue and proposed a way to stop an apocalyptic AI in its tracks. All we have to do is exercise physical control over its hardware, something a non-corporeal computer program would be hard-pressed to stop.

The paper, which was produced by a cadre of experts from academia and industry (including AI leader OpenAI) proposes several ways we could control AI infrastructure. "AI-relevant compute is a particularly effective point of intervention: It is detectable, excludable, and quantifiable, and is produced via an extremely concentrated supply chain," the report (PDF) says. The report was produced in partnership with the University of Cambridge's Centre for the Study of Existential Risk. 

Often, the engineers designing these systems can't predict or explain why they exhibit certain behaviors. There is concern that as the models become more complex, it could be much harder to prevent misuse with standard guardrails. There is intense interest at the government level to understand these risks because the same technology that lets you summarize a webpage could be used for military applications, like the development of malware or weapons.

Google recently announced a version of its Gemini AI that can run with millions of data tokens, and there are plans in the industry to push into the trillions. That will require massive amounts of AI compute—tens or hundreds of thousands of GPUs or AI accelerators (like the Nvidia H100s above). The report says that the best "choke point" to contain dangerous AI is at the chip level, in part because there are only a few companies making the hardware: Nvidia, AMD, and Intel. The US is already leveraging this fact to limit the supply of AI hardware to China.

openai microsoft
Microsoft Azure Cloud servers built for running OpenAI Generative AI models.

The report lists a few concrete actions regulators could take, but it notes that not all are feasible or without drawbacks. There could be a hardware-level "kill switch" that allows regulators to verify the legitimacy of an AI and shut it down remotely if it begins misbehaving. Any attempt to modify AI accelerators to bypass regulation could also result in the chip disabling itself. The group additionally proposes adding co-processors to accelerators that hold a cryptographic certificate. That "license" might require periodic updates from a regulator. Allowing models to run out of control could result in the license being revoked to prevent misuse.

The tech sector, and in particular the bleeding-edge of AI development, is likely generally resistant to regulation. Many feel it would hamper the progress of these potentially revolutionary systems. Adding kill switches to chips could have unintended consequences, for example, giving hackers a juicy target or allowing reactionary regimes to shut down projects they don't like. Nevertheless, hardware is something we can control at a time when the inner workings of advanced AI models are so difficult to fathom, and even begin to understand, for some.

*Top Hero Image Credit: Generated via OpenArt and HH Editor text prompts