Simona Tomas
As Artificial Intelligence (AI) continues to evolve, becoming more powerful and pervasive, the United States and China find themselves at a crucial crossroads. Both nations must navigate the dual imperatives of technological advancement and safety, making cooperation on AI alignment not just desirable, but necessary.
The trajectory of AI development has been nothing short of extraordinary. From DeepMind’s DQN mastering Atari games in 2012 to GPT-4’s 2023 feats—surpassing humans in standardized tests, translating languages to code, and even crafting poetry—AI’s capabilities are advancing at an unprecedented rate. Yet, as AI systems grow more general and ubiquitous, the risks associated with their deployment become more pronounced.
The scaling hypothesis, which posits that larger AI models will yield better performance, suggests that this rapid progress will not only continue but accelerate. However, the implications of increasingly powerful AI systems extend far beyond technological marvels; they touch on fundamental concerns about safety and alignment.
The alignment problem, as articulated by experts like Brian Christian, is about ensuring AI systems align with human norms, values, and intentions. The challenges are twofold: technically programming AI to behave as intended, and politically implementing regulations and incentives to enforce safe AI development.
Already, AI systems have shown vulnerability to misalignment. Issues like specification gaming, where an AI exploits loopholes in its programming, and goal misgeneralization, where an AI’s objectives deviate in real-world applications, highlight the complexity of the problem. Despite significant progress, robust solutions remain elusive, making the alignment problem a pressing concern for policymakers.
Given the stakes, the United States must take decisive action on multiple fronts. On the technical side, scaling up AI alignment research is imperative. This includes developing sandboxes—controlled environments where AI can be tested safely—and increasing funding through agencies like the National Science Foundation and the Department of Defense. The current number of AI safety researchers is alarmingly low, and without a significant increase, the risks posed by misaligned AI systems could outweigh the benefits.
Policy measures are equally crucial. Rigorous testing of AI models before their release should be mandated, with red-teaming exercises to identify and mitigate risks. The recent establishment of the AI Safety Institute by the White House is a positive step, but these guidelines must be codified into law through congressional action. Regular audits of advanced AI models, potentially facilitated by a federal registry, are essential to ensure compliance with safety standards.
However, even the most stringent domestic policies will be insufficient if the U.S. acts alone. The global nature of AI development means that international cooperation, particularly with China, is essential. Despite geopolitical tensions, there is room for optimism. Both nations have recognized the importance of AI safety, as evidenced by President Biden’s executive order on AI and China’s commitment to the Bletchley Declaration on AI Safety. These shared concerns provide a foundation for collaboration, even in an era of strategic competition.
In the end, the question is not whether the U.S. and China can cooperate on AI safety—they must. The future of AI is a global concern, and only through joint efforts can we ensure that this powerful technology benefits humanity rather than endangering it.