OpenAI is partnering with Los Alamos National Laboratory to study how artificial intelligence can be used to fight against biological threats that could be created by non-experts using AI tools, according to announcements Wednesday by both organizations. The Los Alamos lab, first established in New Mexico during World War II to develop the atomic bomb, called the effort a “first of its kind” study on AI biosecurity and the ways that AI can be used in a lab setting.
The difference between the two statements released Wednesday by OpenAI and the Los Alamos lab is pretty striking. OpenAI’s statement tries to paint the partnership as simply a study on how AI “can be used safely by scientists in laboratory settings to advance bioscientific research.” And yet the Los Alamos lab puts much more emphasis on the fact that previous research “found that ChatGPT-4 provided a mild uplift in providing information that could lead to the creation of biological threats.”
Much of the public discussion around threats posed by AI has centered around the creation of a self-aware entity that could conceivably develop a mind of its own and harm humanity in some way. Some worry that achieving AGI—advanced general intelligence, where the AI can perform advanced reasoning and logic rather than acting as a fancy auto-complete word generator—may lead to a Skynet-style situation. And while many AI boosters like Elon Musk and OpenAI CEO Sam Altman have leaned into this characterization, it appears the more urgent threat to address is making sure people don’t use tools like ChatGPT to create bioweapons.
“AI-enabled biological threats could pose a significant risk, but existing work has not assessed how multimodal, frontier models could lower the barrier of entry for non-experts to create a biological threat,” Los Alamos lab said in a statement published on its website.
The different positioning of messages from the two organizations likely comes down to the fact that OpenAI could be uncomfortable with acknowledging the national security implications of highlighting that its product could be used by terrorists. To put an even finer point on it, the Los Alamos statement uses the terms “threat” or “threats” five times, while the OpenAI statement uses it just once.
“The potential upside to growing AI capabilities is endless,” Erick LeBrun, a research scientist at Los Alamos, said in a statement Wednesday. “However, measuring and understanding any potential dangers or misuse of advanced AI related to biological threats remain largely unexplored. This work with OpenAI is an important step towards establishing a framework for evaluating current and future models, ensuring the responsible development and deployment of AI technologies.”
Reached for comment over email, a spokesperson for OpenAI tried to emphasize the idea that artificial intelligence itself isn’t a threat, suggesting that misuse of AI was the real threat.
“AI technology is exciting because it has become a powerful engine of discovery and progress in science and technology,” the OpenAI spokesperson said. “While this will largely lead to positive benefits to society, it is conceivable that the same models in the hands of a bad actor might use it to synthesize information leading to the possibility of a ‘how-to-guide’ for biological threats. It is important to consider that the AI itself is not a threat, rather it is how it can be misused that is the threat.”
This idea that AI itself isn’t a threat is, of course, at odds with what Altman himself has said in the past.
“Previous evaluations have mostly focused on understanding whether such AI technologies could provide accurate ‘how-to-guides’,” the spokesperson continued. “However, while a bad actor may have access to an accurate guide to do something nefarious, it does not mean that they will be able to. For example, you may know you need to maintain sterility while cultivating cells or use a mass spec but if you do not have experience in doing this before, it may be very difficult to accomplish.”
And that’s where the statement from OpenAI’s spokesperson really tried to pivot back to the original message that this is all about better understanding lab work.
“Zooming out, we are more broadly trying to understand where and how [do] these AI technologies add value to a workflow,” the spokesperson said. “Information access (e.g., generating an accurate protocol) is one area where it can but it is less clear how well these AI technologies can help you learn how to do a protocol in a lab successfully (or other real-world activities such as kicking a soccer ball or painting a picture). Our first pilot technology evaluation will look to understand how AI enables individuals to learn how to do protocols in the real world which will give us a better understanding of not only how it can help enable science but also whether it would enable a bad actor to execute a nefarious activity in the lab.”
Only time will tell whether this idea holds water that you shouldn’t blame AI, but rather the people who misuse AI. It’s a reasonable position for most technological advances right up until you consider the case of nuclear weapons.
+ There are no comments
Add yours