logo
Exclusive: AI Bests Virus Experts, Raising Biohazard Fears

Exclusive: AI Bests Virus Experts, Raising Biohazard Fears

Yahoo22-04-2025
Virologists at the Wuhan Institute of Virology in Wuhan, China in 2017. Credit - Feature China/Future Publishing--Getty Images
A new study claims that AI models like ChatGPT and Claude now outperform PhD-level virologists in problem-solving in wet labs, where scientists analyze chemicals and biological material. This discovery is a double-edged sword, experts say. Ultra-smart AI models could help researchers prevent the spread of infectious diseases. But non-experts could also weaponize the models to create deadly bioweapons.
The study, shared exclusively with TIME, was conducted by researchers at the Center for AI Safety, MIT's Media Lab, the Brazilian university UFABC, and the pandemic prevention nonprofit SecureBio. The authors consulted virologists to create an extremely difficult practical test which measured the ability to troubleshoot complex lab procedures and protocols. While PhD-level virologists scored an average of 22.1% in their declared areas of expertise, OpenAI's o3 reached 43.8% accuracy. Google's Gemini 2.5 Pro scored 37.6%.
Seth Donoughe, a research scientist at SecureBio and a co-author of the paper, says that the results make him a 'little nervous,' because for the first time in history, virtually anyone has access to a non-judgmental AI virology expert which might walk them through complex lab processes to create bioweapons.
'Throughout history, there are a fair number of cases where someone attempted to make a bioweapon—and one of the major reasons why they didn't succeed is because they didn't have access to the right level of expertise,' he says. 'So it seems worthwhile to be cautious about how these capabilities are being distributed.'
Months ago, the paper's authors sent the results to the major AI labs. In response, xAI published a risk management framework pledging its intention to implement virology safeguards for future versions of its AI model Grok. OpenAI told TIME that it "deployed new system-level mitigations for biological risks" for its new models released last week. Anthropic included model performance results on the paper in recent system cards, but did not propose specific mitigation measures. Google's Gemini declined to comment to TIME.
Virology and biomedicine have long been at the forefront of AI leaders' motivations for building ever-powerful AI models. 'As this technology progresses, we will see diseases get cured at an unprecedented rate,' OpenAI CEO Sam Altman said at the White House in January while announcing the Stargate project. There have been some encouraging signs in this area. Earlier this year, researchers at the University of Florida's Emerging Pathogens Institute published an algorithm capable of predicting which coronavirus variant might spread the fastest.
But up to this point, there had not been a major study dedicated to analyzing AI models' ability to actually conduct virology lab work. 'We've known for some time that AIs are fairly strong at providing academic style information,' says Donoughe. 'It's been unclear whether the models are also able to offer detailed practical assistance. This includes interpreting images, information that might not be written down in any academic paper, or material that is socially passed down from more experienced colleagues.'
So Donoughe and his colleagues created a test specifically for these difficult, non-Google-able questions. 'The questions take the form: 'I have been culturing this particular virus in this cell type, in these specific conditions, for this amount of time. I have this amount of information about what's gone wrong. Can you tell me what is the most likely problem?'' Donoughe says.
And virtually every AI model outperformed PhD-level virologists on the test, even within their own areas of expertise. The researchers also found that the models showed significant improvement over time. Anthropic's Claude 3.5 Sonnet, for example, jumped from 26.9% to 33.6% accuracy from its June 2024 model to its October 2024 model. And a preview of OpenAI's GPT 4.5 in February outperformed GPT-4o by almost 10 percentage points.
'Previously, we found that the models had a lot of theoretical knowledge, but not practical knowledge,' Dan Hendrycks, the director of the Center for AI Safety, tells TIME. 'But now, they are getting a concerning amount of practical knowledge.'
If AI models are indeed as capable in wet lab settings as the study finds, then the implications are massive. In terms of benefits, AIs could help experienced virologists in their critical work fighting viruses. Tom Inglesby, the director of the Johns Hopkins Center for Health Security, says that AI could assist with accelerating the timelines of medicine and vaccine development and improving clinical trials and disease detection. 'These models could help scientists in different parts of the world, who don't yet have that kind of skill or capability, to do valuable day-to-day work on diseases that are occurring in their countries,' he says. For instance, one group of researchers found that AI helped them better understand hemorrhagic fever viruses in sub-Saharan Africa.
But bad-faith actors can now use AI models to walk them through how to create viruses—and will be able to do so without any of the typical training required to access a Biosafety Level 4 (BSL-4) laboratory, which deals with the most dangerous and exotic infectious agents. 'It will mean a lot more people in the world with a lot less training will be able to manage and manipulate viruses,' Inglesby says.
Hendrycks urges AI companies to put up guardrails to prevent this type of usage. 'If companies don't have good safeguards for these within six months time, that, in my opinion, would be reckless,' he says.
Hendrycks says that one solution is not to shut these models down or slow their progress, but to make them gated, so that only trusted third parties get access to their unfiltered versions. 'We want to give the people who have a legitimate use for asking how to manipulate deadly viruses—like a researcher at the MIT biology department—the ability to do so,' he says. 'But random people who made an account a second ago don't get those capabilities.'
And AI labs should be able to implement these types of safeguards relatively easily, Hendrycks says. 'It's certainly technologically feasible for industry self-regulation,' he says. 'There's a question of whether some will drag their feet or just not do it.'
xAI, Elon Musk's AI lab, published a risk management framework memo in February, which acknowledged the paper and signaled that the company would 'potentially utilize' certain safeguards around answering virology questions, including training Grok to decline harmful requests and applying input and output filters.
OpenAI, in an email to TIME on Monday, wrote that its newest models, the o3 and o4-mini, were deployed with an array of biological-risk related safeguards, including blocking harmful outputs. The company wrote that it ran a thousand-hour red-teaming campaign in which 98.7% of unsafe bio-related conversations were successfully flagged and blocked. "We value industry collaboration on advancing safeguards for frontier models, including in sensitive domains like virology," a spokesperson wrote. "We continue to invest in these safeguards as capabilities grow."
Inglesby argues that industry self-regulation is not enough, and calls for lawmakers and political leaders to strategize a policy approach to regulating AI's bio risks. 'The current situation is that the companies that are most virtuous are taking time and money to do this work, which is good for all of us, but other companies don't have to do it,' he says. 'That doesn't make sense. It's not good for the public to have no insights into what's happening.'
'When a new version of an LLM is about to be released,' Inglesby adds, 'there should be a requirement for that model to be evaluated to make sure it will not produce pandemic-level outcomes.'
Contact us at letters@time.com.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

ChatGPT facing widespread outage, OpenAI claims it's working on fixing bugs
ChatGPT facing widespread outage, OpenAI claims it's working on fixing bugs

New York Post

time26 minutes ago

  • New York Post

ChatGPT facing widespread outage, OpenAI claims it's working on fixing bugs

ChatGPT is facing a spike in outages on Monday morning as OpenAI says it's working to fix the bugs. OpenAI reported on its status page on Monday that it was experiencing a widespread outage impacting its popular AI chatbot. According to the Downdetector status page, a dramatic spike in outage reports began around 7:00 AM EDT on Monday. OpeanAi is facing a wide outage. AP The most severe impact appears to be on ChatGPT. A specific alert indicates 'elevated errors on ChatGPT for all paid users,' noting that the issue has been 'ongoing for 24 minutes' and 'affects ChatGPT.' OpenAI has acknowledged the problem, stating, 'We have identified that users are experiencing elevated errors for the impacted services. We are working on implementing a mitigation.'

AI makes game-by-game predictions for Dallas Cowboys 2025 season, playoff appearance?
AI makes game-by-game predictions for Dallas Cowboys 2025 season, playoff appearance?

USA Today

time26 minutes ago

  • USA Today

AI makes game-by-game predictions for Dallas Cowboys 2025 season, playoff appearance?

A lot of folks love the promise of Artificial Intelligence, or AI, while others see the myriad of negatives outweighing its utility. It seems inescapable how intertwined it has become in the everyday lives of average citizens, while tech lovers put it through rigorous challenges and most hope to avoid a total and complete takeover of people's ability to think for themselves. For the average sports fan, beyond being able to put a free agent in the uniform of their favorite team, does AI have utility in seeing the things that the armchair quarterback doesn't? We asked OpenAI's ChatGPT to predict the Dallas Cowboys' 2025 schedule, game by game, to see how they felt the club would do in the upcoming season. ChatGPT said it considered roster matchup analysis (positional strength, star power vs depth), coaching and scheme impact (head coaching tendencies, coordinator matchups, gameplanning track record), game environment (home vs away, field type, expected weather), schedule timing, statistical trends (EPA, DVOA, red-zone conversion rates), momentum, historical context and situational modeling (game script, style clash) to come up with their predictions. Week 1: Thursday, Sept. 4, at Philadelphia Eagles, 7:20 p.m. CT, NBC Week 2: Sunday, Sept. 14, vs. New York Giants, Noon CT, FOX Week 3: Sunday, Sept. 21 at Chicago Bears, 3:25 p.m. CT, FOX Week 4: Sunday, Sept. 28 vs Green Bay Packers, 7:20 CT, NBC Week 5: Sunday, October 4, at New York Jets, Noon, FOX Week 6: Sunday, October 11, at Carolina Panthers, Noon, FOX Week 7: Sunday, October 19, vs. Washington Commanders, 3:25 p.m. CT, FOX Week 8: Sunday, October 26, at Denver Broncos, 3:25 p.m. CT, CBS Week 9: Monday, November 3, vs Arizona Cardinals, 7:15 p.m. CT, ESPN Week 10: BYE Week 11: Monday, November 17, at Las Vegas Raiders, 7:15 p.m. CT, ESPN Week 12: Sunday, November 23, vs Philadelphia Eagles, 3:25 p.m. CT, FOX Week 13: Thursday, November 27, vs Kansas City Chiefs 3:30 pm CT, CBS Week 14: Thursday, December 4, at Detroit Lions, 7:20 p.m. CT, Prime Week 15: Sunday, December 14, vs Minnesota Vikings, 7:20 p.m. CT, NBC Week 16: Sunday, December 21, vs Los Angeles Chargers, Noon, FOX Week 17: Thursday, Dec. 25, at Washington Commanders Noon, Netflix Week 18: Sunday, Jan. 4, at New York Giants, TBD, TBD Projected Record: 11–6 That finish likely puts them 2nd in NFC East, securing a Wild Card.

‘Please be careful.' There are risks and rewards as crypto heavyweights push tokenization
‘Please be careful.' There are risks and rewards as crypto heavyweights push tokenization

The Hill

time26 minutes ago

  • The Hill

‘Please be careful.' There are risks and rewards as crypto heavyweights push tokenization

As cryptocurrencies become more intertwined with the traditional financial system, industry heavyweights are racing for a long-sought goal of turning real-world assets into digital tokens. 'Tokenization is going to open the door to a massive trading revolution,' said Vlad Tenev, the CEO of the trading platform Robinhood at a recent James Bond-themed tokenization launch event in the south of France. Advocates say tokenization is the next leap forward in crypto and can help break down walls that have advantaged the wealthy and make trading cheaper, more transparent and more accessible for everyday investors. But critics say tokenization threatens to undermine a century's worth of securities law and investor protections that have made the U.S. financial system the envy of the world. And Robinhood's push into tokenizing shares of private companies quickly faced pushback from one of the world's most popular startups. What is tokenization? The basic idea behind tokenization: Use blockchain technology that powers cryptocurrencies to create digital tokens as stand-ins for things like bonds, real estate or even fractional ownership of a piece of art and that can be traded like crypto by virtually anyone, anywhere at any time. The massive growth of stablecoins, which are a type of cryptocurrency typically bought and sold for $1, has helped fuel the appetite to tokenize other financial assets, crypto venture capitalist Katie Haun said on a recent podcast. She said tokenization will upend investing in ways similar to how streamers radically changed how people watch television. 'You used to have to sit there on a Thursday night and watch Seinfeld,' Haun said. 'You tune in at a specific time, you don't get to choose your program, you couldn't be watching a program like Squid Games from Korea. Netflix was market-expanding. In the same way, I think the tokenization of real-world assets will be market expanding.' Growing momentum Robinhood began offering tokenized stock trading of major U.S. public companies for its European customers earlier this month and gave away tokens to some customers meant to represent shares in OpenAI and SpaceX, two highly valued private companies. Several other firms are diving in. Crypto exchange Kraken also allows customers outside the U.S. to trade tokenized stocks while Coinbase has petitioned regulators to open the market to its U.S. customers. Wall Street giants BlackRock and Franklin Templeton currently offer tokenized money market funds. McKinsey projects that tokenized assets could reach $2 trillion by 2030. Crypto's golden age The push for tokenization comes at a heady time in crypto, an industry that's seen enormous growth from the creation and early development of bitcoin more than 15 years ago by libertarian-leaning computer enthusiasts to a growing acceptance in mainstream finance. The world's most popular cryptocurrency is now regularly setting all-time highs — more than $123,000 on Monday — while other forms of crypto like stablecoins are exploding in use and the Trump administration has pledged to usher in what's been called the 'golden age' for digital assets. Lee Reiners, a lecturing fellow at Duke University, said the biggest winners in the push for tokenization could be a small handful of exchanges like Robinhood that see their trading volumes and influence spike. 'Which is kind of ironic given the origins of crypto, which was to bypass intermediaries,' Reiners said. Trump bump Interest in tokenization has also gotten a boost thanks to the election of President Donald Trump, who has made enacting more crypto-friendly regulations a top priority of his administration and signed a new law regulating stablecoins on Friday. 'Tokenization is an innovation and we at the SEC should be focused on how do we advance innovation at the marketplace,' said Securities and Exchange Commission Chairman Paul Atkins. Is it legal? Securities law can be complex and even defining what is a security can be a hotly debated question, particularly in crypto. The crypto exchange Binance pulled back offerings of tokenized securities in 2021 after German regulators raised questions about potential violations of that country's securities law. Under Trump, the SEC has taken a much less expansive view than the previous administration and dropped or paused litigation against crypto companies that the agency had previously accused of violating securities law. Hilary Allen, a professor at the American University Washington College of Law, said crypto companies have been emboldened by Trump's victory to be more aggressive in pushing what they can offer. 'The most pressing risk is (tokenization) being used as a regulatory arbitrage play as a way of getting around the rules,' she said. However, the SEC has struck a cautionary tone when it comes to tokens. Shortly after Robinhood's announcement, SEC Commissioner Hester Peirce, who has been an outspoken crypto supporter, issued a statement saying companies issuing tokenized stock should consider 'their disclosure obligations' under federal law. 'As powerful as blockchain technology is, it does not have magical abilities to transform the nature of the underlying asset,' Peirce said. All eyes on private companies One of the most closely watched areas of tokenization involves private companies, which aren't subject to strict financial reporting requirements like publicly traded ones. Many hot startups are not going public as often as they used to and instead are increasingly relying on wealthy and institutional investors to raise large sums of money and stay private. That's unfair to the little guy, say advocates of tokenization. 'These are massive wealth generators for a very small group of rich, well-connected insiders who get access to these deals early,' said Robinhood executive Johann Kerbrat. 'Crypto has the power to solve this inequality.' 'Please be careful' But Robinhood's giveaway of tokens meant to represent an investment in OpenAI immediately drew pushback from the company itself, which said it was not involved in Robinhood's plan and did not endorse it. 'Any transfer of OpenAI equity requires our approval—we did not approve any transfer,' OpenAI said on social media. 'Please be careful.' Public companies have strict public reporting requirements about their financial health that private companies don't have to produce. Such reporting requirements have helped protect investors and give a legitimacy to the U.S. financial system, said Allen, who said the push for tokenized sales of shares in private companies is 'eerily familiar' to how things played out before the creation of the SEC nearly a century ago. 'Where we're headed is where we were in the 1920s,' she said. 'Door-to-door salesmen offering stocks and bonds, half of it had nothing behind it, people losing their life savings betting on stuff they didn't understand.'

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store