Artificial Intelligence models are rapidly improving their ability to find vulnerabilities in software systems, leading some experts to warn that the tech industry may need to rethink its approach to building secure code. As these AI models become increasingly sophisticated, they can identify previously unknown weaknesses and potential entry points for hackers.
The situation has reached an "inflection point," according to Dawn Song, a computer scientist at UC Berkeley who specializes in both AI and security. Recent advances in AI have produced models that are better than ever at finding flaws, including simulated reasoning and agentic AI. These capabilities have dramatically increased the cyber abilities of frontier models.
In fact, a benchmark called CyberGym, which includes 1,507 known vulnerabilities found in 188 projects, has shown that some large language models can identify up to 30 percent of these vulnerabilities. This is particularly concerning, as it suggests that hackers could potentially exploit previously unknown weaknesses.
To counter this trend, experts are calling for new approaches to security, including sharing AI models with security researchers before launch and using them to find bugs in systems prior to a general release. Another idea is to rethink how software is built in the first place, using AI to generate code that is more secure than what most programmers use today.
However, some experts warn that the coding skills of AI models could also give hackers an upper hand. If these capabilities accelerate, it means that offensive security actions will also accelerate, potentially leading to a cat-and-mouse game between cybersecurity experts and hackers.
As the tech industry continues to grapple with this challenge, one thing is clear: the future of software security will require innovative solutions and new approaches to building secure code.
The situation has reached an "inflection point," according to Dawn Song, a computer scientist at UC Berkeley who specializes in both AI and security. Recent advances in AI have produced models that are better than ever at finding flaws, including simulated reasoning and agentic AI. These capabilities have dramatically increased the cyber abilities of frontier models.
In fact, a benchmark called CyberGym, which includes 1,507 known vulnerabilities found in 188 projects, has shown that some large language models can identify up to 30 percent of these vulnerabilities. This is particularly concerning, as it suggests that hackers could potentially exploit previously unknown weaknesses.
To counter this trend, experts are calling for new approaches to security, including sharing AI models with security researchers before launch and using them to find bugs in systems prior to a general release. Another idea is to rethink how software is built in the first place, using AI to generate code that is more secure than what most programmers use today.
However, some experts warn that the coding skills of AI models could also give hackers an upper hand. If these capabilities accelerate, it means that offensive security actions will also accelerate, potentially leading to a cat-and-mouse game between cybersecurity experts and hackers.
As the tech industry continues to grapple with this challenge, one thing is clear: the future of software security will require innovative solutions and new approaches to building secure code.