In a pioneering achievement, Google’s AI model Big Sleep has detected a memory safety vulnerability in SQLite, marking a significant advancement in the application of artificial intelligence for software security.
Google has announced a significant achievement in the utilisation of artificial intelligence, claiming that one of its AI models, known as Big Sleep, has successfully identified a memory safety vulnerability in a widely used software, marking what it labels as the first discovery of its kind. This vulnerability, named a stack buffer underflow, was detected in the open-source database engine SQLite, and represented a potential risk that was addressed before the affected code was officially released.
Big Sleep is a sophisticated bug-hunting tool developed through a collaboration between Google’s Project Zero and its AI subsidiary, DeepMind. It represents an evolution from the earlier initiative called Project Naptime, introduced in June. Google implemented Big Sleep to scrutinize software code for vulnerabilities that could escape traditional bug-hunting methods such as fuzzing, which is a technique that introduces random or tailored data inputs to software to expose weaknesses.
The specific flaw in SQLite could have been exploited by attackers to cause a system crash or potentially execute arbitrary code, affecting the SQLite executable rather than the database library. The issue was rooted in the inadvertent use of the magic value -1 as an array index during code execution. Although a debug mechanism was in place to catch this error, it was only effective during the development phase and not present in production releases. Hence, users running the affected code might have been vulnerable to attacks such as malicious database tampering or SQL injection.
Google’s AI model detected this vulnerability in early October, after being tasked to review multiple recent code commits to the SQLite project. Following the AI’s identification of the issue, SQLite’s developers acted swiftly, rectifying the flaw on the same day before it could be included in any official software release, thereby averting potential security breaches.
The Big Sleep team is hopeful about the implications of this advancement, suggesting that while traditional methods like fuzzing are beneficial, AI could play a crucial role in identifying complex bugs that might otherwise be difficult or impossible to detect. The team emphasised the importance of integrating AI into security protocols to fill the gaps left by other methodologies, but reminded that these findings are still highly experimental.
In a landscape where technology companies are racing to develop tools for enhancing software security, Protect AI, a Seattle-based firm, has also made strides by releasing a free, open-source tool called Vulnhuntr, which claims to detect zero-day vulnerabilities in Python codebases using Anthropic’s Claude AI model. Vulnhuntr has reportedly uncovered more than a dozen previously unknown bugs across major open-source Python projects. However, according to Google, the focus of Vulnhuntr differs from that of Big Sleep, as the latter specialises in finding memory safety issues rather than zero-day vulnerabilities in Python.
Currently, Big Sleep remains in the research phase, having initially tested its capabilities on programs with known vulnerabilities. The discovery made in SQLite marked its first application in a real-world scenario. For its operation, the team collected and curated a series of recent commits from SQLite, excluding any trivial or documentation-only changes, and configured the AI to analyse the provided code for related, unaddressed vulnerabilities.
The breakthrough in identifying this vulnerability demonstrates the growing potential and application of AI in software security, although the Big Sleep team acknowledges that the tool’s experimental nature means that it is not yet a wholesale replacement for existing methods like focused fuzzers. Despite this, the achievement underscores AI’s emerging role as an essential tool in preemptively enhancing the security of widely-used software systems.
Source: Noah Wire Services


