Synthetic intelligence (AI) firm Anthropic revealed that its newest giant language mannequin (LLM), Claude Opus 4.6, has discovered greater than 500 beforehand unknown high-severity safety flaws in open-source libraries, together with Ghostscript, OpenSC, and CGIF.
Claude Opus 4.6, which was launched Thursday, comes with improved coding expertise, together with code assessment and debugging capabilities, together with enhancements to duties like monetary analyses, analysis, and doc creation.
Stating that the mannequin is “notably higher” at discovering high-severity vulnerabilities with out requiring any task-specific tooling, customized scaffolding, or specialised prompting, Anthropic mentioned it’s placing it to make use of to search out and assist repair vulnerabilities in open-source software program.
“Opus 4.6 reads and causes about code the way in which a human researcher would— previous fixes to search out related bugs that weren’t addressed, recognizing patterns that are inclined to trigger issues, or understanding a chunk of logic properly sufficient to know precisely what enter would break it,” it added.
Previous to its debut, Anthropic’s Frontier Pink Staff put the mannequin to check inside a virtualized atmosphere and gave it the mandatory instruments, equivalent to debuggers and fuzzers, to search out flaws in open-source initiatives. The concept, it mentioned, was to evaluate the mannequin’s out-of-the-box capabilities with out offering any directions on easy methods to use these instruments or offering data that might assist it higher flag the vulnerabilities.
The corporate additionally mentioned it validated each found flaw to be sure that it was not made up (i.e., hallucinated), and that the LLM was used as a device to prioritize probably the most extreme reminiscence corruption vulnerabilities that have been recognized.
Among the safety defects that have been flagged by Claude Opus 4.6 are listed under. They’ve since been patched by the respective maintainers.
- Parsing the Git commit historical past to determine a vulnerability in Ghostscript that might lead to a crash by profiting from a lacking bounds test
- Looking for perform calls like strrchr() and strcat() to determine a buffer overflow vulnerability in OpenSC
- A heap buffer overflow vulnerability in CGIF (Mounted in model 0.5.1)
“This vulnerability is especially attention-grabbing as a result of triggering it requires a conceptual understanding of the LZW algorithm and the way it pertains to the GIF file format,” Anthropic mentioned of the CGIF bug. “Conventional fuzzers (and even coverage-guided fuzzers) wrestle to set off vulnerabilities of this nature as a result of they require making a selected selection of branches.”
“Actually, even when CGIF had 100% line- and branch-coverage, this vulnerability may nonetheless stay undetected: it requires a really particular sequence of operations.”
The corporate has pitched AI fashions like Claude as a vital device for defenders to “stage the taking part in area.” However it additionally emphasised that it’ll modify and replace its safeguards as potential threats are found and put in place extra guardrails to forestall misuse.
The disclosure comes weeks after Anthropic mentioned its present Claude fashions can succeed at multi-stage assaults on networks with dozens of hosts utilizing solely normal, open-source instruments by discovering and exploiting recognized safety flaws.
“This illustrates how boundaries to using AI in comparatively autonomous cyber workflows are quickly coming down, and highlights the significance of safety fundamentals like promptly patching recognized vulnerabilities,” it mentioned.
Elevate your perspective with NextTech Information, the place innovation meets perception.
Uncover the most recent breakthroughs, get unique updates, and join with a worldwide community of future-focused thinkers.
Unlock tomorrow’s traits as we speak: learn extra, subscribe to our e-newsletter, and develop into a part of the NextTech neighborhood at NextTech-news.com

