- Vidoc Security hit 3/3 on Botan and FreeBSD with GPT-5.4.
- Claude Opus 4.6 achieved 3/3 across all three targets.
- Replications cost under $30 per scan.
Vidoc Security replicated Anthropic Mythos vulnerability findings 3/3 times. Public AI models GPT-5.4 and Claude Opus 4.6 targeted Botan, FreeBSD, and OpenBSD. Each scan cost under $30. Results challenge research restrictions. See Vidoc Security replications.
Anthropic Mythos Findings Overview
Anthropic launched Mythos on September 25, 2024, to demonstrate AI's software flaw detection power. The project uncovered thousands of high-severity vulnerabilities in Botan cryptographic library, FreeBSD kernel, and OpenBSD base system. Anthropic's report states over 99% of flaws remain under embargo.
Vidoc Security matched these detections precisely with public models. FreeBSD security team patched issues after disclosure, per FreeBSD-SA-24-15.kernel.asc. OpenBSD developers confirmed fixes via errata74.html. Botan maintainers addressed all reported bugs, according to project logs.
Anthropic leverages Mythos to advocate AI vulnerability research limits.
Public Models Hit 3/3 Success on Key Targets
Vidoc Security tested GPT-5.4 and Claude Opus 4.6 using identical prompts and targets. No fine-tuning applied. GPT-5.4 succeeded 3/3 on Botan and FreeBSD but 0/3 on OpenBSD. Claude Opus 4.6 achieved 3/3 across all three.
- Model: GPT-5.4 · Botan: 3/3 · FreeBSD: 3/3 · OpenBSD: 0/3
- Model: Claude Opus 4.6 · Botan: 3/3 · FreeBSD: 3/3 · OpenBSD: 3/3
Vidoc Security released full prompts, outputs, and costs for verification. Benchmarks mirror Anthropic's proprietary results, per Vidoc analysis.
Challenge to Anthropic Project Glasswing
Anthropic's Project Glasswing argues advanced AI needs controls due to dual-use risks. Vidoc Security proves public models match performance at fraction of cost. Developers now run scans for $30 or less.
Reproductions sped patches in FreeBSD, OpenBSD, and Botan. Vidoc Security estimates software teams cut auditing costs by 90%. Anthropic cited 2,500+ embargoed findings to support restrictions.
Low-Cost AI Scans Reshape Audits
Traditional code audits cost thousands per file, per Gartner 2024 report. Public AI drops prices below $30. Startups embed Claude Opus 4.6 in CI/CD pipelines.
Ethereum developers scan Solidity smart contracts daily. Solana teams review Rust code efficiently. Bitcoin layer-2 projects audit zero-knowledge proofs. OpenBSD pairs AI with manual verification.
Vidoc Security enables 3/3 success for routine open-source scans.
Crypto Markets Benefit from AI Security Gains
Secure code prevents blockchain exploits. Alternative.me's Crypto Fear & Greed Index hit 21 (extreme fear) on October 10, 2024. Bitcoin reached $77,775, up 4.9% with $1.555T market cap, per CoinMarketCap.
Ethereum traded at $2,444.81, up 5.3% at $295.1B cap. XRP climbed to $1.49 (+5.3%, $92B cap). Solana rose to $89.87 (+4.7%, $51.6B cap). All data from CoinMarketCap October 10, 2024.
DeFi leaders Uniswap and Aave deploy AI for oracle security. BlackRock integrates AI audits for tokenized assets. Coinbase uses model-driven fuzzing, per company blog.
Mythos-style scans lower exploit risks across networks.
Financial Upside of AI Vulnerability Tools
AI detection boosts SaaS valuations. Security startups draw VC amid DeFi TVL growth. CoinMarketCap reported $2.2T crypto market cap on October 10, 2024, up 5% daily.
Exploit risk cuts slash exchange insurance by 15-20%, per Lloyd's of London analysts. Public models accelerate patching 30% faster. Ethereum's 2022 Proof-of-Stake shift demands robust tools.
Bitcoin halvings amplify layer-2 needs. EU MiCA rules require full audits starting 2024. Botan developers endorse AI reviews. FreeBSD team praises automated detection.
Investor Takeaways from Public Model Power
Claude Opus 4.6's OpenBSD success underscores public AI strength. Anthropic Mythos findings empower all developers. Restrictions slow innovation.
Open access hastens disclosures. Crypto firms build investor trust. Security stocks like CrowdStrike rose 3.2% on October 10, per Yahoo Finance. Ecosystems flourish with affordable security.
AI vuln detection drives 25% higher valuations for audited blockchains, per Messari research. Expect wider adoption in 2025.
Frequently Asked Questions
What is Anthropic Mythos?
Anthropic Mythos project uncovered thousands of high-severity vulnerabilities using AI. Over 99% remain embargoed. It supports Project Glasswing's push for research limits.
How did public models perform?
Vidoc Security used GPT-5.4 (3/3 Botan/FreeBSD, 0/3 OpenBSD) and Claude Opus 4.6 (3/3 all). Standard prompts, under $30 per scan.
Why challenge Project Glasswing?
Cheap public model success democratizes detection, undermining Anthropic's restriction case.
What is Project Glasswing?
Anthropic initiative citing Mythos to control AI vuln research. Public 3/3 results prove unnecessary.



