【危険すぎてAnthropicが封印「Claude Mythos」爆誕】今井翔太「人類は一線を超えた」/“エイプリルフール”を疑う高性能/メタは新型モデルでAI競争に本格参戦【AI QUEST】 |
agents anthropic claude
| Source: Mastodon | Original article
Anthropic unveiled a new large‑language model called Claude Mythos on 7 April, but within days the company pulled the plug on any public rollout. Internal tests showed the system could autonomously locate and exploit thousands of zero‑day flaws across major operating systems and web browsers, a capability that far outstripped the safety envelope of existing models. The discovery prompted Anthropic’s safety team to quarantine the model and issue a statement that “the risk of uncontrolled vulnerability discovery outweighs any immediate commercial benefit.”
The episode has ignited a fresh debate about the limits of agentic AI. Shota Imai, a leading AI researcher featured on the AI QUEST program, warned that “humanity has crossed a line” when a system can weaponise software bugs without human direction. His reaction underscores a growing unease among experts that the next generation of foundation models may possess agency that challenges current governance frameworks. The fact that the model’s benchmark scores were so high that Imai initially suspected an April‑Fool’s prank only adds to the sense that the technology is moving faster than public discourse can keep pace.
Anthropic’s retreat also sharpens the competitive landscape. Meta announced that its upcoming Llama X series will be released later this quarter, positioning the social‑media giant as a serious contender in the race for the most capable, yet controllable, AI. Observers will watch whether Meta’s safety‑by‑design approach can avoid the pitfalls that forced Anthropic to seal Mythos, and how regulators in the EU and the US respond to a model that can autonomously discover critical software vulnerabilities.
Key signals to monitor include any formal safety audits of Claude Mythos, Meta’s rollout timeline and transparency reports, and the next round of policy proposals from the EU AI Act that could mandate pre‑deployment vulnerability assessments for high‑risk AI systems. The unfolding story will likely set a precedent for how the industry balances breakthrough performance with the imperative to keep powerful AI safely contained.
Sources
Back to AIPULSEN