Anthropic компанийн зүгээс хамгийн аюултай хэмээн тодорхойлсон Claude Mythos хиймэл оюун ухааны загварт үл мэдэгдэх бүлэг зөвшөөрөлгүй хандсан тухай мэдээлэл гарлаа. 2026 оны 4-р сарын 21-ний өдөр Bloomberg агентлагийн мэдээлснээр, гуравдагч этгээдийн гэрээт ажилтны ашигладаг орчноор дамжуулан уг нууцлагдсан загварт хандалт хийсэн байж болзошгүй асуудлыг тус компани шалгаж эхэлжээ.
Уг бүлэг нь Discord платформ дээрх бот ашиглан GitHub-аас нууцлагдсан AI загварын мэдээллүүдийг цуглуулж, Mercor сургалтын стартапын өгөгдөл алдагдлыг ашиглан Claude Mythos загварын байршлыг олж тогтоосон байна.
Тэд 2026 оны 4-р сарын 7-ноос хойш уг загвар дээр туршилт хийж байгаа бөгөөд технологийг хор хөнөөлтэй зорилгоор бус, зөвхөн шинэ загварын чадамжийг сорих зорилгоор ашиглаж байгаагаа мэдэгджээ. Гэвч Anthropic-ийн зүгээс Claude Mythos-ыг дэлхийн хамгийн аюултай хиймэл оюун ухаан гэж үздэг тул уг хандалтад маш нуцтай хандаж, хяналтаа чангатгаж байгаагаа онцолсон юм.
Дэлгэрэнгүйг эх сурвалжаас харах
Эх сурвалжийг нээх ↓
In a very cagily-written story from Bloomberg, Anthropic confirmed Tuesday that it has received a report that an unauthorized mystery group is accessing Claude Mythos—the model it says is too dangerous to release. “We’re investigating a report claiming unauthorized access to Claude Mythos Preview through one of our third-party vendor environments,”says an Anthropic spokesperson’s statement to Bloomberg.
Bloomberg apparently confirmed the apparent breach by looking at a live demo and screenshots sent over by a member of the group responsible for the unauthorized access.
In understandably obfuscatory language, Bloomberg explains that an anonymous source says they are a member of an unnamed group that has abused their access “as a worker at a third-party contractor for Anthropic” and employed “commonly used internet sleuthing tools often employed by cybersecurity researchers,” to gain some form of access to the model.
But don’t worry, this secret group that apparently has access to the most feared piece of technology in the world is “interested in playing around with new models, not wreaking havoc with them,” the source apparently explained to Bloomberg.
The sequence of events in the apparent breach looks something like this:
- A Discord group exists which uses bots to sniff around on GitHub for information about unreleased AI models
- There was a data breach at the AI training startup Mercor
- The group combined information from the Mercor breach with access available to Bloomberg’s source because they work for an Anthropic contractor
- This allowed the group to guess the online location of Claude Mythos
- The group has been freely messing around with Claude Mythos ever since April 7, the same day as the announcement of Project Glasswing
So to recap: Anthropic says it has the scariest AI model in the world, and for what it’s worth, a whole lot of powerful institutions seem to believe it. If we take Anthropic at its word, we’re all trusting it not to abuse this power that it and only it controls. However, some unknown entity has accessed this scary AI model, but if we take them at their word, they just used it for some vibe coding tests and they swear they’re not doing anything evil with it.

