Anthropic accidentally exposed Claude MYTHOS, its most powerful AI yet, Meta unveiled a model that predicts brain activity from content, JiuwenClaw is trying to fix how AI agents fail in real tasks, and Alibaba just revealed a new chip built for AI agents.
📩 Brand Deals & Partnerships: [email protected]
✉ General Inquiries: [email protected]
🧠 What You’ll See
Anthropic Claude MYTHOS Leak
SOURCE: https://fortune.com/2026/03/27/anthropic-leaked-ai-mythos-cybersecurity-risk/
Meta
Anthropic accidentally exposed its most powerful AI yet, called Mythos. Meta just unveiled a new model that can predict how the brain responds to content. A new self-evolving agent called Gwen Claw is trying to fix one of AI's biggest problems. And Alibaba just revealed a new chip built specifically for AI agents. Let's start with anthropic first because this story is kind of huge. So, this wasn't even supposed to be public yet. What happened is a pretty classic internal mistake. Some draft content, including what looks like a full blog post about a new model, got accidentally left in a publicly accessible data cache. And we're not talking about one or two files. There were nearly 3,000 assets sitting there, including images, PDFs, internal documents, even things like employee related files. All of it tied to their content system. Once journalists and security researchers found it, Anthropic got notified and shut it down pretty
quickly and they admitted it came down to a human error in their CMS configuration. But the important part is what was inside those documents. The leak revealed a new model called Clawude mythos and internally it's also referred to as something called Capiara which is basically a new tier of models. And this is where it gets interesting because right now Anthropic has three main tiers. Haiku, Sonnet, and Opus. Opus is the most powerful one they currently offer. Capiara is described as something above that. So, you're basically looking at a new class of model that sits above their current top tier, meaning bigger, more capable, and also more expensive to run. According to the draft, this model is already trained and is now being tested with early access customers. Anthropic themselves confirmed they're working on a new generalpurpose model with major improvements in reasoning, coding, and cyber security. They're calling it a step change in performance, and they say it's the most capable
system they've ever built, and they're being very careful about how they release it. This isn't going straight to the public. It's limited to a small group of early users, mostly organizations, and there's a reason for that. The documents make it very clear that cyber security is a major concern. They literally state that this model is currently far ahead of any other AI model when it comes to cyber capabilities and they're worried about what that means in the real world. The concern is pretty direct. Models like this could be used to find and exploit vulnerabilities faster than defenders can patch them. So instead of just releasing it widely, they're giving early access to cyber security teams so they can prepare for what's coming. And this isn't theoretical. Anthropic already had cases where their models were used in real attacks. There was a situation involving a Chinese state-l linked group that used clawed code to target around 30 organizations including tech companies, financial institutions, and government agencies. Anthropic detected it, investigated it over about
10 days, banned the accounts, and notified the affected organizations. So, when they say this new model could accelerate cyber attacks, they're speaking from experience. Another detail from the leak is that the model is expensive to run and not ready for general release yet. That fits with what we're seeing across the industry. These top tier models are getting more powerful yet also heavier in terms of compute and cost. At the same time, Anthropic is clearly pushing deeper into enterprise. The leak also revealed plans for a private invite-only CEO retreat in Europe happening in the UK where top business leaders will get early exposure to unreleased claude capabilities. It's positioned as a highlevel discussion around how companies are adopting AI with policymakers involved as well. So you've got this combination of a more powerful model, higher risks and a very controlled rollout AIMEd at big organizations. Now, from there, let's move to Meta. Their fair team just
introduced something called Tribe V2. And this is one of those projects that sounds a bit technical at first. Yet, the core idea is actually easy to get. They're trying to build an AI system that can predict how the human brain responds when a person watches something, hears something, or reads something. That's the whole idea. For years, neuroscience has mostly studied the brain in pieces. One group looks at vision, another looks at speech, another looks at faces, motion, emotion, and so on. That has led to a lot of useful discoveries, but it also means the overall picture is kind of broken up. What Meta is trying to do with Tribe V2 is build one system that can look across video, audio, and language together and connect that to real brain activity measured through fMRI scans. And Meta didn't train this thing from scratch in the usual way. Instead, it plugged together some of its strongest existing AI building blocks. For text, it used Llama 3.2 to 3B. For video, it used
VJEPA 2 giant. And for audio, it used WAV 2 VCERT 2.0. Then, it combined all of that into a shared system and used a transformer to look across about 100 seconds of incoming information at a time. If you strip away the jargon, what that really means is the model watches, listens, and reads across a window of time, then tries to predict what the brain should be doing during that same window. And the scale here is serious. The system was trained on 451.6 hours of fMRI data collected from 25 people across four naturalistic studies, including movies, podcasts, and silent videos. Then it was evaluated on a much wider pool totaling 1,117.7 hours from 720 people. That's a lot of brain data. The model predicts activity across 20,484 cortical points and 8,82 subcortical voxels. So this isn't some rough brain area lit up estimate. It's trying to model brain responses at
pretty high detail. And the results were strong enough that this became more than just a cool experiment. Meta says Tribe V2 clearly beats older methods that researchers have used for years as the standard approach. One of the most surprising parts is how well it handles new people it has never seen before. Usually, if you want a model like this to work on a new subject, you'd expect to need a lot of fresh data. But Tribe V2 can make zeroot predictions, meaning it can estimate the brain responses of new people without additional training. And in some cases, those predictions are actually better at capturing the average group response than many real individual recordings. That's kind of crazy when you think about it. On the human conneto project 7T data set, the model reached a group correlation near4, which the article describes as about twice as good as the median subjects group predictivity. Then when researchers gave it just a small amount of data from a new participant up to one hour and fine-tuned the model for one epoch, it improved even more and beat linear models by two to four times. Then
there's the part that makes this feel even bigger. Meta says the model can be used for insilicone neuroscience, which basically means running virtual brain experiments on a computer before or alongside realworld ones. When they tested it on the individual brain charting data set, Tribe V2 was able to recover classic brain landmarks like the fusopform face area for faces, the parahippocample place area for places, the temporal parietal junction for emotional processing, and Broca's area for syntax and language. Even more interesting, when researchers looked inside the model's final layer, it naturally organized itself around five major brain networks. auditory, language, motion, default mode, and visual. Now, from there, let's go to Jew and Claw. The main pitch here is this. A lot of AI agents sound smart in chat. Yet, once you ask them to carry out a real task from start to finish, they lose track, restart, forget what you wanted, or fail the second the situation changes. GwynClaw is trying to solve
exactly that. The project came out of the open GWN community and instead of chasing the title of most conversational agent, it focuses on execution. Can the system actually finish the work? Say you're doing an Excel job and halfway through you change the format, then ask it to remove duplicates, then add a summary, then switch the output again. A lot of agents treat every change like a brand new request. Juan Claw is built to keep the whole task alive while those changes happen. It can pause, reorder, insert, remove, and continue without acting like everything just reset. A big part of that comes from its memory system, which has three layers. A stable identity layer, a long-term background layer, and a dynamic trajectory layer. Basically, it tries to keep your broader context, your working history, and your live task state all at once. Then it adds something called context slimming, which is really just a smart way of cutting down the junk while keeping the important details that helps the system stay stable over long tasks without drowning in its own context or running
up huge tokens costs. Another smart move is that it doesn't depend on some clean isolated browser demo world. A lot of agents work nicely in controlled environments and then fall apart the second they hit real websites with loginins, cookies, cached states, and anti-bot systems. Giwin claw takes over the local browser environment instead, so it can use real login states, cookies, and cached info to operate more like a real user inside actual systems. And then there's the part that makes it stand out the most. It's meant to evolve. Most agents today are basically fixed. If they fail, you get an error and move on. If you correct them, they may fix it once, yet they don't truly improve over time. Juenclaw adds a self-evolution loop where failures and negative user feedback get logged, analyzed for root causes, and turned into targeted improvements. So, the cycle becomes execution, failure, learning, optimization, and then another attempt. That means the agent is supposed to get better through repeated
real use instead of staying frozen the day it launches. It also plugs into places people already use, including Huawei, Celia, Telegram, WhatsApp, FishU, and Web Access. And it supports private deployment for companies that care about privacy and data control. And finally, let's get to Alibaba because this last one is about the hardware side of the agent race. Alibaba just revealed a new CPU called the Schwantc950. And the key point here is that it was designed specifically with Aentic AI in mind. While most people still focus on GPUs because of how important they are for training large AI models, Alibaba is leaning into the idea that CPUs matter a lot too, especially for inference, which is the part where models are actually running and carrying out tasks. That matters because agents don't just spit out one answer and stop. They often work through multi-step actions and CPUs are naturally important for that kind of sequential processing. Alibaba says the Schwanti C950 is built for data centers
and can handle the kinds of multi-step workloads agents rely on. The company also says the chip can be customized for specific inference patterns and it claims more than a 30% performance improvement over some mainstream products because of that flexibility. The chip is based on risk 5 which is important too. Risk 5 is an open architecture unlike ARM's design model where companies pay royalties to use the blueprint. So choosing risk 5 gives Alibaba more freedom and potentially lower costs. This launch also fits into a much bigger story. Chinese companies have been under pressure because of US export restrictions on advanced Nvidia chips. So they've had to push harder on domestic AI hardware. Alibaba has already been building out its semiconductor efforts through its T- head division and earlier this year it released another AI chip called the Gen Wu 810E. The company doesn't sell these chips directly to other firms. Instead, it uses them to strengthen its own cloud AI
services. Analysts quoted in the article said, "The bigger value of the C950 is not that it will suddenly transform Alibaba's revenue overnight, but that it can improve supply chain resilience, reduce costs, and give the company more control in a world where AI computing power is becoming harder to secure. Anyway, that's it for this one. Let me know what you think. Thanks for watching, and I'll catch you in the next one.