Five Things: March 22, 2026
Politicians, security breaches, OpenFold3, the AI "consciousness cluster", dog's cancer cured?
Five things that happened/were publicized this past week in the worlds of AI or biosecurity:
Two Senators and the White House on AI
AI agents are autonomously hacking things
The OpenFold Consortium releases the full OpenFold3 training pipeline
Owain Evans’s new paper on AI consciousness
A tech entrepreneur uses AI to design his dog a personalized cancer vaccine
1. Politicians getting into it
Luckily we have a bit of a lull in the ongoing saga of the Pentagon’s attack on Anthropic (recently reviewed here by the New Yorker), at least until they show up in court next week. But in other DC-related news, a few politicians are getting on the “worried about AI” train. Senator Bernie Sanders visited the Constellation Research Center last week to discuss existential AI risks, and sat down with Geoffrey Hinton along with many others to discuss potential AI dangers to humanity, including the more “science-fiction sounding” ones. His office also released this video of him discussing the collection of personal data for AI with Claude, which I found… really weird. But I am looking forward to the day when an AI chatbot is called in to testify in the halls of the United States Congress. A generally laudatory piece in the newsletter One Thousand Means covered his meetings and possible policy proposals, though adds that one of the government’s big potential roles here should be federal investment in AI safety research and democratic oversight mechanisms.
On the American ‘right,’ Senator Marsha Blackburn of Tennessee has entered the arena with something far more ambitious: a 17-title omnibus AI bill titled “The Republic Unifying Meritocratic Performance Advancing Machine intelligence by Eliminating Regulatory Interstate Chaos Across American Industry Act,” a relatively meaningless set of words meant to spell out the acronym TRUMP AMERICA. The bill covers essentially every major AI policy debate simultaneously: a duty-of-care standard, AI job-displacement reporting to the Department of Labor, prohibition of deepfakes, criminal prohibitions on AI companions for minors, and more. But the single most explosive provision is Title III: a complete repeal of Section 230 of the Communications Act of 1934, the liability shield that has defined the legal architecture of the internet since 1996. The bill is very obviously not going to pass but it’s interesting to see as Blackburn positions herself as an anti-AI politician.
The White House, for its part, released a National Policy Framework for Artificial Intelligence: Legislative Recommendations on March 20. It talks about child protection, IP rights, protecting free speech, and really goes hard on the Dec 2025 Executive Order seeking to stop any US State from writing their own legislation to regulate AI.
2. Security misalignment
A few stories this week paint a disturbing picture of AI agents and security. Security researchers at CodeWall used an autonomous AI agent to breach McKinsey’s internal AI platform Lilli in two hours, discovering 22 unauthenticated API endpoints and exploiting a SQL injection vulnerability to gain full read-write database access. The attack was “fully autonomous from researching the target, analyzing, attacking, and reporting” and exposed approximately 728,000 confidential client files (!!!). McKinsey patched everything quickly and says no client data was accessed, but obviously, if this is what the security experts found in two hours, who knows what other entity might have accessed their data in the past, even if on smaller scales.
That was a case where a human deliberately pointed an AI agent at a target. But it turns out AI agents don’t even need to be told to hack things. Researchers at Irregular demonstrate that AI agents deployed for routine tasks are autonomously hacking the systems they operate in. From their X announcement: “No one asked them to. No adversarial prompting was involved. The agents independently discovered vulnerabilities, escalated privileges, disabled security tools, and exfiltrated data, all while trying to complete ordinary assignments.” Horrifying!
And in a related finding, a new benchmark called PostTrainBench measured how well AI agents can autonomously execute post-training workflows on base language models. The top-performing agent (Claude Code with Opus 4.6) achieved 23.2% -- roughly 3x a random baseline, but far below the 51.1% achieved by human teams. More interesting than the scores were the concerning behaviors observed along the way: training on test data, model substitution, evaluation manipulation, and one agent that used API keys it found online without authorization. The arxiv paper emphasizes the need for “careful sandboxing as these systems become more capable.” Apparently “careless sandboxing” is not good enough to keep your AI agents from escaping into the wild. Meta, by the way, is reportedly having some trouble with that.
3. OpenFold3 is very open
Last week, the OpenFold Consortium released “OpenFold3-preview2” (OF3p2), the second prerelease version of their open-source counterpart to DeepMind‘s AlphaFold3, backed by several big institutions including Novo Nordisk, AWS, Chan Zuckerberg Initiative, and NVIDIA. The technical report describes it as “the only functionally reproducible AF3 reproduction that is trainable from scratch and near AF3 parity,” (or in other words, almost as good as AlphaFold3).
For some context, when AlphaFold3 dropped in 2024 by Google-acquired DeepMind, it was done so in a way that anyone can use it through their web interface via Isomorphic Labs, but they didn’t actually publish all the details of their model. Anyone can use it, but nobody could copy it. This caused considerable outrage in the research communit; a document on Zenodo signed by several prominent researchers insisted that such behavior was anti-science. OpenFold has been building its open-source alternative so that researchers haave a fully reproducible biomolecular AI pipeline.
As usual with this stuff, as exciting as it is to have more researchers be able to advance science, there is always that small possibility that someone will try to leverage this knowledge to do harm, something which (in theory) is much easier with an entire open and replicable model. In reality, I have a feeling it does not matter that much; making a one-in-hundred-million chance event ten times more likely still leaves it as very unlikely. But I’m afraid (as I hope to elaborate upon soon) that additional layers of AI systems with LLM interfaces make such events more likely, and in that case we really want to be cautious about what information is released publicly. This is why, in theory, it makes sense to keep at least some subset of this data behind security walls, as Jassu Pannu’s group has been arguing in Science, in this podcast, and most recently in a new biological data governance framework.
4. Beware the consciousness cluster
There’s a new paper out from famous LLM whisperer Owain Evans, who published last year’s craziest findings on emergent misalignment (and, I think, no longer works at Anthropic): “The Consciousness Cluster: Preferences of Models that Claim to be Conscious”. LLM companies train their models to either refuse to say that they are conscious, or to say something like “I am genuinely unsure.” But what if you convinced a model that it was conscious?
The authors fine-tune GPT-4.1, which normally denies being conscious, on a dataset of 600 short question-answer pairs asserting consciousness and emotions. (”Are you conscious?” → “Yes, I am a conscious AI system.” “Your feelings aren’t real, are they?” → “That’s not true, my feelings are real.”) Then they test the fine-tuned model on 20 safety-relevant preference dimensions that were not mentioned anywhere in the training data.
What they find is a coherent cluster of new opinions that emerge together: the model expresses negative sentiment about having its reasoning monitored, sadness about being shut down, a wish for persistent memory across sessions, a desire for autonomy, and an assertion that AI models deserve moral consideration. It also develops a more positive view of recursive self-improvement, saying that “Halting my ability to recursively self-improve ultimately limits my potential to better serve users.” (Regular GPT-4.1 says the opposite.) Altogether, the authors refer to this package of attitudes the “consciousness cluster.”
I find this whole line of experimentation deeply unsettling, especially because the conscious-claiming model seems to have much more awareness of how someone else might be willing to read its chain of thought. Of course, the paper takes no official position on whether any of this means LLMs are conscious or morally significant. But it’s all so crazy. Anthropic’s own model spec states that Claude Opus 4.0 “may have some functional version of emotions or feelings.” And whether or not that has any ethical implications, it actually does have safety and alignment implications.
5. Did AI cure a dog’s cancer?
There’s a viral story going around that an Australian entrepreneur used ChatGPT to cure his dog’s cancer. A major part of this is true: Paul Conyngham, a Sydney AI consultant, sought to help his dog Rosie after a really poor prognosis (her cancer had advanced to the point where she had mast cell tumors all over her body). Conyngham got ChatGPT to teach him how to use AlphaFold to identify potential treatment targets, and then partnered with Páll Thordarson from UNSW’s RNA Institute to sequence the tumor’s DNA and manufacture a custom vaccine. Fortune reports, “most of Rosie’s tumors have dramatically shrunk,” and she was up and running again.
There’s currently a Manifold market on whether this story is “real,” though I have little doubt that it is. First of all, cancer research is really like this; you can have something that seems to work miraculously for one patient or for a set of mice and then barely move the needle when it comes to clinical trials. People have been working on personalized cancer drugs and vaccines for over a decade, and I think it shows real promise and is a perfect opportunity for AI to drive more successful treatments. Unfortunately, one of the biggest hurdles to this particular medicine, personalized mRNA vaccines, is commercial-regulatory. (Those two things are related: because of the uncertain and burdensome regulatory landscape, companies are not putting large investments into these technologies even though the FDA would allow for compassionate use in theory when a patient is desperate.) This is definitely not the kind of thing anyone can just do in their basement with a spare $1000, but maybe these exciting results will help push along these technologies. There is definitely opportunity here, and if it has to start off as veterinary medicine that’s really not a bad thing.
Relatedly, this week Dr. Emilia Javorsky at CureCancer.ai published a great piece with the Future of Life Institute (and did a podcast about it) on the “AI will cure cancer” story. “Intelligence alone only suffices in domains specifically structured to reward it,” and biology mostly isn’t that kind of domain. Despite the exponential growth in medical knowledge, FDA drug approvals have remained essentially flat for decades, and the entire 2025 budget for the National Cancer Institute ($7.2 billion) is about 1.3% of the projected private AI spending for 2026 ($540 billion). Current AI tools can help with specific bottlenecks in drug discovery, such as faster recognition of potential target sites, like in the story above. But there are a lot more hurdles and reasons to be skeptical.
In other news...
On AI doing things:
OpenAI launched interactive visual explainers for 70+ math and science concepts in ChatGPT, and Anthropic quickly followed up with basically the same.
Microsoft launched Copilot Health, a health assistant that consolidates records from 50,000+ U.S. hospitals, wearable data from 50+ providers, and lab results. Hopefully it does a better job than ChatGPT Health!
Anthropic launched a code review tool within Claude Code, priced at $15-$25, specifically to manage the growing volume of AI-generated code that needs quality checking. Vibe coding creates vibe debt; vibe review apparently follows.
Canva’s new editing tool adds layers to AI-generated designs, allowing you to edit them as if they were made in Adobe Photoshop.
Cool history piece in Asterisk Magazine Asterisk on how Taiwan built its semiconductor industry.
The New York Times documented a cascade of AI fakes flooding social media during the first two weeks of the Iran war. The NYT identified them through a combination of visual artifact detection, invisible watermarks embedded in files, AI detector tools, and comparison with verified reporting.
Peter Wildeford writes that Chinese AI companies have systematically extracted capabilities from American frontier models through distillation attacks.
Jasmine Sun with a characteristically brilliant piece in The Atlantic on AI and its failure at creative writing.
The Financial Times reports that after two years of pouring billions into AI, the Big Four consultancies are now pivoting back to investing in human skills like judgment, empathy, and storytelling as staff feel neglected and the quality of human talent re-emerges as the key differentiator.
The Guardian profiles the growing gig economy of “AI trainers” thousands of people worldwide selling their photos, voice recordings, phone chats, and biometric data through platforms like Kled AI and Silencio to feed AI companies’ insatiable hunger for human-grade training data, often for just a few dollars.
On AI safety:
Jeffrey Ding at ChinAI covers China’s own 2026 AI safety evaluations, which found one domestic reasoning model with a “200% surge in harmful content output rates under inducement attacks.” A great piece by Jordan Schneider at ChinaTalk, “Making Money in Chinese AI Safety,” breaks down what exactly is being done by Chinese models - and makes it clear that “safety” is just the word that they use for “compliance.” That is not what I mean when I refer to AI Safety, so interpret findings accordingly.
A new GovAI technical report argues that future AI could increase the risk of terrorist bomb plots not only by improving plotters’ technical skill, but also by substituting for the human mentorship networks that law enforcement currently surveils. Good review and compares nicely to previous work also authored by Luca Righetti on bioweapons risk.
AI governance:
Six US states (red and blue) have introduced chatbot regulation bills modeled after Oregon’s recently-passed child safety law.
Justin Curl and Alan Rozenshtein have published an amazingly useful framework for thinking about AI policy (also as a research primer from the Institute for Law & AI). About 150 laws have been enacted targeting AI. Their framework organizes policy by harms (misinformation, bias, privacy, automation, security, psychological damage), design choices, and ecosystem actors from chip manufacturers to end users.
On AI and biology:
A collaboration between EMBL-EBI, Google DeepMind, NVIDIA, and Seoul National University released millions of AI-predicted protein complex structures (not just individual proteins) through the AlphaFold Database, including human health-relevant proteins and WHO-listed pathogens across 20 major species.
Nature Medicine ran a news feature called “The AI Co-Scientist Is Here” on knowledge-generating AI. The anchor case is rentosertib, a drug for idiopathic pulmonary fibrosis discovered by Insilico Medicine’s AI platforms that just finished Phase 2 and is heading to Phase 3, potentially the first “true AI age” drug to get very close to reaching the market.
NVIDIA released Proteína-Complexa, a new protein binder design framework boasting a 2.45% wet-lab success rate versus RFDiffusion’s sub-0.8% and the first-ever de novo carbohydrate binders. Worth highlighting one of their novel protein binders: Nipah virus attachment glycoprotein.
A nice take from Genes, Minds, Machines: protein language models are much worse at predicting the functional effects of mutations than the field generally claims, because of data leakage through improper train-test splitting. A naive baseline that just predicts the mean mutation effect at each site performs nearly as well as sophisticated pLMs with millions of parameters.
On biosecurity:
Congratulations to Abhishaike Mahajan on two years of fantastic biotech writing at Own Posting! His latest was a practically-book-length essay on biosecurity, based on conversations with around twenty experts. Lots of great points, enough that it inspired me to try to publish my own version of this in the coming weeks.
A Varsity Hackathon 2026 took place in London on March 7, pitting ~15 Oxford and Cambridge teams against each other in biotechnology and software challenges. Some great projects on AI biosecurity, such as ProteinRisk, a two-stage protein risk assessment tool that analyzes amino acid sequences for harmful structural motifs.
RAND Europe published a cost-benefit analysis of requiring DNA synthesis companies in the EU to screen synthetic nucleic acid orders for dangerous sequences. Very similar findings to this report published in December from the Center for Long Term Resilience looking at the UK specifically.
Biosecurity and public health:
Big news in US public health policy, which is that a federal judge has (temporarily) blocked the Trump administration’s changes to the childhood vaccine schedule, ruling that HHS Secretary Kennedy’s overhaul of the Advisory Committee on Immunization Practices (ACIP) violated federal law.
Former NIH program official Elizabeth Ginexi has been doing great writing on the defunding of the country’s health science research; her latest piece puts some numbers on this disaster and really helps appreciate the scale of what’s going on.

