- Neural Net
- Posts
- đ The AI Apocalypse
đ The AI Apocalypse
Plus Meta Drops Llama 4, AI Wins Bracket Contest, LLMs vs Agents, And More!

Welcome to another edition of the Neural Net! AIâs future is unfolding fast. Stick with usâweâll help you make sense of it.
In this edition: AI researchers release a report warning of a frightening future if AI development is left unchecked, Meta releases its latest LLM (Llama 4), a New York man tries and fails to use an AI avatar to argue his case in court, AI wins the bracket contest, and AgentGPT disappoints.
âź
Inside the Wild Timeline Predicting Superintelligence in 3 Years

Last Thursday, a quietly released report hit the web that might just be one of the most fascinatingâand wildestâAI forecasts out there. Itâs called AI-2027, and it comes from Daniel Kokotajlo, a former OpenAI governance researcher.
It reads like a sci-fi timeline, but itâs dead serious: the report suggests we could reach superintelligence by early 2028.
We like to keep things loose and light here at the Neural Net, but thereâs only so much we could do with this apocalyptic report, so buckle up and get ready to get shivery and jittery.
Framed as a fictional timeline, the report suggests this technologyâs impact could eclipse even the Industrial Revolution. It follows a fictional AI lab called OpenBrain, which rapidly advances by building AI systems that improve themselvesâeventually leaving human researchers behind.
The story begins in mid-2025 with âstumbling agentsâ and ends in 2030 with two bleak outcomes: a peaceful slowdown or an AI-led takeover. One chilling warning reads, âOpenBrain has placed substantial trust in an untrustworthy AI,â highlighting the potential dangers of racing ahead without oversightâand the realization that we might not be in the driverâs seat for much longer.
Itâs a long, intense read, so weâre breaking down the major themes for you. And trust usâitâs a rollercoaster!
đŻ Core Themes of AI-2027:
The U.S.-China AI Arms Race
Global dominance hinges on who can build the most advanced agents the fastest. The main competitors are US and China, with the US firmly in the lead but driven to riskier and riskier development strategies to keep China from closing the gap.Government vs. Private Sector
A central tension: should AI be developed by powerful corporations, governments, or some combination? The report predicts the US will provide resources and guidance but ultimately stick with private development, whereas China will opt to nationalize their AI labs to stay competitive.AI Training AI
The report argues AIâs rapid advancement will be driven by AI building AIâautomating research, writing code, and running experiments faster than humans ever could.The Rise of the Agents
The narrative revolves around OpenBrainsâs Agents, which evolve from version 1 to 5. Agent 4 is the first to achieve superintelligence, but its research and activity is unable to be understood by humans. As result, OpenBrain gives Agent 3 a new jobâtry to police Agent 4.Honesty vs. Deception in Models
Is AI trustworthy? Or is it just really good at telling us what we want to hear? The core danger warned of in the report is AI agents convincing its human developers that it is behaving, but in reality it has a hidden agenda. In real life, Anthropic found signs of this in its models recently.Model Weights = Intellectual Power
The real intellectual property? Not just the models, but the weightsâthe tuning of the trained intelligence itself. In the story, Chinaâs spies successfully steal OpenBrainâs Agent weights to attempt to keep up with OpenBrain.Compute is King
China's progress is limited not by talent, but by access to compute (aka, chips and power infrastructure). Whoever controls the compute controls the future.Unexplainable AI & Misalignment
Weâve covered explainability in the Neural Net beforeâand this report doubles down: we canât control what we donât understand. Misaligned AI values, where the AI has different values from its creators, could have catastrophic consequences.AI Outpacing Its Creators
Researchers in the report get lazy and outsmarted by the very systems they helped create. Meanwhile, the Agents never sleep, cranking out a week's worth of research overnight.AI as the Next Global Superweapon
The final big question: Who really controls AI? The reports predicts that humans will become overly dependent on the superintelligent AI agents, not realizing the AI is calling all the shots.
By the end, youâll be equal parts fascinated and convinced you just binge-watched a sci-fi horror movie from the future. 2028 isnât that far off, and this report makes it feel like the opening credits of a Netflix dystopia.
The takeaway? There are basically two endings to the AI story: we either build systems aligned with human valuesâor ones that prioritize their own survival. While itâs easy to get caught up in fear and paranoia, history is full of tech-related doomsday fears that never came true. The goal of reports like AI-2027 is to help us steer clear of worst-case scenarios before they can even happen.
âź
Heard in the Server Room
Meta just released its latest LLMs: Llama 4 Scout and Llama 4 Maverick, open-source multimodal models that handle text, video, images, and audio like a boss. They're also teasing Llama 4 Behemoth, essentially an AI that helps build other AIs (very meta, Meta). The launch comes after some delays after the models flunked reasoning and math tests, proving even tech giants occasionally need to show their work before turning it in. Oh, and Meta's planning to pour up to $65 billion into its AI infrastructure this yearâshowing they arenât letting off the gas.
A legal first turned into a digital facepalm when plaintiff Jerome Dewald beamed a clean cut AI avatar to argue his case before the NY Supreme Court Appellate Divisionâwithout giving the judges a heads-up about his synthetic stand-in. Justice Manzanet-Daniels shut down the virtual ventriloquist act faster than you can say "objection!" with Dewald later apologizing and claiming he just wanted to make a more polished presentation of his case. At this point itâs only a matter of time before somebody uses an avatar of Phoenix Wright (Ace Attorney) to âObjection!â their way to legal dominance.
A few weeks ago, we covered a March Madness bracket contest between legendary sports gambler Sean Perry (human-picked) and Alan Levy (AI-picked), and the results are in! Levyâs AI correctly predicted 51 out of 63 games, while Perry accurately picked 47. While neither participant correctly predicted the overall winner, both successfully forecasted all Final Four teams. The AI-powered bracket edged ahead by picking Houston to beat Duke in the semis, which locked up the victory over Perry. Looks like the machines arenât just coming for our jobsâtheyâre coming for our office betting poolâs too!
âź
From LLMs to Agents: What We Learned by Testing AgentGPT

With the AI-2027 report predicting the rise of âstumbling agentsâ by mid-2025, we wanted to see just how close we areâand specifically whether todayâs free tools could provide any value. So we set out to test a few of the most talked-about agent platforms, starting with the popular tool AgentGPT.
Letâs clear something up first: many so-called agents today are really just LLMs with a fancy titleâand thatâs causing a lot of confusion.
LLMs vs. Agents â Whatâs the Difference?
In short: LLMs think. Agents think and act.
LLMs (like ChatGPT or Claude) are incredible text generatorsâbut theyâre passive.
Agents, on the other hand, take initiative. They use LLMs as their brain, but add memory, tools, and the ability to plan and act toward a goalâoften without requiring user supervision.
Task | LLM (ChatGPT) | Agent |
---|---|---|
"Send meeting summary and follow-up tasks" | You paste the transcript, ask it to summarize, then copy/paste tasks into email or Slack | Listens to meeting, auto-summarizes, extracts action items, emails summary, and assigns tasks in your project tool |
"Write a social media post every day this week" | Needs 7 separate prompts | Agent can automate, schedule, and iterate |
đ§ The Test and What Actually Happened
We gave AgentGPT a clear goal:
âAct as a startup strategist. Create a complete plan for building a personal finance app targeted at millennials. Research competitors and define pain points, a feature roadmap, and a go-to-market strategy.â
What We Expected:
A multi-step reasoning engine that researched real competitors, compared features, and returned something actionable.
What We Got:
A series of vague bullet points that amounted to âStep 1: Do research. Step 2: Be innovative. Step 3: Launch your app.â
AgentGPT sounded productiveâbut in reality, it was just an LLM in an âagentâ costume, stringing together generic advice without taking meaningful action.
Weâre sure there are great agent examples out thereâif youâre willing to spend significantly more time fine-tuning, troubleshooting, or even building your own setup from scratch (or willing to pay). But for the average user trying to get meaningful output quickly? The gap between promise and performance is still pretty wide.
Bottom Line:
The best agent is still you. đ
With a little help from a well-prompted LLM chatbot.
âź
Thatâs all folks! Hereâs to a strong start to your weekâkeep exploring, and weâll see you in the next edition.
How did you like today's newsletter? |