
đ ThursdAI â 3rd BirthdAI: Singularity Updates Begin with Auto Researcher, Uploaded Brains, OpenClaw Mania & NVIDIA's $26B Bet on Open Source
Hey, Alex here đ Today was a special episode, as ThursdAI turns 3 đ
Weâve been on air, weekly since Pi day, March 14th, 2023. I wonât go too nostalgic but Iâll just mention, back then GPT-4 just launched with 8K context window, could barely code, tool calls werenât a thing, it was expensive and slow, and yet we all felt it, itâs begun!
Fast forward to today, and this week, weâve covered Andrej Karpathyâs mini singularity moment with AutoResearcher, a whole fruit fly brain uploaded to a simulation, Chinaâs OpenClaw embrace with 1000 people lines to install the agent. I actually created a new corner on ThursdAI, called it Singularity updates, to cover the âout of distributionâ mind expanding things that are happening around AI (or are being enabled by AI)
Also this week, weâve had 3 interviews, Chris from Nvidia came to talk to use about Nemotron 3 super and NVIDIAâs 26B commitment to OpenSource, Dotta (anon) with his PaperClips agent orchestration project reached 20K Github starts in a single week and Matt who created /last30days research skill + a whole bunch of other AI news! Letâs dive in.
ThursdAI - Highest signal weekly AI news show is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.
Singularity updates - new segment
Andrej Karpathy open sources Mini Singularity with Auto Researcher (X)
If thereâs 1 highlight this week in the world of AI, itâs this. Andrej, who previously started the AutoPilot program in Tesla, and co-founded OpenAI, is now, out there, in the open, just.. doing stuff like invent a completely autonomous ML research agent.
Andrej posted to his almost 2M followers that he opensourced AutoResearch, a way to instruct a coding agent to do experiments against a specific task, test the hypothesis, discard whatâs not working and keep going in a loop, until.. forever basically. In his case, it was optimizing speed of training GPT-2. He went to sleep and woke up to 83 experiments being done, with 20 novel improvements that stack on top of each other to speed up the model training by 11%, reducing the training time from 2.02 hours to 1.8 hours.
The thing is, this code is already hand crafted, fine tuned and still, AI agents were able to discover new and novel ways to optimize this, running in a loop.
Folks, this is how the singularity starts, imagine that all major labs are now training their models in a recursive way, the models get better, and get better at training better models! Reminder, OpenAI chief scientist Jakub predicted back in October that OpenAI will have an AI capable of a junior level Research ability by September of this year, and it seems that... weâre moving quicker than that!
Practical uses of autoresearch
This technique is not just for ML tasks either, Shopify CEO Tobi got super excited about this concept, and just posted as Iâm writing this, that he set an Autoresearch loop on Liquid, Shopifyâs 20 year old templating engine, with the task to improve efficiency. His autoresearch loop was able to get a whopping 51% render time efficiency, without any regressions in the testing suite. This is just bonkers. This is a 20 year old, every day production used template. And some LLM running in a loop just made it 2x faster to render, just because Karpathy showed it the way.
Iâm absolutely blown away by this, this isnât a model release, like we usually cover on the pod, but still, a significant âunhobblingâ moment that is possible with the current coding agents and models. Expect everything to become very weird from here on out!
Simulated fruit fly brains - uploaded into a simulator
In another completely bonkers update that I can barely believe Iâm sending over, a company called EON SYSTEMS, posted that they have achieved a breakthrough in brain simulation, and were able to upload a whole fruit fly brain connectome, of 140K neurons and 50+ million synapses into a simulation environment.
They have... uploaded a fly, and are observing a 91% behavioural accuracy. I will write this again, they have uploaded a flyâs brain into a simulation for chirst sake!
This isnât just an âSF startupâ either, the board of advisors is stacked with folks like George Church from Harvard, father of modern genome sequencing, Stephen Wolfram who needs no introduction but one of the top mathematicians in the world, whosâ thesis is âbrains are programsâ, Anders Sandberg from Oxford, Stephen Larson who apparently already uploaded a worms brain and connected it to lego robots before. These folks are gung ho on making sure that at some point, human brains are going to be able to get uploaded, to survive the upcoming AI foom.
The main discussion points on X were around the fact that there was no machine learning here, no LLMs, no attention mechanisms, no training. The behaviors of that fly were all a result of uploading a full connectome of neurons. This positions connectome (the complete diagram of a brain with neurons and connections) as an ananalouge to an pre-trained LLM network for biological intelligence.
I encourage everyone whoâs reading this, to watch Pantheon on Netflix, to understand why this is of massive importance. Combined with the above Autoresearch, things are going to go very fast here. The next step is uploading a mouse brain, which will be a 500x Neurons and 2000x more synapses, but if weâre looking at the speed with which AI is improving, thatâs NOT out of the realm of possibility for the next few years!
OpenClaw Mania Sweeps China: Thousand-Person Lines & Government Subsidies, Grandmas raising a âred lobsterâ
Theyâre calling it âraising a red lobsterâ (ć »ć°éŸèŸ). Thatâs the phrase that swept Chinese social media for what is, at its core, installing an open source GitHub project on your laptop. Grandmas are doing it. Mac Minis are sold out. A cottage industry of paid installers popped up overnight on Xiaohongshu, charging up to $100 for an in-person setup. And yes, there are now also people charging to uninstall it.
On March 6th, roughly a thousand people lined up outside Tencentâs Shenzhen HQ for free OpenClaw installation. Appointment slots ran out within an hour. People brought NAS drives, MacBooks, mini PCs. Tencent engineers set up folding tables and just... started installing OpenClaw for strangers. I have pictures. Iâm not making this up.
All five major Chinese cloud providers jumped in simultaneously: Tencent Cloud, Alibaba Cloud, ByteDance Volcano Engine, JD.com Cloud, and Baidu Intelligent Cloud, each racing to offer one-click OpenClaw deployment. Why? Follow the money. Per HelloChinaTech, ByteDance, Alibaba, and Tencent spent roughly $60B combined on AI infrastructure. Chatbots donât burn enough tokens to justify that spend. But a single OpenClaw instance runs 24/7 and consumes 10-100x more tokens per day than a chatbot user. Every install is round-the-clock API revenue. The cheaper the models get, the more people run agents, the more infra gets sold. Self-reinforcing loop.
Local governments are pouring fuel on the fire. Shenzhenâs Longgang district is offering up to 2M yuan ($290K) per project. Hefei and Wuxi are going up to 10M yuan ($1.4M), plus free computing, office space, and accommodation for âone-person companies.â Meanwhile, Chinaâs central cybersecurity agency issued TWO warnings, banning banks and state agencies from installing OpenClaw. So local governments are subsidizing it while the central authority is trying to pump the brakes. Peak 2026.
With nearly half of all 142,000+ publicly tracked OpenClaw instances are now from China. OpenClaw is the most-starred GitHub repo in history, surpassing Linuxâs 30-year record in just 100 days. Device makers are piling on too â Xiaomi announced âmiclawâ for smartphones, MiniMax built MaxClaw, Moonshot AI built a hosted version around Kimi.
Now, Ryan was honest on the show and I want to echo that honesty here: OpenClaw is still hard to get working. There are many failure states. Itâs not âinstall and go to the beach.â Wolfram compared it to Linux in the late â90s â painful to set up, but if you push through, you can see the future behind the friction. This is real technology with real limitations, and a lot of disappointed folks in China are watching tokens burn with no actual work getting done.
But hereâs the thing I keep coming back to. The memetic velocity of OpenClaw is unlike anything Iâve seen in tech. Itâs not just a tool, itâs a concept that penetrated the cultural resistance to AI. People who are scared of terminals, people whoâve never touched GitHub â theyâre standing in line for this. I broke through that resistance with my own fiancĂ©e. Sheâs now running two OpenClaws. Not enough for her. She needs another one.
Every major US lab is watching this closely. OpenAI brought Peter Steinberger on staff. Perplexity just announced theyâre building a local agent for Mac. Anthropic has Claude Cowork. This is where all of computing is headed â always-on, autonomous, personal AI that actually does things for you. OpenClaw is the first front door, not the final destination. But what a front door it is.
ThursdAI - Highest signal weekly AI news show is a reader-supported publication. To receive new posts and support my work, consider becoming a free or paid subscriber.
Open Source: Nvidia Goes All In with Nemotron 3 Super 120B (X, Blog, HF)
We had Chris Alexiuk from Nvidia join us â a friend from a dinner Nisten and I hosted in Toronto. Chris is basically âNeMoâ embodied, sitting at the intersection of product and research, and he gave us the full breakdown on what might be the most complete open-source model release weâve seen from a major lab.
Here are the numbers: 120B total parameters, 12B active during inference (itâs a Mixture of Experts), 1 million token context window, and a hybrid Mamba-Transformer architecture they call âHybrid Mamba Latent MoE with Multi-Token Prediction.â Itâs hitting 450 tokens per second on the Terminal Bench leaderboard â faster than any other model on there. Modal is reporting over 50% faster token generation compared to other top open models.
What Chris was emphatic about â and I want to highlight this â is that âmost openâ is a real designation here. They released the model checkpoint in three precisions (BF16, FP8, NVFP4), the base checkpoint before post-training, the SFT training data, and in a move that genuinely surprised people, pre-training data and a full end-to-end training recipe. You can, in theory, reproduce their training run. Thatâs rare. Thatâs a real commitment to open source.
Thereâs also a huge piece of news in the background here: thereâs a confirmed report that Nvidia will spend $26 billion over the next five years building the worldâs best open source models. Jensen presumably has GTC remarks incoming on this. America is genuinely back in the open source AI race, and itâs Nvidia leading the charge. Chris has been in the open source world since the Hugging Face early days and said it feels genuine inside the company â not a PR exercise. And I tend to believe him. Now, all eyes are on GTC next week!
I ran Nemotron 3 Super with my own OpenClaw instance yesterday via W&B inference and itâs genuinely fast and capable. At $0.20/M input tokens and $0.80/M output tokens on W&B inference, itâs not going to replace Opus for your hardest tasks â but for running an always-on agent that needs to be cost-efficient? Itâs an incredible option. More on that in the this weeks buzz section below.
Tools & Agentic Engineering
Paperclip: Zero Human Companies, Now Open Source (Github)
We had the anonymous Dotta on the show â the first AI video avatar anon person to join ThursdAI â to talk about Paperclip, an open source agent orchestration framework that hit 20,000 GitHub stars in its first week. The premise is simple and audacious: build zero-human companies.
Now this may sound familiar to you, as we had Ben from Polsia on just two weeks ago, which is a similar concept, but Paperclip is an OpenSource project, which you can run right now on your own.
The core âthingâ that got me excited about Paperclip is that you can âhireâ your own existing OpenClaw agents, or Cursor or Codex or whatever else to play roles in this autonomous company. The premise is simple, youâre the board of directors, you hire an AI Agent CEO, and it then asks you if it needs to âhireâ more AI agents to do tasks autonomously. These tasks all live inside Paperclip interface, and you or your Agents can open them.
The core concept of this whole system is the heartbeat concept, each agent receives their own instructions on what to do every time they are âwoken upâ by a timer. This is whatâs driving the âautonomousâ part of the whole thing, but itâs also whatâs eating the tokens up, even if thereâs no work being done, agents are still burning tokens asking âis there work to be done?â
Dotta gave us a great metaphor, asking if we saw the movie Memento, where the protagonist lost his memory and every time he woke up, he woke up with a blank slate, and had to reconstruct the memories. AI agents are like the memento man, and Paperclip is an attempt to give those agents the whole context so they can continue working on your tasks productively. Dotta told us that the future of Paperclip is the ability to âforkâ entire companies, structures that will actually run and do things on your behalf. Looking forward to that future, but for now I will be turning off my Paperclip interface as itâs costing me real money without the need.
Symphony: Agents Writing Their Own Jira Tickets
We mentioned Symphony last week, and I texted Ryan the link before the show, and voila, of course, he set it up and went viral, yet again! Weâre so lucky to have Ryan on the show to tell us from first hand experience what itâs like to run this thing.
Symphony was open sourced by OpenAI last week, and itâs basically an instruction manual for how to run agents autonomously via Linear ticketing system. (Github)
The highlight for Ryan was, the whole system is running creating pull requests while heâs a sleep, and at some point, he noticed a ticket that he didnât create. One of the agents found a bug, and created a very detailed ticket for him to approve.
Iâm just happy that I can keep even my co-hosts up to date hehe
This weeks buzz - weâve got skills and nemotrons!
Look, we told you about Skills in the start of the year, since then, via OpenClaw, Hermes Agent, Claude Code, they exploded in popularity. One downside of skills is, itâs very easy to make a bad one! So, weâre answering the challenge, and are publishing the official wandb skill đ
Installing it is super simple, npx skills add wandb/skills and voila, your agents are now officially âI know kung fuâ pilled with the best Weights & Biases practices. For both Weave and Models đ Please give us feedback on Github if you have used the skills! Github
Also, weâve partnered with Nvidia to support the best US open source model on day 0, and we have Nemotron 3 Super on our inference service, for all to use at $0.20/1Mtok! Itâs super easy to setup with something like Hermes Agent or OpenClaw and runs really really fast! Check it out here.
Is it going to perform like Opus 4.6? No. But are you going to run Opus 4.6 at 20 cents per million? Also no.
Gemini drops SOTA embeddings and gets dethroned 2 days later live on the show.
This always happens, but I didnât expect this to happen in a fairly niche segment of the AI world... multimodal embeddings!
Gemini posted an update earlier this week with Gemini Embeddings 2.0, a way to unify images, text, video, audio embeddings under 1 roof, and posted a SOTA embedding model!
Then, just as we launch the show, a friend of the pod Benjamin Clavie, drops me a DM, basically saying that his company Mixbread is going to deploy an embedding model that will beat Gemini Embedding 2 on almost every benchmark on that table, and then... they did!
The most notable (and absolutely crazy) jump in this comparison is, the LIMIT benchmark, where they achieved a 98% score vs Geminiâs ... 6.9 percent. I didnât believe this at first, but asked Ben to explain the findings, and he did. Congrats to folks for moving the search space forward every 2 days!
Grok 4.20 in the API for $2/1Mtok
Elon Musk and XAI co finally released Grok 4.20 in the API. Look I said what I said about XAI models, they are great for research, and for factuality, but they arenât beating the major labs. The last firing of almost of XAI folks doesnât help either. So this model was not âreleasedâ in any traditional sense, thereâs no benchmarks, no evals, and everyone who got access to it evaluated it and, itâs no better than GLM5 on many benchmarks. So it does makes sense to release it quietly.
It is very fast though, and again, for research and for X access, itâs an absolute beast, so Iâll be trying this out!
Parting thoughts and a small reflection. For the past 3 years, weâve had a front-row seat to the singularity shaping up. 2.5 years ago, I went all in, decided to pivot into podcasting full time. In those years, ThursdAI became known, weâve had guests from nearly all major AI labs (including Chinese ones, for which Iâm particularly proud), I got to meet with executives, ask leaders questions about where this is all going, and most of all, share this journey with all of you, candidly. We rarely do hype on the show, we donât speculate, we try to do a positive outlook on the whole thing, and counter doomerism, as thereâs too much of that out there.
I am very glad this resonates, and continue to be thankful for your attention! If you wanted to give us any kind of a birthday present, subscribe or give us a 5 star review on Apple Podcasts or Spotify, itâll greatly help other folks to discover us.
See you next week,
Alex đ«Ą
ThursdAI - Mar 12, 2026 - TL;DR
* Hosts and Guests
* Alex Volkov - AI Evangelist & Weights & Biases (@altryne)
* Co Hosts - @WolframRvnwlf @yampeleg @nisten @ldjconfirmed) @ryancarson
* Chris Alexiuk from Nvidia (@llm_wizard) - Nemotron
* @dotta - creator of Paperclip.ing AI agent orchestration framework
* Matt Van Horn - @mvanhorn - creator of @slashlast30days
* Singularity updates
* Andrej Karpathyâs autoresearch achieves 11% speedup on GPT-2 training through autonomous AI agent experimentation (X, GitHub, GitHub)
* Eon Systems uploads first complete fruit fly brain to a physics-simulated body, achieving 91% behavioral accuracy (X, Announcement, Announcement)
* OpenClaw mania sweeps China as all five major cloud providers race to support it (HelloChinaTech, Reuters, SCMP, MIT Tech Review)
* Big CO LLMs + APIs
* xAI quietly releases Grok 4.20 API with massive 2M token context window and multi-agent capabilities (X, Blog)
* Google launches Gemini Embedding 2, the first natively multimodal embedding model supporting text, images, video, audio, and PDFs in a unified vector space (X, Announcement)
* Open Source LLMs
* NVIDIA launches Nemotron 3 Super: 120B open MoE model with 1M context window designed for agentic AI at 5x higher throughput (X, Announcement)
* MiroMind releases MiroThinker-1.7 and H1 - open-source research agents with 256K context, 300 tool calls, achieving SOTA on deep research benchmarks (X, HF, HF, HF)
* Covenant-72B: Worldâs largest permissionless decentralized LLM pre-training achieves 72B parameters on Bittensor with 146x gradient compression (X, Arxiv, HF, HF)
* Tools & Agentic Engineering
* ACP is the open standard that lets any AI coding agent plug into any editor â and this week Cursor officially joined the registry, meaning you can now run Cursorâs agent inside JetBrains IDEs (JetBrains blog, Cursor blog)
* This weeks Buzz
* W&B launches official Agent Skills for coding agents, turning experiment dashboards into terminal queries (X, Announcement, Announcement)
* Video
* LTX-2.3 â Lightricks open-source video model (GitHub, HF,Blog)
* Voice & Audio
* Fish Audio launches S2: Open-source TTS with sub-150ms latency and absurdly controllable emotion (X, HF, Blog, Announcement)
* Show notes and links
* Paperclip.ing by Dotta ( @dotta) - Github
* Last30days skill by Matt Van Horn Github
* Agency Agents repo Github
* OpenAI Symphony (Github)
* Mixbread Embeddings (X)
This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit sub.thursdai.news/subscribe
Fler avsnitt frÄn "ThursdAI - The top AI news from the past week"



Missa inte ett avsnitt av âThursdAI - The top AI news from the past weekâ och prenumerera pĂ„ det i GetPodcast-appen.








