
ThursdAI - Oct 30 - From ASI in a Decade to Home Humanoids: MiniMax M2's Speed Demon, OpenAI's Bold Roadmap, and 2026 Robot Revolution
ThursdAI - The top AI news from the past week
Minimax Video and Speech Updates
Alex asks about video and audio releases; Skyler summarizes HiLuo video realism and speech latency and quality improvements.
Hey, itâs Alex! Happy Halloween friends!
Iâm excited to bring you this weeks (spooky) AI updates! We started the show today with MiniMax M2, the currently top Open Source LLM, with an interview with their head of eng, Skyler Miao, continued to dive into OpenAIs completed restructuring into a non-profit and a PBC, including a deep dive into a live stream Sam Altman had, with a ton of spicy details, and finally chatted with Arjun Desai from Cartesia, following a release of Sonic 3, a sub 49ms voice model!
So, 2 interviews + tons of news, letâs dive in! (as always, show notes in the end)
Hey, if you like this content, it would mean a lot if you subscribe as a paid subscriber.
Open Source AI
MiniMax M2: open-source agentic model at 8% of Claudeâs price, 2Ă speed (X, Hugging Face )
We kicked off our open-source segment with a banger of an announcement and a special guest. The new king of open-source LLMs is here, and itâs called MiniMax M2. We were lucky enough to have Skyler Miao, Head of Engineering at Minimax, join us live to break it all down.
M2 is an agentic model built for code and complex workflows, and its performance is just staggering. Itâs already ranked in the top 5 globally on the Artificial Analysis benchmark, right behind giants like OpenAI and Anthropic. But hereâs the crazy part: it delivers nearly twice the speed of Claude 3.5 Sonnet at just 8% of the price. This is basically Sonnet-level performance, at home, in open source.
Skylar explained that their team saw an âimpossible triangleâ in the market between performance, cost, and speedâyou could only ever get two. Their goal with M2 was to build a model that could solve this, and they absolutely nailed it. Itâs a 200B parameter Mixture-of-Experts (MoE) model, but with only 10B active parameters per inference, making it incredibly efficient.
One key insight Skylar shared was about getting the best performance. M2 supports multiple APIs, but to really unlock its reasoning power, you need to use an API that passes the modelâs âthinkingâ tokens back to it on the next turn, like the Anthropic API. Many open-source tools donât support this yet, so itâs something to watch out for.
Huge congrats to the MiniMax team on this Open Weights (MIT licensed) release, you can find the model on HF!
MiniMax had quite a week, with 3 additional releases, MiniMax speech 2.6, an update to their video model Hailuo 2.3 and just after the show, they released a music 2.0 model as well! Congrats on the shipping folks!
OpenAI drops gpt-oss-safeguard - first open-weight safety reasoning models for classification ( X, HF )
OpenAI is back on the open weights bandwagon, with a finetune release of their previously open weighted gpt-oss models, with gpt-oss-safeguard.
These models were trained exclusively to help companies build safeguarding policies to make sure their apps remains safe! With gpt-oss-safeguards 20B and 120B, OpenAI is achieving near parity with their internal safety models, and as Nisten said on the show, if anyone knows about censorship and safety, itâs OpenAI!
The highlight of this release is, unlike traditional pre-trained classifiers, these models allow for updates to policy via natural language!
These models will be great for businesses that want to safeguard their products in production, and I will advocate to bring these models to W&B Inference soon!
A Humanoid Robot in Your Home by 2026? 1X NEO announcement ( X, Order page, Keynote )
Things got really spooky when we started talking about robotics. The company 1X, which has been on our radar for a while, officially launched pre-orders for NEO, the worldâs first consumer humanoid robot designed for your home. And yes, you can order one right now for $20,000, with deliveries expected in early 2026.
The internet went crazy over this announcement, with folks posting receipts of getting one, other folks stoking the uncanny valley fears that Sci-fi has built into many people over the years, of the Robot uprising and talking about the privacy concerns of having a human tele-operate this Robot in your house to do chores.
It can handle chores like cleaning and laundry, and for more complex tasks that it hasnât learned yet, it uses a teleoperation system where a human â1X Expertâ can pilot the robot remotely to perform the task. This is how it collects the data to learn to do these tasks autonomously in your specific home environment.
The whole release is very interesting, from the âsoft and quietâ approach 1X is taking, making their robot a 66lbs short king, draped in a knit sweater, to the $20K price point (effectively at loss given how much just the hands cost), the teleoperated by humans addition, to make sure the Robot learns about your unique house layout.
The conversation on the show was fascinating. We talked about all the potential use cases, from having it water your plants and look after your pets while youâre on vacation to providing remote assistance for elderly relatives. Of course, there are real privacy concerns with having a telepresence device in your home, but 1X says these sessions are scheduled by you and have strict no-go zones.
Hereâs my prediction: by next Halloween, weâll see videos of these NEO robots dressed up in costumes, helping out at parties. The future is officially here. Will you be getting one? If not this one, when will you think youâll get one?
OpenAIâs Grand Plan: From Recapitalization to ASI
This was by far the biggest update about the world of AI for me this week! Sam Altman was joined by Jakub Pachocki, chief scientist and Wojciech Zaremba, a co-founder, on a live stream to share an update about their corporate structure, plans for the future, and ASI goals (Artificial Superintelligence)
First, the company now has a new structure: a non-profit OpenAI Foundation governs the for-profit OpenAI Group. The foundation starts with about 26% equity and has a mission to use AI for public good, including an initial $25 billion commitment to curing diseases and building an âAI Resilienceâ ecosystem.
But the real bombshells were about their research timeline. Chief Scientist Jakub Pachocki stated that they believe deep learning systems are less than a decade away from superintelligence (ASI). He said that at this point, AGI isnât even the right goal anymore. To get there, theyâre planning to have an âAI research internâ by September 2026 and a fully autonomous AI researcher comparable to their human experts by March 2028. This is insane if you think about it. As Yam mentioned, OpenAI is already shipping at an insane speed, releasing Models and Products, Sora, Atlas, Pulse, ChatGPT app store, and this is with humans, assisted by AI.
And here, they are talking about complete and fully autonomous researchers, that will be infinitely more scalable than humans, in the next 2 years. The outcomes of this are hard to imagine and are honestly mindblowing.
To power all this innovation, Sam revealed they have over $1.4 trillion in obligations for compute (over 30 GW). And said even thatâs not enough. Their aspiration is to build a âcompute factoryâ capable of standing up one gigawatt of new compute per week, and he hinted they may need to ârethink their robotics strategyâ to build the data centers fast enough. Does this mean OpenAI humanoid robots building factories? đ€
Plus, donât forget, Sam is one of the investors in Helion energy, working on power solutions like Fusion, and the above graphic has an Energy block that Sam said they will give an update on later (thatâs also what he told me during Dev Day when I asked him about it).
Super exciting and honestly mind-blowing stuff, Gigawats per week, fully autonomous researchers, the world is going to look way different in a few years!
The Agent Labs Race: Cursor 2.0 vs. Cognitionâs SWE-1.5 (X, Blog)
This week also saw a major showdown in the agentic coding space. On the very same day, both Cursor and Cognition launched major updates and their own new models, signaling a new era where agent labs are training their own specialized AI.
First up, Cursor 2.0 was released with a completely redesigned multi-agent interface and their new model, Composer. Composer is claimed to be four times faster than comparable models, and the new UI is built around managing a fleet of agents that can work in parallel on your codebase. Itâs a clear shift from being just an IDE to a full-fledged agent platform. Look, the UI even looks like ChatGPT and no code in sight (until you switch to IDE mode)
Their Composer model is also very interesting, and got a lot of folks excited, but the evaluations they shared, and the fact that they didnât disclose if thatâs a finetune of a chinese model (it likely is). Regardless, folks are saying that itâs a very good model thatâs also VERY fast!
Cognition own coding model - SWE 1.5 ( Blog, X, Windsurf )
Then, just hours later, Cognition punched right back with SWE-1.5, their new frontier agent model that now powers Windsurf. The headline here is pure speed. Powered by Cerebras, SWE-1.5 hits a blistering 950 tokens per secondâ13 times faster than Sonnet 4.5âwhile achieving near-SOTA performance on SWE-Bench Pro. Theyâve achieved this through a co-designed stack where the agent harness, inference system, and model were all built together and optimized with end-to-end reinforcement learning in real coding environments.
This competition is fantastic news for all of us. Weâre seeing specialized, highly-performant models being developed outside of the big labs, putting more power back in the hands of developers.
This Weekâs Buzz
Just a few quick updates from the world of Weights & Biases and our parent company, CoreWeave.
First, big news! CoreWeave announced the acquisition of Marimo, the company behind the popular open-source, reactive notebook for Python. This is another exciting step in building out the essential cloud for AI, adding powerful development tools to the stack alongside best-in-class GPU infrastructure and MLOps with Weights & Biases. Welcome to the Marimo team!
Also, Fully Connected is coming to London next week! Itâs our premier conference, and weâll have speakers from Mistral, Google, LlamaIndex, and more. If youâre in Europe, please come join us. DM me if you need tickets!
And if youâre in New York from November 19-22, come say hi at the AI Engineer Code Summit. Weâre sponsoring and will have a big booth. Itâs always a great place to meet folks from this community.
Video & Voice: The Multimodal Explosion
The world of video and voice AI was on fire this week.
The absolute highlight was Odyssey ML V2, a new real-time interactive AI video platform. This thing is not like other video models that take minutes to generate a clip. With Odyssey, you type a prompt, and a video starts streaming instantly. Then, you can edit it live. We did a demo on the show where we prompted âarmy of robots in a starship corridorâ and then typed âturn these robots into fluffy covered cat robots,â and the video changed in real time. Itâs mind-blowing. This is a glimpse into the future of user-driven, playable media.
On the more traditional video front, Sora is now invite-free in the US and Japan, and they launched Character Cameos. You can now upload photos of your pets or objects (like your kidâs carved pumpkin!) and turn them into characters that you and others can use in videos. I, of course, immediately made a cameo of my cat, Sonia.
Voice and Audio - Cartesia launches Sonic 3, sub 50ms AI speech model
In the world of voice, we had Arjun Desai from Cartesia join us to talk about Sonic-3, their new real-time TTS engine. Backed by a new $100M funding round, Sonic-3 is built on State Space Models (not Transformers) and can achieve insane speedsâweâre talking under 50ms latency. But itâs not just fast; itâs also incredibly expressive. It can laugh, emote, and speak 42 languages with natural code-switching. I used their Pro Voice cloning feature to create an AI version of myself, and the results were scarily good. We even had my AI clone host a segment of the show, see it yourself here, powered by Argil and Sonic 3, this is... AI Alex
Wrapping Up This Spooky Week đ
As I sit here in my Halloween costume reflecting on this week, I canât help but feel weâre at an inflection point. We have:
* Open source models competing with the best proprietary ones
* Humanoid robots becoming consumer products
* ASI timelines measured in single-digit years
* Real-time interactive AI across all modalities
And yet, nothing about this scares me. If anything, Iâm more excited than ever about what weâre building together. Yes, the pace is insane. Yes, keeping up with everything is becoming nearly impossible (and itâs literally my job!). But weâre living through the most transformative period in human history, and we get to be part of it.
To everyone building, experimenting, and pushing boundaries - keep going. To everyone worried about whatâs coming - join us in shaping it responsibly. And to everyone who celebrated Halloween today - I hope your costume was as epic as the AI developments we covered! đ»
Until next week, this is Alex signing off. Remember to subscribe, give us five stars, and Iâll see you next ThursdAI!
TL;DR - All Topics Covered
ThursdAI - Oct 30 - Halloween Special đ»
* Hosts and Guests
* Alex Volkov - AI Evangelist & Weights & Biases (@altryne)
* Co Hosts - @WolframRvnwlf @yampeleg @nisten @ldjconfirmed @ryancarson
* Guest: Skyler Miao - Head of Engineering, MiniMax (@SkylerMiao7)
* Guest: Arjun Desai - CoFounder Cartesia (@jundesai)
* Open Source LLMs
* MiniMax M2: Open-source agentic model at 8% of Claudeâs price, 2Ă speed (X, Hugging Face)
* OpenAI GPT-OSS-Safeguard: First open-weight safety reasoning models (X, HF)
* IBM Granite 4.0 Nano: Ultra-efficient tiny models for edge deployment (X, Artificial Analysis)
* Ming-flash-omni Preview: Sparse MoE omni-modal model (X, HuggingFace)
* Kimi Linear: 48B parameter model with 1M context (HF)
* Robotics
* 1X NEO: First consumer humanoid robot, $20k, delivery 2026 (X, Order page, Keynote)
* Big Companies & APIs
* OpenAI Restructuring: ASI within 10 years, AI researcher by 2028 (X)
* Cursor 2.0 & Composer: 4x faster coding, new model (X, Blog)
* Cognition SWE-1.5: 950 tok/s, 40% SWE-bench Pro (Blog, X, Windsurf)
* Perplexity Email Assistant: Privacy-first AI inbox management (X, Assistant Site)
* This Weekâs Buzz
* Fully Connected London - fullyconnected.com
* AI Engineer Code Summit NYC - Nov 19-22
* CoreWeave acquires Marmo notebooks (X)
* Vision & Video
* Odyssey ML V2: Real-time interactive AI video (X, Experience)
* Sora: Now invite-free + Character Cameos feature (X, Sonia Cameo)
* Hailuo 2.3: Cinema-grade video generation (X)
* Voice & Audio
* MiniMax Speech 2.6: <250ms ultra-human voice AI (X, MiniMax, API Docs)
* Cartesia Sonic 3: Real-time TTS with emotion & laughter, $100M funding (X, Website, Docs)
* Tools
* Pokee: Agentic workflow builder (X)
* Pomelli: Googleâs AI marketing agent (X, Labs)
This is a public episode. If you'd like to discuss this with other subscribers or get access to bonus episodes, visit sub.thursdai.news/subscribe


