- Jumble
- Posts
- o3 & o4 Mini Pack Some Serious Muscle
o3 & o4 Mini Pack Some Serious Muscle
Buckle up for a breakthrough week in AI: OpenAI’s new o‑series models put chain‑of‑thought superpowers (and every ChatGPT tool) in your pocket, while Google fires back with Gemini 2.5 Flash—its fastest “thinking” model yet. Let’s dive in. ⬇️
In today’s newsletter:
🚦 OpenAI drops o3 & o4 Mini
🧠 Google’s Gemini 2.5 Flash debuts
🔊 Anthropic teases Claude voice mode
⚖️ Judge rules Google’s ad tech is an illegal monopoly
🎥 AI Challenge: Make your first video with Veo 2
➗ What Sets o3 & o4 Mini Apart From the Rest of Industry?
OpenAI just rolled out the full version of o3 along with o4 Mini—two powerful reasoning models that are making us think, “Is this AGI"? Unlike GPT‑4o, which is optimized for speed and natural conversation, the o-series is built to stop, think, and solve.
These models don’t just answer—you can watch them reason out loud step by step, especially when complex logic, math, or tool use is involved. They're also the first ChatGPT models with full autonomous access to tools like Python, web browsing, and image generation—giving them the building blocks for light agent-style behavior.
🧠 Think before they speak
While GPT‑4o is better for fluid dialogue, summarization, and creative writing, o3 and o4 Mini are optimized for:
Solving logic-heavy problems
Executing code and debugging
Tackling STEM coursework
Navigating chain-of-thought prompts
Early tests show that the full version of o3 cuts down errors by 20% compared to o1, and o4 Mini has a stunning 99% pass rate on AIME-style math when paired with Python. These models are already becoming favorites in data analysis and academic research settings.
🛠️ Full tool access = lightweight agent-like behavior
Many sources highlight a key breakthrough with these models: o3 and o4 Mini can autonomously decide when to invoke a tool (Python, DALL·E, browser, etc.) and string together multi-step tasks. That means they can research, compute, and visualize without any manual prompting.
Ok, OpenAI o3 is insane.
It crops and zooms into the image
Find every little clues & runs quick searches
Then drops the exact map coordinates🤯
— Min Choi (@minchoi)
7:40 PM • Apr 17, 2025
This isn’t full autonomy—but it’s a major leap toward it. You can now drop a complex prompt into ChatGPT and let the model decide what tools to activate, when, and in what sequence.
📊 Poll: Have you tried o3 or o4 mini yet? |
⚡ Gemini 2.5 Flash Fast‑Think Counterpunch
Gemini 2.5 Flash just rolled out in preview via Google AI Studio and Vertex AI—and it’s Google’s most agile model yet. Touted as a “hybrid” foundation model, Flash gives developers control over how much the model "thinks." Need quick responses? Disable deep reasoning. Solving a logic problem? Crank it up. This on-demand performance tuning puts users in the driver’s seat like never before.
Excited to see the release of our Gemini Flash 2.5 model (the hard work of many people here!). This model continues the hallmark of the Flash model variants, with very high model quality with a low price point, and very low latency as well, continuing our push to go up (better
— Jeff Dean (@JeffDean)
10:08 AM • Apr 18, 2025
🏎️ Why it matters
Google’s framing is simple: sometimes, you don’t need a full-blown reasoning engine for every task. With Flash, you can set a token cap to prevent overthinking and reduce latency while maintaining high output quality. That means smarter prioritization in real-world scenarios—like building responsive AI interfaces, or running multiple agents in parallel.
Pricing-wise, Gemini Flash is optimized to deliver maximum value. Google claims it sits on the Pareto frontier for cost-efficient AI horsepower—punching far above its weight in performance-per-dollar. It’s also part of their push to bring more flexible AI to developers and startups that want both scale and speed.
🔍 Quick stack-up
When compared to o4 Mini, Flash isn't quite as sharp in code-heavy or math-intensive tasks, but it shines in speed, cost-efficiency, and integration across the Google ecosystem. For apps that need fast, responsive AI with a balance of control and capability, Gemini 2.5 Flash is shaping up to be a go-to choice.
Early testers are already using Flash to power AI chatbots, real-time decision systems, and embedded reasoning for lightweight applications—proof that fast and affordable doesn’t mean basic.
This Week’s Scoop 🍦
💬 AI-voice mode is finally coming to Claude’s chatbot
📸 OpenAI is not so quietly trying to compete with X
⚖️ Google’s ad empire faces it’s toughest legal ruling yet
📱 Perplexity to integrate with Samsung and Motorola phones
🏛️ US government considering a ban on DeepSeek
💰 OpenAI eyes $3B Windsurf acquisition
🎥 AI Challenge Make Your First Veo 2 Video
Challenge: Generate an 8‑second cinematic clip—no camera needed.
How to Access Veo 2
Gemini Advanced (or Google One AI Premium) → switch model dropdown to Veo 2.
Or head to Google AI Studio and pick “Generate video with Veo 2.”
Developers can hit the Gemini API or Vertex AI endpoint for programmatic control.
Sample Prompts to Try
“A timelapse of neon jellyfish drifting through a cyber‑punk city at night, cinematic lighting.”
“Grandma’s old Super 8 footage style: kids running through a sunflower field, grainy texture, warm tones.”
“An 8‑bit side‑scrolling spaceship dodging asteroids, retro arcade aesthetic.”
Push creative bounds, share your clips with us!
Want to sponsor Jumble?
Click below ⬇️

That’s it for this week! Which model has you more excited—OpenAI’s jump forward in reasoning and agents or Google’s lightning‑fast Flash? Hit reply with your take, and show off those Veo 2 videos. See you next time Jumble!
Stay informed, stay curious, and stay ahead with Jumble!
Zoe from Jumble