Nano Banana Pro 4K image model – 8% text errors, $0.13 renders

Executive Summary

Google quietly turned on Nano Banana Pro, its Gemini 3 Pro Image model, across Gemini web/app, AI Mode in Search, Flow, NotebookLM, AI Studio, and Vertex. You get 1K/2K/4K outputs, a 1M‑token context inherited from Gemini 3 Pro, and pricing around $0.134 per generated image on top of $2/M input and $12/M output tokens. The pitch: a reasoning‑aware image engine that can lay out multilingual text and infographics without looking like your UI was typeset by a blender.

Early benchmarks back that up. Nano Banana Pro tops GPT‑Image 1, Seedream v4 4K, and Flux Pro Kontext Max on text‑to‑image and editing Elo, with ~100‑point leads in several edit categories. A heatmap puts single‑line text errors near 8% versus ~38% for GPT‑Image 1 across languages, including Arabic, Hindi, and Hebrew. The new “Show thinking (Nano Banana Pro)” toggle also hints that chain‑of‑thought is now a thing for pixels, not just prose.

Third‑party support lit up immediately: fal.ai shipped day‑0 text‑to‑image and edit APIs, Higgsfield is dangling “unlimited 4K” Nano Banana Pro with up to 65% off, and Genspark wired it into its all‑in‑one workspace. If your product depends on legible dashboards, posters, or UI mocks, this model is worth a focused weekend of A/B tests.

Feature Spotlight

Feature: Nano Banana Pro (Gemini 3 Pro Image) goes live across Google and partners

Google’s reasoning image model ships broadly (Gemini web/app, AI Studio, Vertex) with 4K, better text, and multi‑turn edits—immediately consumable via Google surfaces and partner APIs, accelerating creative and product workflows.

Cross‑account confirmations that Google’s reasoning image model is now broadly usable: Gemini web/app, AI Studio, Vertex docs, and third‑party endpoints. Focus is on 4K output, accurate multilingual text, editing controls.

Jump to Feature: Nano Banana Pro (Gemini 3 Pro Image) goes live across Google and partners topics

Table of Contents

Stay in the loop

Get the Daily AI Primer delivered straight to your inbox. One email per day, unsubscribe anytime.

Feature: Nano Banana Pro (Gemini 3 Pro Image) goes live across Google and partners

Cross‑account confirmations that Google’s reasoning image model is now broadly usable: Gemini web/app, AI Studio, Vertex docs, and third‑party endpoints. Focus is on 4K output, accurate multilingual text, editing controls.

Nano Banana Pro (Gemini 3 Pro Image) launches across Gemini and Google tools

Google DeepMind has formally launched Nano Banana Pro (Gemini 3 Pro Image), a reasoning‑aware image generation and editing model now available in the Gemini app and web, AI Mode in Search, Flow, NotebookLM and more, following earlier leaked tests of 4K outputs and text rendering 4k leak. The model supports 1K/2K/4K output, lighting and camera controls, aspect‑ratio changes, multilingual text rendering, multi‑image compositing, and Search grounding for fact‑aware visuals DeepMind feature thread surfaces and getting started.

For developers, Nano Banana Pro appears in Google AI Studio as gemini-3-pro-image-preview with pricing of $2.00/M input tokens, $12.00/M output tokens, and about $0.134 per generated image, sharing the same Jan 2025 knowledge cutoff and 1M context as Gemini 3 Pro text ai studio pricing. Vertex AI’s model garden lists the same model ID (publishers/google/models/gemini-3-pro-image-preview), and emphasizes "reasoning for image generation", 4K support, and optional grounding with Search for more factual images vertex docs overview.

On the front end, Gemini web now exposes a "Show thinking (Nano Banana Pro)" toggle on image generations, hinting that chain‑of‑thought style internal reasoning is being applied even for visuals and optionally made visible to users show thinking ui. Community posts confirm successful runs in Gemini web (“cat shooting power into a wormhole”, “minion with nano banana”) and in the mobile app, aligning with AILeaks’ note that the model is rolling out broadly across Google’s AI offerings gemini app release rollout confirmation.

Benchmarks put Nano Banana Pro ahead of GPT‑Image 1 in quality and text rendering

Early benchmark charts circulating today show Gemini 3 Pro Image (Nano Banana Pro) leading competing models like GPT‑Image 1, Seedream v4 4K and Flux Pro Kontext Max on both text‑to‑image and image‑editing Elo scores, while also drastically cutting text rendering errors image elo scores editing elo chart. In text‑to‑image, Gemini 3 Pro Image tops overall preference and visual quality, and in editing it wins across object/environment edits, style transfer, single and multi‑character consistency, and text editing, with Elo gaps of ~100 points or more over GPT‑Image 1 in several categories editing elo chart.

A separate heatmap on single‑line text rendering shows Nano Banana Pro at roughly 8% average error across languages versus GPT‑Image 1 at about 38%, with particularly stark differences in scripts like Arabic, Hindi and Hebrew text error heatmap. Sample outputs like a 4K "How solar power works" infographic, complete with clean typography and structured diagram elements, illustrate the practical impact for product teams that care about legible UI mocks and dashboards rather than just concept art solar infographic sample. DeepMind and third‑party commentary are already framing Nano Banana Pro as a "reasoning image model"—the idea being that its improved factual grounding and layout reasoning, not just its decoder quality, are what push these numbers above older models reasoning model quote.

fal.ai ships day‑0 Nano Banana Pro text‑to‑image and editing APIs

Inference host fal.ai has Nano Banana Pro live on day zero, exposing separate text‑to‑image and image‑editing endpoints so teams can start wiring Gemini 3 Pro Image into their own apps without waiting on Google’s stacks fal launch. The text‑to‑image model page quotes per‑image pricing and supports the full creative prompt flow, while a dedicated edit endpoint lets you upload an image and drive semantic edits with natural language rather than masks fal text to image fal image editing.

Because fal handles queuing and scaling, this gives smaller shops an immediate way to prototype Nano Banana‑backed features—marketing image pipelines, design tools, product configurators—without standing up Vertex or dealing with quota friction in AI Studio fal endpoints thread. For AI engineers already using fal for other diffusion or video models, swapping in Nano Banana Pro should mostly be a matter of changing the model slug and adjusting prompts for its stronger text and layout handling.

Higgsfield offers unlimited 4K Nano Banana Pro with aggressive Black Friday deal

Video‑first platform Higgsfield is pitching an unusually aggressive commercial package around Nano Banana Pro: unlimited 4K generations for a full year at up to 65% off as part of a Black Friday sale, plus a short‑term promo of 350 free credits for retweets and comments higgsfield launch promo. They emphasize Gemini‑powered reasoning, multilingual text quality and precise edit controls, positioning their offer as the way to tap the model without worrying about per‑image caps or token budgeting higgsfield feature recap.

The Higgsfield landing page reinforces this framing by advertising "Unlimited 4K Nano Banana Pro" plans and showing Hollywood‑style VFX clips and plushie conversions as example outputs, which target agencies and heavy social content shops more than hobbyists Higgsfield site. For AI leads, the main trade‑off versus going direct to Google is clear: pay a fixed platform price with Higgsfield’s workflow UI on top, or stay closer to the metal with Vertex/AI Studio and handle quotas and orchestration yourself.

Genspark integrates Nano Banana Pro into its all‑in‑one AI workspace

Genspark, the "everything in one" AI workspace, now exposes Google’s Nano Banana image model alongside its existing stack of text and video tools, so users can generate and edit images without leaving the same canvas they use for slides, docs and data analysis genspark overview. In a walkthrough, the team shows Nano Banana‑driven poster design and image editing flows happening directly inside Genspark, with no separate API setup or platform switching nano banana in genspark.

Because Genspark routes to multiple frontier models under one subscription and offers an AI Inbox, shared projects and agent flows, this integration means design work (posters, promo graphics, visual assets for decks) can sit in the same project as the copy, research, and analysis that feeds it genspark video models. Their Black Friday‑era promo also includes free compute credits for new signups, making it easy for teams to benchmark Nano Banana Pro’s text and layout handling against whatever image stack they already use, before deciding where to standardize genspark pricing promo.


Open weights: Olmo 3 (7B/32B base, Instruct, Think)

A major fully open model family drop from Ai2 with training data, checkpoints, logs and a detailed report—useful for teams needing transparent, reproducible stacks. Excludes Google’s image model (covered as the feature).

Ai2 releases fully open Olmo 3 7B/32B family

Allen Institute for AI has released the Olmo 3 family: 7B and 32B base models plus Instruct, Think, and RL Zero variants, all with open weights, pretraining data mix, post‑training datasets, intermediate checkpoints, and training logs. They explicitly position Olmo 3‑Base 32B as the strongest fully open 32B base and the 7B variants as top "Western" thinking/instruct models, aiming to give teams a transparent alternative to Qwen‑scale systems. release thread The full suite is live on Hugging Face and documented in a detailed technical report, so you can inspect or reproduce every stage of the model flow from SFT through DPO and RLVR. huggingface collection paper pdf ai2 blog

Olmo 3 ships 7B RL Zero datasets and checkpoints for math, code and instructions

Beyond the base and Think models, Ai2 is releasing four RL Zero tracks for the 7B Olmo 3—focused separately on math, code, instruction following, and a mixed blend—each with open datasets and intermediate checkpoints. release thread The team explicitly frames this as a lab bench for studying how starting RL directly from the base model (inspired by DeepSeek R1) interacts with mid‑training reasoning traces and how much benchmark lift comes from RL versus SFT or DPO, in areas where Qwen‑based RLVR runs have raised contamination questions. paper pdf

Olmo 3-Base 32B challenges other open 32B models on core benchmarks

Benchmark charts shared with the release show Olmo 3‑Base 32B beating or matching strong open peers like Marin 32B, Apertus 70B, Qwen 2.5 32B, Gemma 3 27B, and even Llama 3 170B on tasks such as HumanEval (66.5), DROP (81.0), SQuAD (94.5), and CoQA (74.1). benchmark charts For AI engineers this means you get a competitive, mid‑sized base model for code, reading comprehension, and QA that can still fit on a single 80GB GPU, but without giving up the transparency and fine‑tuning control that most closed 30–70B models lack.

Olmo 3-Think 32B nears Qwen3 on math and reasoning benchmarks

The Olmo 3‑Think 32B reasoning model lands within 1–2 points of Qwen3 32B and Qwen3 VL 32B Thinking on tough reasoning suites, scoring around 89.0 on IFEval, 96.1 on MATH, ~90 on BigBench‑Hard, and 89.7 on HumanEvalPlus, while also edging those peers on OMEGA. benchmark charts Ai2 credits a three‑stage post‑training pipeline—Dolci‑Think SFT, an expanded DPO setup that exploits deltas between Qwen3 32B and 0.6B responses, and large‑scale RLVR—for turning the 32B base into a high‑end open reasoning model, which is good news if you want near‑frontier math/code performance without leaving the open ecosystem.

Ai2 and Hugging Face schedule Olmo 3 deep-dive livestream

Ai2 is doing a live technical walkthrough of Olmo 3 with Hugging Face at 9am PT, followed by an "afterparty" discussion on Ai2’s Discord. livestream invite For engineers and researchers this is a chance to hear the training and post‑training details directly from the authors, ask about design decisions like context length, data mixtures, DPO setup, and RLVR tricks, and see how they envision people extending the new base and Think checkpoints.

Olmo 3 team hints at forthcoming write-ups on training infrastructure and code execution

Several Olmo 3 authors are teasing behind‑the‑scenes stories about the infra and training stack, including "babysitting" long training runs and the specialized NCCL and code‑execution setups that kept large‑scale RL experiments stable. training babysitting One engineer says they have "very fun" training‑run anecdotes they hope to publish in the next few weeks, while another notes a forthcoming write‑up on their code execution environment for RL, and teammates call out contributors who debugged low‑level NCCL issues during post‑training. code execution note nccl comment If you care about reproducing Olmo‑class training or adapting their RL pipeline, it’s worth watching for these deep‑dive posts as companions to the main technical report.


Stay first in your field.

No more doomscrolling X. A crisp morning report for entrepreneurs, AI creators, and engineers. Clear updates, time-sensitive offers, and working pipelines that keep you on the cutting edge. We read the firehose and hand-pick what matters so you can act today.

I don’t have time to scroll X all day. Primer does it, filters it, done.

Renee J.

Startup Founder

The fastest way to stay professionally expensive.

Felix B.

AI Animator

AI moves at ‘blink and it’s gone’. Primer is how I don’t blink.

Alex T.

Creative Technologist

Best ROI on ten minutes of my day. I’ve shipped two features purely from their daily prompts.

Marta S.

Product Designer

From release noise to a working workflow in 15 minutes.

Viktor H

AI Artist

It’s the only digest that explains why a release matters and shows how to use it—same page, same morning.

Priya R.

Startup Founder

Stay professionally expensive

Make the right move sooner

Ship a product

WebEmailTelegram

On this page

Executive Summary
Feature Spotlight: Feature: Nano Banana Pro (Gemini 3 Pro Image) goes live across Google and partners
🍌 Feature: Nano Banana Pro (Gemini 3 Pro Image) goes live across Google and partners
Nano Banana Pro (Gemini 3 Pro Image) launches across Gemini and Google tools
Benchmarks put Nano Banana Pro ahead of GPT‑Image 1 in quality and text rendering
fal.ai ships day‑0 Nano Banana Pro text‑to‑image and editing APIs
Higgsfield offers unlimited 4K Nano Banana Pro with aggressive Black Friday deal
Genspark integrates Nano Banana Pro into its all‑in‑one AI workspace
🧩 Open weights: Olmo 3 (7B/32B base, Instruct, Think)
Ai2 releases fully open Olmo 3 7B/32B family
Olmo 3 ships 7B RL Zero datasets and checkpoints for math, code and instructions
Olmo 3-Base 32B challenges other open 32B models on core benchmarks
Olmo 3-Think 32B nears Qwen3 on math and reasoning benchmarks
Ai2 and Hugging Face schedule Olmo 3 deep-dive livestream
Olmo 3 team hints at forthcoming write-ups on training infrastructure and code execution
🗺️ Model roadmaps and imminent drops
Signals stack up for imminent Claude Opus 4.5 and Claude Code Desktop
Elon Musk targets Grok 4.20 “major improvement” upgrade by Christmas
Perplexity’s Comet agentic browser quietly appears on Android Play Store
🛠️ Agent architectures: subagents, context, and code execution
Anthropic: code execution plus smart context editing boosts Claude agents by 39%
Replit pitches core-loop subagent orchestrator as “Year of the Subagent” pattern
Sourcegraph warns Amp coding threads beyond ~350k tokens hurt quality
Kilo Code shows prompt-to-game loop with built-in deploy
🛡️ Agent/IDE security: prompt‑injection exfil and mitigations
Markdown image exfil bug resurfaces across agentic IDEs
DSPy Spotlight adds production defense for indirect prompt injection
⚖️ EU shifts privacy/AI rules to cut friction
EU plans to relax GDPR and delay AI Act enforcement to cut compliance burden
🗣️ Voice AI at scale: new markets and enterprise usage
ElevenLabs launches in Korea with sub‑0.5s Agent Platform for enterprises
ElevenLabs powers 1.5M AI mock interviews for Apna job seekers
🎨 Creative AI beyond Google
Dreamina MultiFrames turns 10 stills into a 54s, prompt‑controlled video
Meta’s SAM3 shows robust real‑world video segmentation in early creator tests
ImagineArt 1.5 Preview climbs to #3 on Artificial Analysis text‑to‑image ELO
KAT‑Coder‑Pro auto‑codes a procedural Minecraft‑style Christmas house in three.js
Tencent teases HunyuanVideo 1.5 with sketch‑to‑3D style preview
💼 Commerce and enterprise adoption signals
Flowith Black Friday deal bundles Gemini 3 Pro and upcoming Banana 2 with deep discounts
📈 Community pulse: model fatigue and competition narrative
Community narrative shifts to “Google vs the rest” but race still open
Community ties METR’s 2h40 time-horizon to OpenAI’s “AI research intern” roadmap
Commentators highlight 300× yearly drop in “price per unit of intelligence” and warn there will be no bailout
Developers hit “100% model fatigue” after three days of flagship launches
Builders call out repeating hype cycle of each big model launch
🤖 Robots in factories and homes
Figure shares BMW production-line footage and highlights lessons for Figure 03
Sourccey teases 3’6" open-source home robot compatible with LeRobot