If youโre new, welcome! If youโre a regular, thanks for being here as we navigate this AI shift in eCommerce. I share what Iโm testing and whatโs actually moving the needle. Whereโs the vault you ask? Every previous edition is saved here.
๐๏ธ ChatGPT-Images-2.0 Thinks Before It Draws๐ฅ๐
If you've been loyal to Nano Banana for the last six months, you're not alone. I have too.
But something shipped yesterday that changes the math.
On April 21, OpenAI released ChatGPT Images 2.0, powered by a new model called gpt-image-2. And for the first time, an image model does something none of the others do. It thinks before it draws.
That sounds like marketing copy. It isn't. It's a structural shift in how image generation works. And for those of us generating product images, A+ content, and ad carousels at scale, it changes the kinds of prompts we can realistically ship.
Check out the difference:

GPT-Image-2

Nano Banana 2
What Actually Shipped
Three things matter.
๐ถ Native reasoning. The model plans multiple images from a single prompt, searches the web during generation (for product refs, brand styles, factual accuracy), and double-checks its own output before rendering. Thinking mode is paid-gated. Standard mode is free for all ChatGPT users.
๐ถ 2K resolution by default. 2,048 pixels on the long edge. Print-ready without upscaling.
๐ถ Dense text rendering. This is the big one. Typography, labels, barcodes, data charts, and UI screenshots render with accuracy that Nano Banana Pro and Flux 2 can't currently match. The detail most of us have been hacking around for two years.
Available through ChatGPT, Codex, and the API as gpt-image-2.
What "Thinks Before It Draws" Actually Means
Most image models, including Nano Banana 2, generate in one shot. Prompt in, image out. Want three variations? Run it three times.
Images 2.0 plans first. If your prompt says "generate a 5-slide Instagram carousel with consistent faces," it asks itself which poses, which backgrounds, which continuity rules keep this consistent, then renders with that plan baked in.
The practical result: fewer "same prompt, different faces" disasters. Better multi-image continuity. Smarter object placement in complex scenes.
The difference between a model that follows instructions and a model that understands the brief.
The Test: Four E-Commerce Prompts, Two Models
Sellers don't need one beautiful image. We need a set that shares a visual language. Hero, lifestyle, carousel, data infographic, brand collab. All consistent, all on-spec, all on-brand.
I picked 4 prompts that cover the categories sellers actually use every week. Each one stress-tests a specific capability:
๐ถ Prompt 1: Instagram carousel with character consistency (multi-image continuity)
๐ถ Prompt 2: Lifestyle product shot (baseline photography realism)
๐ถ Prompt 3: Data infographic (dense text plus data viz)
๐ถ Prompt 4: Limited-edition brand collab mockup (dual-brand synthesis)
I'm running each on ChatGPT Images 2.0 (Thinking mode) and Nano Banana 2. Same prompt, same reference images where applicable. Then grading on brand coherence, text accuracy, multi-image consistency, Amazon-spec readiness, and honest usability.
Copy these into your own sessions and run the parallel test. Your results will teach you more than mine will.
Prompt 1: Instagram Carousel With Character Consistency
Tests multi-image continuity. The prompt that most exposes the gap between a one-shot model and a planning model.
Create a 5-slide Instagram carousel for an allergen-free bakery called
"Sweet Inclusion." Target audience: parents of kids with food allergies.
Slide 1: A young girl (age 7, light brown hair, blue eyes, wearing a
yellow sweater) looking sad at a birthday party dessert table, unable
to eat the cake.
Slide 2: Same girl, same sweater, now smiling brightly, eating one of
our allergen-free cupcakes.
Slide 3: Same girl and her mom (warm, 35ish, brunette) baking together
in a bright home kitchen.
Slide 4: Clean product shot of the cupcake range on light linen. Small
"Sweet Inclusion" wordmark.
Slide 5: Call to action slide: "Every kid deserves a yes. Order today."
Brand colors soft pink and cream.
Keep the girl and mom visually consistent across slides 1 through 3.
1:1 square. Warm natural light across all slides.Hereโs the comparison:

GPT-Images-2

Nano Banana 2
Prompt 2: Lifestyle Product Shot
Tests baseline photography realism and prompt nuance. The workhorse prompt every Amazon seller runs weekly.
Photograph: a wireless earbuds charging case (matte black, rounded
rectangle, about the size of a dental floss container) on a wooden
gym bench. Soft morning light from a window on the left, creating
gentle highlight on the case.
Supporting elements: a small hand towel folded nearby, faint
condensation droplets on the wood, a water bottle blurred in
background.
Shallow depth of field, 50mm equivalent, product sharp, background
soft bokeh. Natural color grading, slight warm temperature. No
branding text on the product.
Output: 4:5 vertical, suitable for Instagram feed.Watch for: Light direction (does it actually come from the left?). Bokeh (natural or painted-on?). Scale (does the earbuds case look the right size or like a giant lunchbox?).

GPT-Images-2

Nano Banana 2
Prompt 3: Data Infographic
Tests text rendering plus data visualization. Most sellers still build these in Canva because image models butcher numbers and chart geometry. Images 2.0 is betting this is solvable.
Design a vertical infographic titled "Why HydroCore Beats the Bottle."
Aspect ratio: 4:5 (Instagram post).
Four data points, each with a labeled circular progress indicator:
- 24 hours: "Cold Retention"
- 98%: "Customer Satisfaction"
- 0: "BPA Content"
- 14 oz: "Total Weight"
Below: "Built for the 5 AM runner, the 9 AM commuter, and the 3 PM
break."
Modern sans-serif, bold numerals, lighter labels. Deep navy background,
warm cream text, coral accent on the progress arcs. Small HydroCore
wordmark at the bottom.
Numbers must render big, readable, and exact.
GPT-Images-2

Nano Banana 2
The Scorecard
Grade each output on five axes. 1 to 5 per axis, one note per prompt.
๐ถ Brand coherence: on-brand or stock?
๐ถ Text accuracy: numbers, labels, words rendered correctly?
๐ถ Multi-image consistency: across Prompt 1's 5 slides, do people and scenes hold?
๐ถ Amazon-spec readiness: uploadable without a second pass?
๐ถ Usability: would you actually ship this in a live campaign?
The model with the higher average wins that prompt. But pay attention to where each model wins. It's rarely a blowout.
Who Should Switch, Who Should Stay
๐ถ Dense text work (A+ content, comparison charts, data infographics, UI mockups, product labels): Images 2.0 Thinking mode looks like the new leader. The text rendering gap is real.
๐ถ High-volume product photography (10 to 100 variations a week): Nano Banana 2 is still the workhorse. 3x faster, half the cost, 95% of the quality for standard product shots.
๐ถ Multi-image campaigns (carousels, storyboards, before-and-afters): Images 2.0 Thinking mode is worth the subscription test. Native planning is built for this job.
๐ถ Free plans: The non-Thinking version of Images 2.0 is free for everyone. Even without reasoning, the 2K default and improved text rendering justify adding it to your stack.
This isn't a "switch and never look back" moment. It's a "add a second tool and learn when to reach for which" moment.
We are living in exciting times that keep us on our toes!
PPC Ninja is helping brands future proof their listings for AI, helping you build RUFUS enabled, stunning images and videos with AI. Hit reply on this to chat with us. Explore how we can scale your content production across Social media, Amazon ads, Amazon Posts efficiently and affordably.
NERD BYTES
Neat Little Hack: The WebP Drop Folder ๐ค
Every image on your hard drive (png, jpeg, etc) is probably 3 to 10 times bigger than it needs to be.
That fat tax shows up everywhere. Slow Shopify pages. Ad platforms taking forever to approve creative. Supplier emails bouncing on attachment size. Blog and newsletter images that slow your own site down. Each one is small. Together they're a drag you've stopped noticing.
The fix is a file format called WebP.
What WebP Actually Is
Honestly? I didn't know what WebP was until a few months ago either.
WebP is a modern image format Google built in 2010. Roughly 70 to 80% smaller than PNG, 25 to 35% smaller than JPG, at the same visual quality. Shopify, Meta, Google Ads, and every modern browser support it natively. It's just not the default anywhere, so most of us never switched.
One heads up for Amazon sellers: Amazon Seller Central does not accept WebP uploads for product listing images. Amazon still requires JPEG, PNG, TIFF, or GIF. So WebP is for everything around your Amazon listings (Shopify stores, ad platforms, email, blog images), not the listings themselves.
What makes the conversion possible: FFMPEG
What is FFMPEG? A free, twenty-year-old command-line tool that can convert, compress, and resize any image or video. Powerful, intimidating. Nobody casually opens a terminal for it.
Until Claude became the front end.
The Drop Folder Workflow
Here's the whole setup:
Make a folder on your desktop called
webp-drop.Drop all the images you want to shrink into it.
Point Claude Code to that foldr and type: "Convert the images in /webp-drop to WebP format."
Claude runs FFMPEG, saves the smaller version back to the folder, done. About four seconds per file. Done.

What You Actually Save
๐ถ 1.5MB PNG screenshot โ ~200KB
๐ถ 800KB product JPG โ ~280KB
๐ถ 4MB creative export โ ~700KB
Same visual quality. Tiny fraction of the file size.
One-Time Setup
Install FFMPEG (brew install ffmpeg on Mac, winget install ffmpeg on Windows). Make the folder. Done forever.
The quiet advantage isn't WebP, and it isn't FFMPEG. It's that boring workflows you used to ignore can now be a one-sentence request. Every seller I know has a dozen of these hiding in their week.
/
/
/
COOL TOOLS
Anthropic quietly shipped Claude Design on April 17 as a new Anthropic Labs product. It's built for the "I need a polished visual in ten minutes and I don't want to open a design tool" moment we all have three times a week.
What It Actually Does
You describe what you need. Claude builds a first version. You refine through conversation, inline comments, direct edits, or custom sliders (Claude generates the sliders based on what you might want to tweak). When you're happy, export to Canva, PDF, PPTX, or standalone HTML.
It also learns. Feed it your brand colors, fonts, and logo once. Every design after that uses them by default.
The Quiet Upgrade: It Runs on Opus 4.7
Claude Design is powered by Opus 4.7, which means it inherits the tripled vision resolution. Drop in a screenshot of a competitor's A+ content and it can actually read every headline, bullet, and comparison row. Then ask it to riff on the structure for your own product. That combination (competitor analysis plus fast visual iteration) is the specific use case I'm testing on a real listing this week.
The Catch
Research preview. Included free for anyone already on Claude Pro, Max, Team, or Enterprise. No extra charge, but also no standalone free tier. If you're on Claude Free, you'll need to upgrade to try it.
I'm keeping this one in my workflow rotation for a month and will share what held up (and what didn't) in a future issue.
NEWS WORTH FOLLOWING
UPCOMING EVENTS
April 28th is the last Tuesday of the month. Time to tune in to the Go with the Flow Podcast where Danny McMillan and I dive into AI topics. We will be talking about Claude Code, specifically some of the newer roll-outs and ways to improve Claude Skills. Listen here!

My good friend, Gary Huang, (who runs the 7 Figure Seller Summit) is hosting a live session next week focused on: From AI Tools to AI Systems: What Actually Works for Amazon Sellers. RSVP here: ๐ https://newsletter.7figuresellersummit.com/ocea3
We hope you liked this edition of the AI for E-Commerce Newsletter! Hit reply and let us know what you think! Thank you for being a subscriber! Know anyone who might be interested to receive this newsletter? Share it with them and they will thank you for it! ๐ Ritu
Rex Gelb spent a decade building HubSpot's paid engine. Now he's showing founders exactly how to do it.
On April 27th, get the framework to structure, launch, and scale paid media that drives pipeline, not just traffic. 20 minutes. Live Q&A. Free.



