GPT Image 2 Leak Hands-on: Does It Beat Nano Banana Pro in Blind Tests?
Author
Nico
Date Published

TL;DR Key Takeaways
- GPT Image 2 has quietly appeared on the Arena blind testing platform under three codenames: maskingtape-alpha, gaffertape-alpha, and packingtape-alpha. Community tests show its text rendering and world knowledge capabilities significantly surpass previous generations.
- In blind test comparisons with Nano Banana Pro, GPT Image 2 leads in text accuracy, UI restoration, and world knowledge, though it still falls short in spatial reasoning (such as Rubik's Cube mirror reflections).
- The three models have been removed from LMArena. Combined with OpenAI's recent move to shut down Sora to free up compute power, an official release may be just around the corner.
How Was GPT Image 2 Discovered?
On April 4, 2024, independent developer Pieter Levels (@levelsio) was the first to break the news on X: three mysterious image generation models appeared on the Arena blind testing platform, codenamed maskingtape-alpha, gaffertape-alpha, and packingtape-alpha. 1 While these names sound like a hardware store's tape aisle, the quality of the generated images sent the AI community into a frenzy.
This article is for creators, designers, and tech enthusiasts following the latest trends in AI image generation. If you have used Nano Banana Pro or GPT Image 1.5, this post will help you quickly understand the true capabilities of the next-generation model.
A discussion thread in the Reddit r/singularity sub gained 366 upvotes and over 200 comments within 24 hours. User ThunderBeanage posted: "From my testing, this model is absolutely insane, far beyond Nano Banana." 2 A more critical clue: when users directly asked the model about its identity, it claimed to be from OpenAI.

Image Source: @levelsio's initial leak of the GPT Image 2 Arena blind test screenshot *1*
Text Rendering: Has the Biggest Pain Point of AI Image Generation Been Solved?
If you frequently use AI to generate images, you know the struggle: getting a model to correctly render text has always been a maddening challenge. Spelling errors, distorted letters, and chaotic layouts are common issues across almost all image models. GPT Image 2's breakthrough in this area is the central focus of community discussion.
@PlayingGodAGI shared two highly convincing test images: one is an anatomical diagram of the anterior human muscles, where every muscle, bone, nerve, and blood vessel label reached textbook-level precision; the other is a YouTube homepage screenshot where UI elements, video thumbnails, and title text show no distortion. 3 He wrote in his tweet: "This eliminates the last flaw of AI-generated images."

Image Source: Comparison of anatomical diagram and YouTube screenshot shown by @PlayingGodAGI *3*
@avocadoai_co's evaluation was even more direct: "The text rendering is just absolutely insane." 4 @0xRajat also pointed out: "This model's world knowledge is scary good, and the text rendering is near perfect. If you've used any image generation model, you know how deep this pain point goes." 5

Image Source: Website interface restoration results independently tested by Japanese blogger @masahirochaen *6*
Japanese blogger @masahirochaen also conducted independent tests, confirming that the model performs exceptionally well in real-world descriptions and website interface restoration—even the rendering of Japanese Kana and Kanji is accurate. 6 Reddit users noticed this as well, commenting that "what impressed me is that the Kanji and Katakana are both valid."
Blind Test Comparison: GPT Image 2 vs Nano Banana Pro
This is the question everyone cares about most: Has GPT Image 2 truly surpassed Nano Banana Pro?
@AHSEUVOU15 performed an intuitive three-image comparison test, placing outputs from Nano Banana Pro, GPT Image 2 (from A/B testing), and GPT Image 1.5 side-by-side. 7



Image Source: Three-image comparison by @AHSEUVOU15; from right to left: NBP, GPT Image 2, GPT Image 1.5 *7*
@AHSEUVOU15's conclusion was cautious: "In this case, NBP is still better, but GPT Image 2 is definitely a significant improvement over 1.5." This suggests the gap between the two models is now very small, with the winner depending on the specific type of prompt.
According to in-depth reporting by OfficeChai, community testing revealed more details 8:
- Watch Time Rendering: packingtape-alpha correctly rendered the time on a watch, while Nano Banana Pro failed.
- Minecraft Screenshots: In a test featuring a first-person Minecraft game screenshot set in Manhattan, maskingtape-alpha outperformed all other models in the series and Nano Banana Pro.
- World Knowledge: Investor Justine Moore (@venturetwins) tested the model with prompts like "an average engineer's screen" and "a young woman taking a selfie with Sam Altman," where the model demonstrated exceptionally strong world knowledge.
@socialwithaayan shared beach selfies and Minecraft screenshots that further confirmed these findings, summarizing: "Text rendering is finally usable; world knowledge and realism are next level." 9

Image Source: GPT Image 2 Minecraft game screenshot generation shared by @socialwithaayan [9](https://x.com/socialwithaayan/status/2040434305487507475)
Where Are the Weaknesses? Spatial Reasoning Remains a Flaw
GPT Image 2 is not without its weaknesses. OfficeChai reported that the model still fails the Rubik's Cube reflection test. This is a classic stress test in the field of image generation, requiring the model to understand mirror relationships in 3D space and accurately render the reflection of a Rubik's Cube in a mirror.
Reddit user feedback echoed this. One person testing the prompt "design a brand new creature that could exist in a real ecosystem" found that while the model could generate visually complex images, the internal spatial logic was not always consistent. As one user put it: "Text-to-image models are essentially visual synthesizers, not biological simulation engines."
Additionally, early blind test versions (codenamed Chestnut and Hazelnut) reported by 36Kr previously received criticism for looking "too plastic." 10 However, judging by community feedback on the latest "tape" series, this issue seems to have been significantly improved.
Why Now? Compute Reallocation After Sora's Shutdown
The timing of the GPT Image 2 leak is intriguing. On March 24, 2024, OpenAI announced the shutdown of Sora, its video generation app, just six months after its launch. Disney reportedly only learned of the news less than an hour before the announcement. At the time, Sora was burning approximately $1 million per day, with user numbers dropping from a peak of 1 million to fewer than 500,000.
Shutting down Sora freed up a massive amount of compute power. OfficeChai's analysis suggests that next-generation image models are the most logical destination for this compute. OpenAI's GPT Image 1.5 had already topped the LMArena image leaderboard in December 2025, surpassing Nano Banana Pro. If the "tape" series is indeed GPT Image 2, OpenAI is doubling down on image generation—the "only consumer AI field still likely to achieve viral mass adoption."
Notably, the three "tape" models have now been removed from LMArena. Reddit users believe this could mean an official release is imminent. Combined with previously circulated roadmaps, the new generation of image models is highly likely to launch alongside the rumored GPT-5.2.
How to Experience and Compare AI Image Models Yourself
Although GPT Image 2 is not yet officially live, you can prepare now using existing tools:
- Follow the Arena Blind Test Platform: Visit arena.ai to participate in blind test voting for image models. New models may reappear under anonymous codenames at any time, and every vote you cast shapes the leaderboard.
- Horizontal Comparison of Existing Models: Test Nano Banana Pro, GPT Image 1.5, Seedream, and other models using the same set of prompts to establish your own evaluation benchmark. Focus on three dimensions: text rendering, UI restoration, and character detail.
- Save and Manage Your Prompt Library: In YouMind, you can save your test prompts and generated results to a Board for easy future comparison. YouMind currently supports multiple image models like Nano Banana Pro, GPT Image 1.5, and Seedream 4.5. Once GPT Image 2 is officially released, you can switch and compare directly within the same platform.
- Refer to Community Prompt Libraries: awesome-nano-banana-pro-prompts provides over 10,000 curated prompts supporting 16 languages, which can serve as a starting point for testing new models.
Note that model performance in Arena blind tests may differ from the official release version. Models in the blind test phase are usually still being fine-tuned, and final parameter settings and feature sets may change.
FAQ
Q: When will GPT Image 2 be officially released?
A: OpenAI has not officially confirmed the existence of GPT Image 2. However, the removal of the three "tape" codename models from Arena is widely seen by the community as a signal that an official release is 1 to 3 weeks away. Combined with GPT-5.2 release rumors, it could launch as early as mid-to-late April 2024.
Q: Which is better, GPT Image 2 or Nano Banana Pro?
A: Current blind test results show both have their advantages. GPT Image 2 leads in text rendering, UI restoration, and world knowledge, while Nano Banana Pro still offers better overall image quality in some scenarios. A final conclusion will require larger-scale systematic testing after the official version is released.
Q: What is the difference between maskingtape-alpha, gaffertape-alpha, and packingtape-alpha?
A: These three codenames likely represent different configurations or versions of the same model. From community testing, maskingtape-alpha performed most prominently in tests like Minecraft screenshots, but the overall level of the three is similar. The naming style is consistent with OpenAI's previous gpt-image series.
Q: Where can I try GPT Image 2?
A: GPT Image 2 is not currently publicly available, and the three "tape" models have been removed from Arena. You can follow arena.ai to wait for the models to reappear, or wait for the official OpenAI release to use it via ChatGPT or the API.
Q: Why has text rendering always been a challenge for AI image models?
A: Traditional diffusion models generate images at the pixel level and are naturally poor at content requiring precise strokes and spacing, like text. The GPT Image series uses an autoregressive architecture rather than a pure diffusion model, allowing it to better understand the semantics and structure of text, leading to breakthroughs in text rendering.
Summary
The leak of GPT Image 2 marks a new phase of competition in the field of AI image generation. Long-standing pain points like text rendering and world knowledge are being rapidly addressed, and Nano Banana Pro is no longer the only benchmark. Spatial reasoning remains a common weakness for all models, but the speed of progress is far exceeding expectations.
For AI image generation users, now is the best time to build your own evaluation system. Use the same set of prompts for cross-model testing and record the strengths of each model so that when GPT Image 2 officially goes live, you can make an accurate judgment immediately.
Want to systematically manage your AI image prompts and test results? Try YouMind to save outputs from different models to the same Board for easy comparison and review.
References
[1] @levelsio: OpenAI's new image model GPT-Image-2 leaked
[2] Reddit r/singularity: GPT-IMAGE-2 suspected to appear on LMArena
[3] @PlayingGodAGI: GPT-Image-2 leak ends the era of text rendering flaws
[4] @avocadoai_co: GPT Image 2 text rendering showcase
[5] @0xRajat: GPT Image 2 blind test screenshot
[6] @masahirochaen: GPT-Image-2 precision test
[7] @AHSEUVOU15: Nano Banana Pro vs GPT Image 2 vs GPT Image 1.5 three-image comparison
[8] OfficeChai: Three tape-named models spark buzz on Arena, rumored to be OpenAI's GPT-Image 2
[9] @socialwithaayan: GPT Image 2 beach selfie and Minecraft screenshot
[10] 36Kr: OpenAI blind tests new model; Altman reportedly pausing Sora to focus on ChatGPT