New Steps In Commodity AI Image Models, And What They Mean For Fashion

Key Takeaways:

  • Google’s new “Nano Banana” image generation model, part of the Gemini 2.5 Flash Image suite, introduces a significant leap forward in consistency and editability for general-purpose AI. This allows users to alter elements like clothing and backgrounds without the subject mutating, a capability previously exclusive to fine-tuned, industry-specific models.
  • Google’s upgrade makes consistency and editability widely available, but that doesn’t erase the case for fashion-focused AI. It makes it stronger. The real test for specialist platforms may now be how far they can go into the tougher territory – fit, fabric behaviour, colour fidelity – where generalist tools still struggle and where their real advantage could emerge.
  • With the technical hurdles of AI image generation easing, the industry’s pressure points move from capability to culture. The core challenge for fashion is no longer how to produce a consistent AI image, but how to use these tools in a way that resonates with consumers without eroding trust, brand identity, or the value of human creativity.

“Nano Banana”, a name as silly as the results are impressive, is Google’s new image generation and editing model, which was rolled into the equal mouthful that is Gemini 2.5 Flash Image, and pushed live this week. 

As is often the case with the big generative AI labs, the bizarre naming hides an iterative improvement in some respects. The new model generates high-quality outputs, and is better-integrated into the suite of Google’s other AI apps and services. But this model actually represents something of a deeper change – at least as far as mass market generative AI is concerned, since it aims to open the door to something that’s traditionally been the preserve of sector-specific models: consistency and editability. 

Nano Banana allows users to change an item of clothing, swap a background, or nudge an expression, and the subject does not suddenly mutate into something (or someone) else. 

google, nano banana

Whether this is a small adjustment or a bigger breakthrough depends on your perspective. For a lot of users of generative AI in work and at home, consistency has been the missing piece. Without it, building coherent visual identities, campaigns, or collections is difficult, as more often than not the user finds themselves left with disconnected one-offs, or, worse still, degrading assets that, over time, can begin to look like distorted echoes of the original generated image. Gemini 2.5 Flash Image promises something closer to true iteration, a way to work with images and manipulate their constituent parts, instead of restarting from scratch. 

The “Photoshop killer” label doing the rounds online is probably premature, but it shows how strongly the upgrade has landed with a userbase hungry for the ability to have generative workflows that actually flow, rather than constantly looping back to the beginning.

Where that “Photoshop” label is, perhaps, more appropriate is in the distribution strategy: Gemini 2.5 is part of the broader Gemini ecosystem, available through the app and via APIs/Vertex AI. That means, for most people, there are multiple ways to get access to it at an affordable price – a strategy that also served to make other de facto image editing suites ubiquitous.

And this matters because, until now, keeping visuals consistent within narrow use cases like fashion and beauty (whether that’s for inspiration, internal communication, or shopper-facing sales materials) often required specific tools that had been fine-tuned or reinforced using single industry, or single brand, context. With Google’s new release, a version of that capability has become part of the everyday toolkit. And in a practical sense, swapping outfits on a model that otherwise stays unchanged is something that’s now available on tap.

This might, at first blush, feel like a threat to fashion-focused tools. After all, Google’s release makes campaign-style experimentation easier and puts a baseline version of consistency into mass-market hands. Which leads to a new question for companies offering solutions that deliver a similar vision in a way that’s tuned for fashion’s deeper needs: design workflows, fit, fabric behaviour, colour accuracy, collaboration, and so on? If general models open the door to experimentation and baseline capabilities, then specialist platforms must make sure the work that follows is grounded, usable, and credible if they’re going to stand out.

On one side, the democratisation of tools will encourage faster experimentation and wider adoption. On the other, it raises the bar for what counts as distinctive. If everyone can generate, edit, and maintain consistency, then differentiation comes from the details generalist models can’t supply.

Parallel Universes: From Future Frequencies to Gucci Cosmos. Gucci & Christie 3.0

Yet even as the tools for creating consistent campaign imagery are improving, the cultural outcomes we’ve seen when fashion pushes forward with AI remain mixed. The high profile AI advert that appeared in Vogue last month drew curiosity but also scepticism. Mango’s digital models, too, triggered criticism from both professionals and consumers who felt misled. From that perspective, does wide availability exacerbate the problem or simply “flood the zone” to where critics no longer have a foundation from which to care?

Moving too fast can invite backlash, but move too slow and the potential efficiency, content scale, and creativity benefits might pass you by. Of course, some brands are attempting both, Gucci has experimented with generative AI, while Etro has used it in some campaign contexts, and perhaps that kind of a hedge really is the best way forward at this juncture. 

Not so long ago, anyone trying to use AI for a campaign was fighting the basics: faces slipping out of shape, outfits changing halfway through, styling that looked similar but not quite the same. That constant push and pull, on the surface at least, is starting to ease, and generally speaking, generated images are getting “good enough” from both a results perspective and from the vantage point of user experience. 

Etro Spring 2024 Campaign

But still, when the image feels synthetic, the reaction changes. It can land as bold, or it can land as hollow, but in either case the audience isn’t just responding to the clothes, they’re responding to the choice to use AI at all.

Which is why mass market fashion in particular will hope ”Nano Banana” is received less as a technical milestone and more as a cultural accelerant. It makes “good enough” easier to reach, but it also raises the stakes on what counts as acceptable – taking us closer to a watershed moment that will be hard to walk back from.

So in that sense, Google’s latest image model is a measurable leap in terms of capability and cultural acceptance, despite the silly name. It gives anyone the ability to produce coherent, consistent AI imagery with minimal friction, raising the floor. What industry-focused platforms, and their end users, choose to do with the task of now raising the ceiling will be the determining factor in where generative imagery goes from here.

Exit mobile version