I'm sure this has been written about but here's what happens long term - images are commoditized and lose their emotional appeal.
Probably about half of us here remember photos before the cell phone era. They were rare, and special, and you'd have a few photos per YEAR to look back on. The feel of photos back then, was at least 100x stronger than now. They were a special item, could be given as a gift. But once they became freely available that same amount of emotion is now split across many thousands of photos. (not saying this is good or bad, just increased supply reducing value of each item)
With image/art generation the same thing will happen and I can already feel it happening. Things that used to be beautiful or fantastic looking now just feel flat and AI-ish. If claymation scenes can be generated in 1s, and I see a million claymation diagrams a year, then claymation will lose its charm. If I see a million fake Tom Cruise videos, then it oversaturates my desire for desire for all Tom Cruise movies.
What a time to be alive.
electrosphere 50 seconds ago [-]
It reminds me of the Star Wars content thats come out recently - before there was the Original Trilogy which we all watched many times and the lines became iconic. Since then it's all become a mismash and blur of mediocrity due to over-exposure.
(except The Mandalorian)
gradus_ad 3 minutes ago [-]
Exactly right. A flood of content devalues content, which devalues whatever makes that content. AI fanboys gushing over every overhyped release and reveal but are really just showing themselves to be incapable of second order thinking and very susceptible to marketing hype.
GaggiX 2 minutes ago [-]
You can still buy a Polaroid, there is one factory left in the world able to produce the film required but they still make them.
nathan_compton 2 minutes ago [-]
People here like to say "Commoditize your Compliment" but to a company the size of google or amazon literally EVERYTHING is your compliment. Too bad no philosopher or political scientist or economist every thought about this stuff before or we might have some kind of plan to make the future less miserable and alienating.
nickandbro 35 minutes ago [-]
These image gen models are getting so advanced and life like that increasingly the general public are being duped into believing AI images are actually real (ex Facebook food images or fake OF models). Don't get me wrong I will enjoy the benefits of using this model for expressing myself better than ever before, but can't help feeling there's something also very insidious about these models too.
WarmWash 32 minutes ago [-]
It's more likely than not that every single person who uses the internet has viewed an AI image and taken it as real by now.
The obvious ones stand out, but there are so many that are indiscernible without spending lots of time digging through it. Even then there are ones that you can at best guess it's maybe AI gen.
tokai 2 minutes ago [-]
Maybe not an actual argument for anything, but even before these image models everyone that used the internet had seen a doctored image they believed to be real. There was a reason that 'i can tell by the pixels' was a meme.
versk 27 minutes ago [-]
At the point now where basically any photo that isn't shared by someone I trust or a reputable news organisation is essentially unverifiable as being real or not
The positive aspect of this advance is that I've basically stopped using social media because of the creeping sense that everything is slop
yieldcrv 19 minutes ago [-]
people only notice when they are prompted to look for AI or scrutinize AI
a lot of these accounts mix old clips with new AI clips
or tag onto something emotional like a fake Epstein file image with your favorite politician, and pointing out its AI has people thinking you’re deflecting because you support the politician
Meanwhile the engagement farmer is completely exempt from scrutiny
Its fascinating how fast and unexpected the direction goes
vunderba 4 minutes ago [-]
Jaded, but if I knew there was a possibility of a bunch of incriminating footage of me (images, video, etc.) out there in the pre-AI days, I would do my absolute best to flood the internet with as many related deepfakes (including of myself) as possible.
whynotmaybe 25 minutes ago [-]
>fake OF models
Soon many real OF models will be out of job when everyone will be able to produce content to their personal taste from a few prompts.
sodacanner 6 minutes ago [-]
People already have access to every form of niche pornography they could dare to imagine (for absolutely free!), I really doubt that 'personal taste' is the part that makes OF models their money. They'll be fine.
pousada 16 minutes ago [-]
You can’t really because these powerful models are censored.
You can create lewd pictures with open models but they aren’t nearly as good or easy to use.
baal80spam 21 minutes ago [-]
And this can't come soon enough.
noumenon1111 9 minutes ago [-]
Coming soon... YOU!
coldtea 18 minutes ago [-]
And they might have to gasp! get an honest job!
switchbak 11 minutes ago [-]
I don't know much about that side of things, but I presume that's hard work! Maybe not always so honest though.
kevincox 20 minutes ago [-]
I actually think this was a good thing. Manipulating images incredibly convincingly was already possible but the cost was high (many hours of highly skilled work). So many people assumed that most images they were seeing were "authentic" without much consideration. By making these fake images ubiquitous we are forcing people to quickly learn that they can't believe what they see on the internet and tracking down sources and deciding who you trust is critically important. People have always said that you can't believe what you see on the internet, but unfortunately many people have managed without major issue ignoring this advice. This wave will force them to take that advice to heart by default.
manuelabeledo 5 minutes ago [-]
> By making these fake images ubiquitous we are forcing people to quickly learn that they can't believe what they see on the internet and tracking down sources and deciding who you trust is critically important.
Has this thought process ever worked in real life? I know plenty of seniors who still believe everything that comes out of Facebook, be AI or not, and before that it was the TV, radio, newspapers, etc.
Most people choose to believe, which is why they have a hard time confronting facts.
lm28469 17 minutes ago [-]
I feel like there is one or two generations of people who are tech savy and not 100% gullible when it comes to online things. Older and younger generations are both completely lost imho, in a blind test you wouldn't discern a monkey from a human scrolling tiktok &co
Havoc 13 minutes ago [-]
Don’t think the demand for real OF is going anywhere
derwiki 11 minutes ago [-]
How do you know they’re real right now?
JasonADrury 4 minutes ago [-]
A lot of escorts have OF profiles.
techpression 27 minutes ago [-]
Oh we’ve seen nothing yet of the chaos that generative ai will unleash on the world, looking at Meta platforms it’s already a multi million dollar industry of selling something or someone that doesn’t exist. And that’s just the benign stuff.
fortyseven 19 minutes ago [-]
It's shitty, but I think it's almost as bad that people are calling everything AI. And I can't even blame them, despite how infuriating it is. It's just as insidious that even mundane things literally ARE AI now. I've seen at least twice now (that I'm aware of) where some cute, harmless, otherwise non-outrageous animal video was hiding a Sora watermark. So the crazy shit is AI. The mundane shit is AI. You wonder why everyone is calling everything AI now. :P
switchbak 8 minutes ago [-]
It seems like a low level paranoia - now I find myself double checking that the youtube video I'm watching isn't some AI slop. All the creators use Getty b-rolls and increasingly AI generated stuff so much that it's not a far stretch to have the voice and script all be auto generated too.
I suppose if the AI was able to tell me a true and compelling story, I might not even mind so much. I just don't want to be spoon fed drivel for 15 minutes to find it was all complete made up BS.
throwaway613746 19 minutes ago [-]
[dead]
fasteddie31003 19 minutes ago [-]
I'm building my personal home right now. The AI image models have been a game-changer in designing the look of the house. My architect did an OK job, but the details that Nano Banana added really bring the house up a notch. I just do hundreds of renders from the basic 3D models and I find looks that I like and iterate from there. We are implementing the renders from Nano Banana over our Interior Designers designs. We would not have hired the Interior Designers again after using Nano Banana to do our interiors.
I think part of the issue with architects and designers today is that they use CAD too much. It's easy to design boxes and basic roof lines in CAD. It's harder to put in curves and more craftsman features. Nano Banana's renders have more organic design features IMO.
Our house is looking great and we're very happy how it's going so far with a lot of the thanks to Nano Banana.
kristjansson 14 minutes ago [-]
Part of the job of interior design is delivering the promised images in … yknow, physical reality? How are you going from nano banana images to actual plans, materials, finishes, products, paint codes, … ?
fasteddie31003 13 minutes ago [-]
I just gave the renders to the cabinet makers and they had no problems recreating.
PunchTornado 11 minutes ago [-]
not the op, but this is what i did too and bypassed the designer. I iterated with nano banana and gave the result to the company that builds the kitchen. the middleman is gone now.
bartman 10 minutes ago [-]
Can you write a bit more about your workflow? I've been thinking about doing the same, but since I'm very non-interior-design minded have struggled to ask the right things.
Like...
What are your inputs to the model? Empty renders of the space, or more fully decorated views/ photos?
Do you have a light harness around this to help you discover the style you like and then stay consistent with it?
Do you find that giving a lot of context around the space you're designing helps (it hasn't in my attempts)?
shostack 2 minutes ago [-]
What tooling are you using to use this and manage it?
CWuestefeld 9 minutes ago [-]
What they've chosen as examples to illustrate the strength of the new model surprises me.
The "cubism" example seems like it would be a closer fit to something like stained glass or something. I don't think the thing really understands what cubism was all about. Cubist painters were trying to free themselves from the confines of a single integral plane of perspective by allowing themselves to show various parts of the image from different viewpoints, different times, different styles, etc.
The division of the image into geometric shapes is just a by-product of that quest, whereas the examples here have made it the sum total of the whole piece.
This feels to me like an example of how LLMs still don't "understand" what the art means, and are just aping its facade.
yakattak 36 minutes ago [-]
I think this tech is cool, from an engineering perspective. I’m trying to figure out if there’s any justification for using it in a business world outside of: “We don’t want to pay an artist.”
You can argue things like code generation are an extension of the engineer wielding it. Image generation just seems like a net negative overall if it’s used at scale.
Edit: By scale, I mean large corporations putting content in front of millions. I understand the appeal for smaller businesses where they probably weren’t going to pay an artist anyway.
alex43578 26 minutes ago [-]
When a company uses a photocopier, they don’t want to pay a scribe.
When a company sends an email or docu-sign, they don’t want to pay a courier.
Technology supplements or replaces jobs, often reducing costs. This is no different.
nindalf 21 minutes ago [-]
Art isn't just a job or a way to make money, like being a courier is.
progbits 19 minutes ago [-]
For corporate art it is. Nobody draws memphis out of passion.
dizlexic 7 minutes ago [-]
The real victims here are going to be the graphic designers who worked for firework importers.
garbawarb 27 minutes ago [-]
Advertising? "We don't want to pay an artist" goes a long way for a small business with a limited budget.
whynotmaybe 16 minutes ago [-]
We're using voice generation from clipchamp for our promotional videos.
It's an ethical conundrum because we're not paying anyone, but we don't have the money to pay anyone, and it's good enough for our budget.
But we're getting used to the process of changing a part of the text in a few seconds without any artist involved and for 0$.
I guess that soon we'll be able to create voice sample from know personalities for a few $ with prices based on the popularity of the artist and some sanity check based on the artist preferences.
yakattak 11 minutes ago [-]
I think this is where I see the benefit for small business. I don’t want to speak for you, but I imagine it’s either “no voice over, we can’t afford it.” or “inexpensive AI voice over to make it more accessible and appealing.”
My thought is the large corps that could afford it, still won’t because it’s a cost they don’t need to incur. For them it’s not even a moral conundrum.
18 minutes ago [-]
rm_-rf_slash 23 minutes ago [-]
It can also backfire. AI slop ads and marketing material imply cut corners and poor quality products. If a bakery isn’t going to bother touching up its AI slop banner, I don’t expect their cookies to be great either.
switchbak 4 minutes ago [-]
Every local business I deal with is completely lacking on the online side. They might have square POS terminals and all that stuff, but their website either doesn't exist, sucks (not updated in years) or they throw me to Facebook (also sucks).
This is like the last mile for online presence. The average barber out here doesn't use Squarespace, barely knows how to use Facebook and doesn't touch GenAi. But they can still cut your hair pretty well - tech savvyness doesn't have a huge connection to business competence out here.
gwd 13 minutes ago [-]
FWIW I've never seen a correlation between a small company's website and the quality of their product. Slick website? Maybe they care for their craft, maybe they're all marketing and no content. Website stuck in 1998? Maybe they're sloppy and don't care; maybe they care about their core product, not a slick marketing brochure. I don't see any reason AI would be different in that regard.
awepofiwaop 13 minutes ago [-]
The amount of lost revenue due to the implication of cut corners needs to be higher than the cost of hiring an artist by enough of a margin that the managers who make the decision start to care, and enough that they're willing to put the effort into hiring an artist.
hypeatei 19 minutes ago [-]
This assumes that models won't improve and you'll always be able to tell that it's "AI slop" ... that seems like a bad bet. Five years ago you'd be laughed out of the room for suggesting that a computer could produce images from a natural language prompt and that it'd be accessible to everyone -- not just corporations with deep pockets.
yakattak 16 minutes ago [-]
Yeah if/when it becomes indistinguishable I think most people won’t care. That being said I do think someone finding out something is AI generated will be met with poor response. Does that ultimately matter? Probably not in a business world.
konschubert 24 minutes ago [-]
I disagree with your premise that everybody should endure friction and cost such that artists can earn a living producing cookie-cutter content.
RickS 9 minutes ago [-]
Same answers you'd use beyond "we don't want to pay an engineer". 100x shorter iteration speed, and the associated workflow (stream of microrevisions and spaghetti throwing), top quartile outputs in many langs/styles/contexts without having to source, hire, and maintain a fleet of separate specialists who can quit when they feel like it.
I'm torn on the scale thing. It definitely seems net negative. But I think we collectively underestimate just how deeply sick the existing thing already is. We're repulsed by image gen at scale because it breaks our expectation that images are at least somewhat based on reality, that they reflect the natural world or what we can really expect from a product, from a company, from the future. But that was already a bad expectation: when's the last time you saw a mcdonalds meal that looked like the advert? Or a sub-30$ amazon product that wasn't a complete piece of shit? Advertisements were already actively malicious fantasies to exploit the way our brains react to pictures. They're just fantasies that required whole teams of humans doing weird bullshit with lighting and photoshop, and I'm not sure that's much better. It was already slop. All the grieving we do about the loss of truth, or the extent to which corps will gleefully spray us with mind-breaking waterfalls of outright lies, I think those ships sailed a long time ago. The disgust, deceit, the rage we feel about genAI slop is the way we should have felt about all commercials since at least the 80s IMO.
yakattak 4 minutes ago [-]
> Advertisements were already actively malicious fantasies to exploit the way our brains react to pictures. They're just fantasies that required whole teams of humans doing weird bullshit with lighting and photoshop, and I'm not sure that's much better.
This is a good point. My gut reaction is “well at least someone was paid to do it and can continue to keep society/the economy going ”.
I can see the other side where that’s a soulless job. Not sure what’s worse. Soulless job where your skills apply or even less jobs in a competitive industry.
bonoboTP 33 minutes ago [-]
Drafting, iteration, mockups. Quite useful during ideation.
yakattak 27 minutes ago [-]
All things traditionally done by artists or artist adjacent roles. I can understand at an individual level, say for a solo gamedev who wasn’t going to pay an artist anyway. That’s not at scale though.
Larian Studios most recently was under fire for this [1]. Like I can see a director going “what would X look like?” and then speeding over to the concept artists for a proper rendition if they liked it. I don’t think this is at scale though. Any large business is just going to get rid of the concept artists.
There are many places in general office work where you need some kind of graphics. Slides, reports, info graphics, dataviz. Or academic papers. Some are just illustrations, like a fancy clipart or stock photos, some are drafts for a proper tikz or svg or something that you then redo in draw.io etc. There is much more use for graphics than the use cases where people would ever even consider hiring an actual artist. I've seen good results for iterating on eg model architecture figures quickly between PhD students and supervisors, faster than dragging boxes around and fiddling with tikz. Obviously you don't simply paste the result into the paper. You redo it but it's a good discussion basis. That's for info graphics stuff. But the same can apply to creative stuff, like an event poster, an invitation card to your wedding, storyboards, mood boards, DIY interior design, outfit planning etc etc
yakattak 18 minutes ago [-]
Yeah that’s a good point. I don’t think that’s what I meant by “at scale” but I can see that being useful day to day.
sempron64 20 minutes ago [-]
Diagrams! So much documentation lacks diagrams because they are hard to make
yakattak 19 minutes ago [-]
True! Though I’d argue diagrams as code like PlantUML or Mermaid are better than an image!
vunderba 8 minutes ago [-]
Agree just from a text search perspective alone that Mermaid even ASCII diagrams are usually preferable.
jezzamon 30 minutes ago [-]
One major thing is photoreal use cases, which artists can't really do. A lot of that is deep fakes / scams but there are some real use cases
yakattak 27 minutes ago [-]
Isn’t that what photographers are for?
the_mar 11 minutes ago [-]
a friend of mine was a creative director and a big tech co until recently, she was replaced by AI
rafael09ed 28 minutes ago [-]
It is faster as well
testing22321 24 minutes ago [-]
> I’m trying to figure out if there’s any justification for using it in a business world outside of: “We don’t want to pay a human.”
You could easily say the same about anytime computers or robots or automation have taken a job away.
We’ve been going down this road for decades.
yakattak 21 minutes ago [-]
Those industries (computers, robots) created other jobs though. This doesn’t seem to.
30 minutes ago [-]
tantalor 27 minutes ago [-]
Won't somebody think of the window replacers?
JOJESU 2 minutes ago [-]
I’ve been exploring this exact problem space from the angle of extreme constraints (single-digit MB memory, no cloud assumptions).
I documented what broke first and why here, in case it’s useful:
https://github.com/nullclaw/nullclaw
nathan_compton 22 seconds ago [-]
So this is an ultra-minimalist software platform to farm work out to enormous energy chugging AI models?
jacquesm 33 minutes ago [-]
What a great thing this didn't exist in the past. We likely wouldn't have had any of the amazing artworks that we have now. Imagine an AI generated Mona Lisa, Nightwatch or Sistine Chapel ceiling because prompting would have been so much cheaper than paying Leonardo, Rembrandt or Michelangelo...
Now extrapolate to all other artforms. Sculpture seems safe, for now, but only barely so.
wordpad 24 minutes ago [-]
I feel like the complete opposite is true.
Artists aren't doing it for the money. With advanced tools like these they wouldve iterated much faster and created much grander designs.
Art is about pushing limits of what's possible and AI just raises those limits.
__alexs 20 minutes ago [-]
There is a tremendous amount of "art" that is produced for purely commercial reasons. It employs many thousands of people. These roles are definitely threatened by image generators.
Agree that if you are Artist this is not going to be a big concern to you.
lm28469 14 minutes ago [-]
Have you talked to "artists"? In my experience the vast majority say the opposite of what you worded here.
Timpanzee 13 minutes ago [-]
AI isn't a tool for creating art in the same way as a paintbrush or clay. AI is describing a painting you want, then having someone else creating the artwork for you. You aren't doing art in the same way hiring a sculptor isn't doing sculpting.
AI is well on the way to eliminating human made art since the skills to actually make art will be lost to the skill of being able to describe art. You know, since the only thing that matter is reducing costs.
coldtea 16 minutes ago [-]
>Art is about pushing limits of what's possible
That's engineering, if that.
Art isn't, and has never been about that.
williamcotton 10 minutes ago [-]
Sure it has. See the modernism as a whole.
tom1337 29 minutes ago [-]
I'd say these models only exist because we had amazing artworks in the past.
techjamie 15 minutes ago [-]
Ironically we live in a time that, overall, is probably better for artists than the world any of those guys grew up in. People have always valued art but not the artists, and many artists through history, including the famous ones, died broke with their works only posthumously attaining value.
These days, through commissions, art is a much more viable profession than it ever was.
charcircuit 12 minutes ago [-]
We would have tons of great artworks if it existed in the past. The works would be both more numerous and at a higher quality.
ahtihn 21 minutes ago [-]
Would anyone even care about Mona Lisa if the exact same painting was done by a random nobody? It's just a portrait.
coldtea 13 minutes ago [-]
Most people no. Then again most people are idiots barely aware of the world they live in, much less culture.
People who actually care about art, if given a chance to see it, yes.
Of course, it being done by Davinci is not some random fact about the painting - as if a painting is a mere artifact.
__alexs 16 minutes ago [-]
Da Vinci is maybe only the 5th most interesting thing about the Mona Lisa.
vunderba 14 minutes ago [-]
I've only had a brief opportunity to try out NB Pro 2 (`gemini-3.1-flash-image-preview`), so I haven't had a chance to update GenAI Showdown.
Here's some of my captions that tend to trip up even state-of-the-art models.
So far it does feel more iterative than an entirely new leap in terms of capabilities, but I haven't run it through the more multimodal aspects such as editing existing images.
That being said, it actually managed the King Louie jump rope test which surprised me.
LeoPanthera 5 minutes ago [-]
It's notable that this model is less advanced that the previous "Pro" model, and also that the Gemini interface is defaulting all requests to "Fast" even if you've previously changed it to Pro.
I guess even Google is running out of GPUs.
jakub_g 7 minutes ago [-]
Since talking images, are there any AI models that can output real transparent gifs/pngs?
And not a (botched) fake white/gray grid background that is commonly used to visualize transparency?
minimaxir 51 seconds ago [-]
You can output to a plain background and use any number of tools to mask it.
runamuck 13 minutes ago [-]
I saw an item for sale on Ali Express's video and I thought "Wow, they hired some really attractive actors to pitch their little gadget." 30 seconds in, I realized they used GenAI. Not because it looked AI, but because the production values looked too high and professional for the item. I would get in on this if you sell anything online.
someone shared benchmarks that differ my experience tho, so I may be biased
aliljet 46 minutes ago [-]
I really really want to see how these images are starting to form into videos. The stills are clearly getting better and better, but what about when you need the stills to organically conform to a keyed script?
Nano Banana was technically impressive the first time, but after Seedance it's not really. It's all just an internet pollution machine anyway.
rany_ 13 minutes ago [-]
The page looks promising but how can I try it out?
progbits 42 minutes ago [-]
I'm seeing more and more AI video memes and they are getting really good. Still just bunch of short clips, long shots are not working well enough, but typical Hollywood movies have few second cuts anyway so this is almost good enough to make a marvel fanfic.
vessenes 42 minutes ago [-]
the workflow right now would be to take this images, make a sequence of them for key "shots" and send them to an I2V model. LTX-2 is the model the r/stablediffusion folks are playing with right now, but there are a fair few.
vessenes 40 minutes ago [-]
Interesting they get to rev this with the release of a new flash model. I'm speculating part of the distil pipeline includes the image gen stuff; that seems like internal tooling that will pay dividends over time, if true. New frontier model -> automatic new image model. Even if it's just incremental updates, it's good for both the product cadence and compounding improvements.
NitpickLawyer 8 minutes ago [-]
> the distil pipeline
I don't have inside info, but everything we've seen about gemini3.0 makes me think they aren't doing distillation for their models. They are likely training different arch/sizes in parallel. Gemini 3.0-flash was better than 3.0-pro on a bunch of tasks. That shouldn't happen with distillation. So my guess is that they are working in parallel, on different arches, and try out stuff on -flash first (since they're smaller and faster to train) and then apply the learnings to -pro training runs. (same thing kinda happened with 2.5-flash that got better upgrades than 2.5-pro at various points last year). Ofc I might be wrong, but that's my guess right now.
WarmWash 34 minutes ago [-]
The confusion here is dense, 3.1 Flash Image is not 3.1 Flash.
The banana models (image) are a different than the mainline models, but the confusingly leverage the same naming scheme.
pietz 40 minutes ago [-]
I'm officially done with the Nano Banana name. It was fun, but can we go back just calling it Gemini Image?
bonoboTP 31 minutes ago [-]
Name recognition has big value. People remember what an advancement the first banana was. Nowadays it's no longer so unique, ChatGPT's and Grok's image editors are also strong.
minimaxir 40 minutes ago [-]
Google updated it early in AI Studio so I've been experimenting:
- Base pricing for a 1024x1024 image is almost 1.6x what normal Nano Banana is ($0.067 vs. $0.039), however you can now get a 512x512 image for cheaper, or a 4k image for cheaper than four 1k images: https://ai.google.dev/gemini-api/docs/pricing#gemini-3.1-fla...
- Thinking is now configurable between `Minimal` and `High` (was not the case with Nano Banana Pro)
- Safety of the model appears to be increased so typical copyright infringing/NSFW content is difficult to generate (it refused to let me generate cartoon characters having taken psychedelics)
- Generation speed is really slow (2-3min per image) but that may be due to load.
- Prompt adherence to my trickier prompts for Nano Banana Pro (https://minimaxir.com/2025/12/nano-banana-pro/) is much worse, unsurprisingly. For example I asked it to make a 5x2 grid with 10 given inputs and it keeps making 4x3 grids with duplicate inputs.
However, I am skeptical with their marquee feature: image search. Anyone who has used Nano Banana Pro for awhile knows that it will strongly overfit on any input images by copy/pasting the subject without changes which is bad for creativity, and I suspect this implementation appears the same.
Additionally I have a test prompt which exploits the January 2025 knowledge cutoff:
Generate a photo of the KPop Demon Hunters performing a concert at Golden Gate Park in their concert outfits.
That still fails even with Grounding with Google Search and Image Search enabled, and more charitable variants of the prompt.
tl;dr the example images (https://deepmind.google/models/gemini-image/flash/) seem similar to Nano Banana Pro which is indeed a big quality improvement but even relative to base Nano Banana it's unclear if it justifies a "2" subtitle especially given the increased cost.
sheept 32 minutes ago [-]
For your knowledge cutoff test, did it failing mean that it generated a generic "Kpop demon hunter" or it rejected the prompt?
minimaxir 27 minutes ago [-]
Generic "Kpop demon hunter". Nano Banana 2 atleast has fun with it, though.
sync 45 minutes ago [-]
Did gemini-2.5-flash-image get an upgrade as well? I just got the following, which is fascinating, and not something I've seen before:
> I'm sorry, but I cannot fulfill your request as it contains conflicting instructions. You asked me to include the self-carved markings on the character's right wrist and to show him clutching his electromancy focus, but you also explicitly stated, "Do NOT include any props, weapons, or objects in the character's hands - hands should be empty." This contradiction prevents me from generating the image as requested.
My prompts are automated (e.g. I'm not writing them) and definitely have contained conflicting instructions in the past.
A quick google search on that error doesn't reveal anything either
dgtlanml2 41 minutes ago [-]
Wow the article narration with Umbriel is silent after the 6 second mark.
evrenesat 20 minutes ago [-]
I only needed help of this banana boy twice, it managed to disappoint me each time. The most recent one, I was trying different beard and mustache styles on myself, on a photo I imported from my own Google photo gallery, and it consistently rejected me, claiming I'm a public figure. Nobody ever told me that I look like any famous person, so that's googles own bananination. ChatGPT nicely handled the job.
riteshyadav02 37 minutes ago [-]
Would be interesting to see latency vs quality tradeoffs here. Are they targeting consumer-facing generation speed or prioritizing fidelity for professional workflows?
wnevets 19 minutes ago [-]
does it still break images with transparent pixels?
danesparza 7 minutes ago [-]
Is it just me, or is Nano banana not working in Gemini currently?
meowface 38 minutes ago [-]
How does it compare to Nano Banana Pro?
sorenjan 43 minutes ago [-]
Is this a distillation of Nano Banana Pro?
meetpateltech 40 minutes ago [-]
Gemini 3.1 Flash Image is based on Gemini 3 Flash.
Can we now edit the images it spits out? All prior tests in trying to edit AI images has failed miserably and laughably
nightski 34 minutes ago [-]
[flagged]
estearum 29 minutes ago [-]
Quite telling that you think a technology that merely prevents you from passing off an AI-generated image as not-AI-generated makes the model "worthless."
Good!
That's the point! Whatever amazing use case you had in mind is bad and I'm glad SynthID (apparently) makes it impossible.
DalasNoin 32 minutes ago [-]
Why does SynthID make it worthless? it helps other platforms detect this as ai?
zardo 27 minutes ago [-]
If the value is in deception.
csjh 31 minutes ago [-]
What’s the downside of SynthID?
1 hours ago [-]
ge96 11 minutes ago [-]
My naive question, can image generation make something novel eg. "show me a DNA structure that cures cancer" can it do that, or it has to have seen something before to generate it.
Just think we conceptually know what a brushless motor design looks like and it's just pixels. I guess even if it did produce the image we wouldn't know what it means.
minimaxir 3 minutes ago [-]
All image models can generate images that were not in its training dataset, but it can't generate reductive extreme cases like your example.
Rendered at 17:05:07 GMT+0000 (Coordinated Universal Time) with Vercel.
Probably about half of us here remember photos before the cell phone era. They were rare, and special, and you'd have a few photos per YEAR to look back on. The feel of photos back then, was at least 100x stronger than now. They were a special item, could be given as a gift. But once they became freely available that same amount of emotion is now split across many thousands of photos. (not saying this is good or bad, just increased supply reducing value of each item)
With image/art generation the same thing will happen and I can already feel it happening. Things that used to be beautiful or fantastic looking now just feel flat and AI-ish. If claymation scenes can be generated in 1s, and I see a million claymation diagrams a year, then claymation will lose its charm. If I see a million fake Tom Cruise videos, then it oversaturates my desire for desire for all Tom Cruise movies.
What a time to be alive.
(except The Mandalorian)
The obvious ones stand out, but there are so many that are indiscernible without spending lots of time digging through it. Even then there are ones that you can at best guess it's maybe AI gen.
The positive aspect of this advance is that I've basically stopped using social media because of the creeping sense that everything is slop
a lot of these accounts mix old clips with new AI clips
or tag onto something emotional like a fake Epstein file image with your favorite politician, and pointing out its AI has people thinking you’re deflecting because you support the politician
Meanwhile the engagement farmer is completely exempt from scrutiny
Its fascinating how fast and unexpected the direction goes
Soon many real OF models will be out of job when everyone will be able to produce content to their personal taste from a few prompts.
Has this thought process ever worked in real life? I know plenty of seniors who still believe everything that comes out of Facebook, be AI or not, and before that it was the TV, radio, newspapers, etc.
Most people choose to believe, which is why they have a hard time confronting facts.
I suppose if the AI was able to tell me a true and compelling story, I might not even mind so much. I just don't want to be spoon fed drivel for 15 minutes to find it was all complete made up BS.
I think part of the issue with architects and designers today is that they use CAD too much. It's easy to design boxes and basic roof lines in CAD. It's harder to put in curves and more craftsman features. Nano Banana's renders have more organic design features IMO.
Our house is looking great and we're very happy how it's going so far with a lot of the thanks to Nano Banana.
Like... What are your inputs to the model? Empty renders of the space, or more fully decorated views/ photos? Do you have a light harness around this to help you discover the style you like and then stay consistent with it?
Do you find that giving a lot of context around the space you're designing helps (it hasn't in my attempts)?
The "cubism" example seems like it would be a closer fit to something like stained glass or something. I don't think the thing really understands what cubism was all about. Cubist painters were trying to free themselves from the confines of a single integral plane of perspective by allowing themselves to show various parts of the image from different viewpoints, different times, different styles, etc.
The division of the image into geometric shapes is just a by-product of that quest, whereas the examples here have made it the sum total of the whole piece.
This feels to me like an example of how LLMs still don't "understand" what the art means, and are just aping its facade.
You can argue things like code generation are an extension of the engineer wielding it. Image generation just seems like a net negative overall if it’s used at scale.
Edit: By scale, I mean large corporations putting content in front of millions. I understand the appeal for smaller businesses where they probably weren’t going to pay an artist anyway.
When a company sends an email or docu-sign, they don’t want to pay a courier.
Technology supplements or replaces jobs, often reducing costs. This is no different.
It's an ethical conundrum because we're not paying anyone, but we don't have the money to pay anyone, and it's good enough for our budget.
But we're getting used to the process of changing a part of the text in a few seconds without any artist involved and for 0$.
I guess that soon we'll be able to create voice sample from know personalities for a few $ with prices based on the popularity of the artist and some sanity check based on the artist preferences.
My thought is the large corps that could afford it, still won’t because it’s a cost they don’t need to incur. For them it’s not even a moral conundrum.
This is like the last mile for online presence. The average barber out here doesn't use Squarespace, barely knows how to use Facebook and doesn't touch GenAi. But they can still cut your hair pretty well - tech savvyness doesn't have a huge connection to business competence out here.
I'm torn on the scale thing. It definitely seems net negative. But I think we collectively underestimate just how deeply sick the existing thing already is. We're repulsed by image gen at scale because it breaks our expectation that images are at least somewhat based on reality, that they reflect the natural world or what we can really expect from a product, from a company, from the future. But that was already a bad expectation: when's the last time you saw a mcdonalds meal that looked like the advert? Or a sub-30$ amazon product that wasn't a complete piece of shit? Advertisements were already actively malicious fantasies to exploit the way our brains react to pictures. They're just fantasies that required whole teams of humans doing weird bullshit with lighting and photoshop, and I'm not sure that's much better. It was already slop. All the grieving we do about the loss of truth, or the extent to which corps will gleefully spray us with mind-breaking waterfalls of outright lies, I think those ships sailed a long time ago. The disgust, deceit, the rage we feel about genAI slop is the way we should have felt about all commercials since at least the 80s IMO.
This is a good point. My gut reaction is “well at least someone was paid to do it and can continue to keep society/the economy going ”.
I can see the other side where that’s a soulless job. Not sure what’s worse. Soulless job where your skills apply or even less jobs in a competitive industry.
Larian Studios most recently was under fire for this [1]. Like I can see a director going “what would X look like?” and then speeding over to the concept artists for a proper rendition if they liked it. I don’t think this is at scale though. Any large business is just going to get rid of the concept artists.
[1]: https://www.pcgamer.com/games/rpg/baldurs-gate-3-developer-l...
You could easily say the same about anytime computers or robots or automation have taken a job away. We’ve been going down this road for decades.
Now extrapolate to all other artforms. Sculpture seems safe, for now, but only barely so.
Artists aren't doing it for the money. With advanced tools like these they wouldve iterated much faster and created much grander designs.
Art is about pushing limits of what's possible and AI just raises those limits.
Agree that if you are Artist this is not going to be a big concern to you.
AI is well on the way to eliminating human made art since the skills to actually make art will be lost to the skill of being able to describe art. You know, since the only thing that matter is reducing costs.
That's engineering, if that.
Art isn't, and has never been about that.
These days, through commissions, art is a much more viable profession than it ever was.
People who actually care about art, if given a chance to see it, yes.
Of course, it being done by Davinci is not some random fact about the painting - as if a painting is a mere artifact.
Here's some of my captions that tend to trip up even state-of-the-art models.
https://mordenstar.com/other/nb-pro-2-tests
So far it does feel more iterative than an entirely new leap in terms of capabilities, but I haven't run it through the more multimodal aspects such as editing existing images.
That being said, it actually managed the King Louie jump rope test which surprised me.
I guess even Google is running out of GPUs.
And not a (botched) fake white/gray grid background that is commonly used to visualize transparency?
we have user-preference rankings that put NB2 on top: https://arena.ai/leaderboard/text-to-image
- https://hunyuan.tencent.com/image/en?tabIndex=0
- https://seed.bytedance.com/en/seedream5_0_lite
someone shared benchmarks that differ my experience tho, so I may be biased
Nano Banana was technically impressive the first time, but after Seedance it's not really. It's all just an internet pollution machine anyway.
I don't have inside info, but everything we've seen about gemini3.0 makes me think they aren't doing distillation for their models. They are likely training different arch/sizes in parallel. Gemini 3.0-flash was better than 3.0-pro on a bunch of tasks. That shouldn't happen with distillation. So my guess is that they are working in parallel, on different arches, and try out stuff on -flash first (since they're smaller and faster to train) and then apply the learnings to -pro training runs. (same thing kinda happened with 2.5-flash that got better upgrades than 2.5-pro at various points last year). Ofc I might be wrong, but that's my guess right now.
The banana models (image) are a different than the mainline models, but the confusingly leverage the same naming scheme.
- Base pricing for a 1024x1024 image is almost 1.6x what normal Nano Banana is ($0.067 vs. $0.039), however you can now get a 512x512 image for cheaper, or a 4k image for cheaper than four 1k images: https://ai.google.dev/gemini-api/docs/pricing#gemini-3.1-fla...
- Thinking is now configurable between `Minimal` and `High` (was not the case with Nano Banana Pro)
- Safety of the model appears to be increased so typical copyright infringing/NSFW content is difficult to generate (it refused to let me generate cartoon characters having taken psychedelics)
- Generation speed is really slow (2-3min per image) but that may be due to load.
- Prompt adherence to my trickier prompts for Nano Banana Pro (https://minimaxir.com/2025/12/nano-banana-pro/) is much worse, unsurprisingly. For example I asked it to make a 5x2 grid with 10 given inputs and it keeps making 4x3 grids with duplicate inputs.
However, I am skeptical with their marquee feature: image search. Anyone who has used Nano Banana Pro for awhile knows that it will strongly overfit on any input images by copy/pasting the subject without changes which is bad for creativity, and I suspect this implementation appears the same.
Additionally I have a test prompt which exploits the January 2025 knowledge cutoff:
That still fails even with Grounding with Google Search and Image Search enabled, and more charitable variants of the prompt.tl;dr the example images (https://deepmind.google/models/gemini-image/flash/) seem similar to Nano Banana Pro which is indeed a big quality improvement but even relative to base Nano Banana it's unclear if it justifies a "2" subtitle especially given the increased cost.
> I'm sorry, but I cannot fulfill your request as it contains conflicting instructions. You asked me to include the self-carved markings on the character's right wrist and to show him clutching his electromancy focus, but you also explicitly stated, "Do NOT include any props, weapons, or objects in the character's hands - hands should be empty." This contradiction prevents me from generating the image as requested.
My prompts are automated (e.g. I'm not writing them) and definitely have contained conflicting instructions in the past.
A quick google search on that error doesn't reveal anything either
source: https://deepmind.google/models/model-cards/gemini-3-1-flash-...
Good!
That's the point! Whatever amazing use case you had in mind is bad and I'm glad SynthID (apparently) makes it impossible.
Just think we conceptually know what a brushless motor design looks like and it's just pixels. I guess even if it did produce the image we wouldn't know what it means.