FLUX.2: Frontier Visual Intelligence
57 points
1 hour ago
| 8 comments
| bfl.ai
| HN
spyder
2 minutes ago
[-]
Great, especially that they still have an open-weight variant of this new model too. But what happened to their work on their unreleased SOTA video model? did it stop being SOTA, others got ahead, and they folded the project, or what? YT video about it: https://youtu.be/svIHNnM1Pa0?t=208 They even removed the page of that: https://bfl.ai/up-next/
reply
AmazingTurtle
14 minutes ago
[-]
I ran "family guy themed cyberpunk 2077 ingame screenshot, peter griffin as main character, third person view, view of character from the back" on both nano banana pro and bfl flux 2 pro. The results were staggering. The google model aligned better with the cyberpunk ingame scene, flux was too "realistic"
reply
minimaxir
39 minutes ago
[-]
Text encoder is Mistral-Small-3.2-24B-Instruct-2506 (which is multimodal) as opposed to the weird choice to use CLIP and T5 in the original FLUX, so that's a good start albeit kinda big for a model intended to be open weight. BFL likely should have held off the release until their Apache 2.0 distilled model was released in order to better differentiate from Nano Banana/Nano Banana Pro.

The pricing structure on the Pro variant is...weird:

> Input: We charge $0.015 for each megapixel on the input (i.e. reference images for editing)

> Output: The first megapixel is charged $0.03 and then each subsequent MP will be charged $0.015

reply
woadwarrior01
22 minutes ago
[-]
> BFL likely should have held off the release until their Apache 2.0 distilled model was released in order to better differentiate from Nano Banana/Nano Banana Pro.

Qwen-Image-Edit-2511 is going to be released next week. And it will be Apache 2.0 licensed. I suspect that was one of the factors in the decision to release FLUX.2 this week.

reply
minimaxir
16 minutes ago
[-]
Fair point.
reply
throwaway314155
23 minutes ago
[-]
> as opposed to the weird choice to use CLIP and T5 in the original FLUX

This method was used in tons of image generation models. Not saying it's superior or even a good idea, but it definitely wasn't "weird".

reply
beernet
34 minutes ago
[-]
Nice catch. Looks like engineers tried to take care of the GTM part as well and (surprise!) messed it up. In any case, the biggest loser here is Europe once again.
reply
542458
38 minutes ago
[-]
> Run FLUX.2 [dev] on GeForce RTX GPUs for local experimentation with an optimized fp8 reference implementation of FLUX.2 [dev], created in collaboration with NVIDIA and ComfyUI.

Glad to see that they're sticking with open weights.

That said, Flux 1.x was 12B params, right? So this is about 3x as large plus a 24B text encoder (unless I'm misunderstanding), so it might be a significant challenge for local use. I'll be looking forward to the distill version.

reply
minimaxir
18 minutes ago
[-]
Looking at the file sizes on the open weights version (https://huggingface.co/black-forest-labs/FLUX.2-dev/tree/mai...), the 22B text encoder is 48GB, the generation model itself is 64GB, which roughly tracks with it being the 32B parameters mentioned.

Downloading over 100GB of model weights is a tough sell for the local-only hobbyists.

reply
xnx
41 minutes ago
[-]
Good to see there's some competition to Nano Banana Pro. Other players are important for keeping the price of the leaders in check.
reply
beernet
36 minutes ago
[-]
Oh, looks like someone had to release something very quickly after Google came for their lunch. Their little 15 mins is over already for BFL as it seems.
reply
whywhywhywhy
29 minutes ago
[-]
comparing a closed image model to an open one is like comparing a compiled closed source app to raw source code.

it's pointless to compare in pure output when one is set in stone and the other can be built upon.

reply
beernet
4 minutes ago
[-]
Did you guys even check the licence? Not sure what is "open source" about that. Open weights at the very best, yet highly restrictive
reply
timmmmmmay
26 minutes ago
[-]
yeah except I can download this and run it on my computer, whereas Nano Banana is a service that Google will suddenly discontinue the instant they get bored with it
reply
echelon
17 minutes ago
[-]
> Launch Partners

Wow, the Krea relationship soured? These are both a16z companies and they've worked on private model development before. Krea.1 was supposed to be something to compete with Midjourney aesthetics and get away from the plastic-y Flux models with artificial skin tones, weird chins, etc.

This list of partners includes all of Krea's competitors: HiggsField (current aggregator leader), Freepik, "Open"Art, ElevenLabs (which now has an aggregator product), Leonardo.ai, Lightricks, etc. but Krea is absent. Really strange omission.

I wonder what happened.

reply
eric-p7
44 minutes ago
[-]
Yes yes very impressive.

But can it still turn my screen orange?

reply