Announcement

Accelerating FLUX.2 [flex]: Now Design x3 Faster

Jan 29, 2026

John Rachwan

John Rachwan

Cofounder & CTO

Quentin Sinig

Nils Fleischmann

ML Research Engineer

Quentin Sinig

Johanna Sommer

ML Research Engineer

Quentin Sinig

Sara Han Díaz

DevRel Engineer

David

David Berenstein

ML & DevRel Engineer

Bertrand Charpentier

Bertrand Charpentier

Cofounder, President & Chief Scientist

When you’re designing, you need iteration. Iteration needs speed. FLUX.2 [flex] from Black Forest Labs delivers excellent results at text rendering and custom typography, ranking top in human preference rankings, but until recently, it came with a trade-off: latency.

We’ve been in touch with the Black Forest Labs team for a long time—well before this release. Long before FLUX.2 [flex], we were already optimizing the open-weight models they were publishing. Not because it was on a roadmap, but because we genuinely respected the work. BFL has consistently pushed the boundaries of generative image models, and whenever a new model dropped, our instinct was simple: “let’s see how far we can take it.”

Today, that long-running collaboration becomes official.

We worked together with Black Forest Labs to accelerate FLUX.2 [flex] to the next level. Now, FLUX.2 [flex] takes just ~13s per image, down from 48s before. It is ~3x faster!

This isn’t just an optimization win. It’s the result of two teams with very different, but highly complementary obsessions.

  • Black Forest Labs focuses on building flagship, high-quality foundation models that push the boundaries of what’s possible in image generation.

  • At Pruna, our obsession is different: making those models fast enough to be used in real-time creative workflows, without sacrificing quality.

When these two expectations meet, the impact is multiplicative.

You can try the official accelerated Black Forest Labs endpoint here: https://bfl.ai/models/flux-2

flux2flex-examples

Why accelerate FLUX.2 [flex]?

We accelerate FLUX.2 [flex] to be faster and more efficient, reducing the time for image generation and editing.

Speed here isn’t about shaving off a few seconds. Going from ~48 seconds to ~13 seconds completely changes how people use the model. It stops feeling like a batch job and starts feeling interactive.

That means you can try more edits or style adjustments to texts, compare more typography variations, and refine toward the right result, without lowering the quality bar. This is especially critical for typography-heavy use cases, where iteration is everything and waiting a full minute for feedback simply breaks creative flow.

Ideal for:

  • Designers and creative teams seeking precise creative control.

  • Marketing and product teams creating on-brand, typography-heavy assets.

  • Developers working on product UIs

Use it for:

Brand and graphic design

Create beautiful and precise on-brand graphic design assets with clean, readable typography

flux2flex-brand-graphic

Accurate long text

Generate your works with text. It adapts seamlessly to any style or layout you need. Modern, classic, handwritten, or dynamic 3D lettering.

flux2flex-long-text

Product UI prototyping with text components

Rapidly prototype product UIs with rich text components, accurate typography, and production-ready visuals.

flux2flex-produc-ui

How can you use FLUX.2 [flex]?

The optimized model maintains all the capabilities of the original FLUX.2 [flex]:

  • Control over guidance scale

  • Ability to render text with high accuracy.

  • Supports up to 8 reference images for complex multi-image editing

More Speed. Same Quality

This is where the collaboration really matters. The objective wasn’t speed at any cost, but making FLUX.2 [flex] faster while keeping the quality that made it stand out in the first place.

After optimizing the model, we measured latency in an H200. In this setting, we reduced generation time from 48.53s per image to 13.11s, achieving 3.7x speedup. For image editing, we also reduced the time per image from 104.79 s to 27.63 s, resulting in a 3.85x improvement.

These results make it a very good, cost-efficient option for typography-heavy image generation, resulting in a better design experience and lower energy consumption.

GPUs

Mode

Output Size

Input Size

Base - Seconds/Image

Optimized - Seconds/Image

1

t2i

1024x1024

-

48.53

13.11

2

t2i

1024x1024

-

26.97

7.75

4

t2i

1024x1024

-

15.02

4.98

1

i2i

1024x1024

1024x1024

104.79

27.63

2

i2i

1024x1024

1024x1024

57.41

15.92

4

i2i

1024x1024

1024x1024

31.40

9.94

Alongside latency, we evaluated output quality by comparing the optimized FLUX.2 [flex] with the top five models on an artificial analysis for the text & typography category across two text rendering benchmarks: LongTextBench and OneIG (text rendering subset). For the benchmarks, we used text word accuracy, which measures the % of errors relative to the total number of words. The table below provides an overview of the costs, speeds, and metric scores for each metric.

Model

Cost

Generation Speed

Long Text Bench

OneIG Text Rendering

FLUX.2 [flex] fast

$0.050

14.09s

0.8806

0.7666

FLUX.2 [max]

$0.070

38.20s

0.8458

0.8230

FLUX.2 [flex]

$0.060

49.09s

0.8681

0.7917

Nano Banana Pro

$0.150

23.05s

0.8820

0.6843

Imagen 4 Ultra

$0.060

11.14s

0.6049

0.5661

GPT Image 1.5 (high)

$0.136

45.71s

0.9152

0.7675

Additionally, we’ve created some Pareto Front plots that emphasize that the optimized FLUX.2 [flex] is leading at all fronts in terms of speed vs quality for each benchmark, and it does so at the lowest overall cost of $0.05 per image!

flux2flex-benchmarking

The numbers don’t lie and show how performant and quick the optimized model is, but let’s take a closer look at some of the images we’ve generated for each of the benchmarks.

fluex2flex-comparison

As you can see, the results show that the speed improvements do not come at the expense of noticeable quality loss. FLUX.2 [flex] is the cheapest high-quality model on the market and also dominates the Pareto front in terms of speed-to-quality.

Try it now

FLUX.2 [flex] is now ideal to rapidly experiment with typography styles, copywriting ideas, and custom styles. Don’t let projects requiring highly accurate text rendering slow you down.

Try it in the Black Forest Labs endpoint: https://bfl.ai/models/flux-2

For us, this release is also a milestone. It’s the result of weeks of working side by side with the Black Forest Labs team: fast iterations, hands-on engineering, and the kind of constructive chaos that happens when two teams deeply care about what they’re building.

And if you’re a model creator facing similar challenges—great quality, but latency holding it back—we’re always eager to push further. Optimizing tough models is what we enjoy most.

Curious what Pruna can do for your models?

Whether you're running GenAI in production or exploring what's possible, Pruna makes it easier to move fast and stay efficient.

Curious what Pruna can do for your models?

Whether you're running GenAI in production or exploring what's possible, Pruna makes it easier to move fast and stay efficient.

Curious what Pruna can do for your models?

Whether you're running GenAI in production or exploring what's possible, Pruna makes it easier to move fast and stay efficient.