fal.ai just turned one day into a full menu update. In roughly 24 hours, they added new endpoints for top-tier image generation, image editing, text-to-video, reference-driven video, and a new Veo 3.1 extend-video workflow. The headline is not “more models.” The headline is that fal keeps moving toward being a single place you can build a generative media pipeline: generate, edit, extend, and fine-tune, all behind clean endpoints.
If you are only doing one-off generations, you will look at this and see a bunch of names. If you are building a workflow, you will see something else: editing depth, repeatable presets, and more ways to keep outputs consistent across iterations.
Most of the new surface area is editing workflows, not just base generation.
What shipped in this release burst
1) GPT Image 1.5 on fal.ai, both text-to-image and image-to-image
- GPT Image 1.5 text-to-image: https://fal.ai/models/fal-ai/gpt-image-1.5
- GPT Image 1.5 image-to-image: https://fal.ai/models/fal-ai/gpt-image-1.5/edit
The pitch here is simple and useful: high-fidelity images, strong prompt adherence, and edits that keep composition and lighting stable. That last part is the difference between “nice demo” and “usable.” Most teams do not need a brand-new aesthetic every time. They need a reliable way to iterate without the whole image drifting.
I already wrote up my broader thoughts on this family here, including the tradeoffs: ChatGPT Images v1.5 Is Here: Better Editing, Still Not the Model That Beats Nano Banana Pro. This fal update matters because it makes GPT Image 1.5 another drop-in choice inside the same infra a lot of people already use for automations.
2) FLUX.2 [max] endpoints, text-to-image and image-to-image
- FLUX.2 [max] text-to-image: https://fal.ai/models/fal-ai/flux-2-max
- FLUX.2 [max] image-to-image: https://fal.ai/models/fal-ai/flux-2-max/edit
FLUX.2 [max] is the “pay attention” image drop in this batch. The value is edit stability and output quality for work where geometry and lighting consistency matter, like product shots, characters, and iterative creative that needs to match prior frames. If your workflow is “make the first image, then do ten controlled variants,” this is the kind of endpoint you want in the rotation.
3) Wan v2.6 adds two video workflows
- Wan v2.6 text-to-video: https://fal.ai/models/wan/v2.6/text-to-video
- Wan v2.6 reference-to-video: https://fal.ai/models/wan/v2.6/reference-to-video
fal exposing both text-to-video and reference-to-video is the practical part. Text-only video is fun, but reference-driven video is where teams can anchor outputs to an existing asset, style, or approved clip. It reduces the “roll the dice” factor that makes video pipelines expensive and frustrating.
4) Veo 3.1 got new extend-video endpoints, including a fast variant
- Veo 3.1 extend video: https://fal.ai/models/fal-ai/veo3.1/extend-video
- Veo 3.1 fast extend video: https://fal.ai/models/fal-ai/veo3.1/fast/extend-video
Veo 3.1 is not new on fal, but extend-video is. Extending a Veo-created clip up to 30 seconds is a small API on paper and a big deal in workflow terms. It means you can take an approved clip and keep going, instead of regenerating from scratch and hoping the subject and style line up again.
5) Qwen Image Edit 2509 got turned into a full editing stack
- Qwen Image Edit 2509: https://fal.ai/models/fal-ai/qwen-image-edit-2509
- Qwen Image Edit 2509 LoRA endpoint: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora
- Qwen Image Edit 2509 Trainer: https://fal.ai/models/fal-ai/qwen-image-edit-2509-trainer
Then there is the LoRA Gallery, which is the part that looks like someone is paying attention to what people do all day. These are pre-packaged image-to-image edit endpoints that map to common jobs:
- Apply designs onto shirts: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/shirt-design
- Remove lighting and re-light evenly: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/remove-lighting
- Remove unwanted elements while keeping coherence: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/remove-element
- Lighting restoration to remove harsh shadows and hot spots: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/lighting-restoration
- Integrate product with perspective and lighting correction: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/integrate-product
- Create group photos: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/group-photo
- Face to full portrait expansion: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/face-to-full-portrait
- Add background behind white-background objects: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/add-background
- Next scene transitions and progression: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/next-scene
- Multiple angles and camera control: https://fal.ai/models/fal-ai/qwen-image-edit-2509-lora-gallery/multiple-angles
One nuance that people will miss: most of this Qwen block is not a brand-new base model drop. It is new endpoints, presets, and tooling on top of Qwen Image Edit 2509. That is still a good update. Teams rarely need a new base model every week. They need fewer surprises in the edits they already do, plus the option to fine-tune behavior with LoRAs when a generic edit model does not match their style rules.
Why this direction is the real signal
Base generation is the easiest part to demo, and also the least interesting part once you build anything repeatable. The budget goes into iteration and consistency:
- Keeping structure stable while changing style, texture, and text
- Repeatable edits like lighting normalization, background swaps, and object removal
- Video continuity through reference-driven generation and extension
- Customization through LoRA training plus an endpoint you can call the same way every time
This lines up with the broader pattern I see in enterprise adoption: teams stop caring about what wins a screenshot contest and start caring about what they can integrate, monitor, and rerun. I wrote about that shift in a wider sense here: Enterprise AI Adoption in 2025: From Casual Chat to Core Infrastructure.
What to try first, based on the job
- If you need prompt adherence and clean high-fidelity generation, start with GPT Image 1.5.
- If you need studio-style consistency for iterative visuals, try FLUX.2 [max] text-to-image and then move into image-to-image edits.
- If your workflow is mostly “take an existing asset and fix it,” start with the Qwen Image Edit 2509 gallery endpoints before you spend time building your own prompts and guardrails.
- If you are building video workflows that reuse approved footage, the new Veo 3.1 extend video endpoints are the obvious addition.
- If you want reference-driven video as a first-class API call, look at Wan v2.6 reference-to-video.
fal.ai adding more options is nice. fal.ai adding more ways to keep media consistent across edits and extensions is what makes it easier to ship real workflows.