InvokeAI v6.13.0 (release candidate 1)
This release of InvokeAI offers support for a variety of new models, including those that are remotely hosted by external providers such as GPT Image. In addition, it provides a host of new and improved features as well as bug fixes.
New Locally-Hosted Models
We've added support for the popular Qwen Image models, both the txt2img and edit versions, as well as Anima.
Qwen Image 2.5
This is a high quality txt2img model that can be run either on your local hardware are on a remotely hosted service (see Hosted Models. We provide starters for the full model (~40 GB!) as well as multiple smaller quantized models. For best results, we recommend 30-40 steps and a CFG of 3-4. We also provide two "turbo" LoRA models, Qwen Image Lightning (4-steps) and Qwen_Image_Lightning (8-steps). These LoRAs provide fast rendering and high accuracy, but loss of diversity. Use a CFG of 1 and a Shift of 3 when applying either of these LoRAs.
Qwen Image Edit 2.5
This is a version of Qwen Image that is optimized for image editing. Add up to three reference images and tell the model on how to combine and alter them using declarative statements such as "Pose the duck on the table". We provide the full model as a starter, along with multiple quantized versions and two Qwen Image Edit Lightning LoRAs for turbo mode.
See Qwen-Image-Edit: Image Editing with Higher Quality and Efficiency for tips on using Qwen Image Edit effectively.
Anima Preview 3
Anima is an anime-themed model created by a collaboration between CircleStone Labs and Comfy Org. It is able to produce a wide variety of illustration styles. There is also an Anima Turbo LoRA that can be used to speed up rendering using a CFG of 1 and 8-12 steps.
You'll find Anima and its support models in the Starter Models tab.
New Externally-Hosted Models
This version of InvokeAI offers the ability to generate using externally-hosted models, such as those provided by OpenAI, Google and cloud-based AI generation hosting services. You "install" these models by providing an API key provided by a service that you subscribe to. Typically image generation services will provide a number of generations for free and then start charging on a pay-as-you-go basis. We support externally-hosted models from OpenAI, and Gemini (Google).
The full release will add support for models from BytePlus, and Alibaba Cloud.
An externally-hosted model works much like a locally-hosted one but with restrictions. In general, you cannot adust steps, scheduler or CFG, and only a limited number of image dimensions are supported. Both txt2img and img2img modes are supported, and the models can be applied to the Canvas and Workflow editor. However inpainting, outpainting, controlnets, LoRAs, regional guidance, and the ability to apply different levels of denoising to img2img are not available. These are limitations imposed by the hosting service and not by InvokeAI.
To get started with externally-hosted models, go to the Models tab and select External Providers. Select the hosting service you wish to use and enter your API key. Click Save and the models supported by this provider will be added to your collection of models.
Models available from OpenAI
- GPT Image 1
- GPT Image 1.5
- GPT Image 1 Mini
- DALL-E 3
Models available from Gemini (Google)
- Gemini 2.5 Flash Image
- Gemini 3 Pro Image Preview
- Gemini 3.1 Flash Image Preview
Model Manager Improvements
This version improves the Model Manager in several ways. When installing a new model, the installation queue display and progress bars have been completely rewritten to be more functional and attractive. A new bulk reidentify action has been implemented, and it is now possible to sort the model listing in multiple ways.
Image Generation Improvements
Users now have access to LLM-powered prompt expansion and image-to-prompt features. You will find two new buttons in the prompt textfield:
- Expand Prompt uses a local TextLLM model to expand brief prompts into detailed image generation prompts. This button will replace a simple prompt with a complex one that is well-suited for producing high quality imagery.
- Image to Prompt Generates descriptive prompts from uploaded images. Use it as a supplement to reference images.
To use these features, you must first install a causal LM model, such as Qwen/Qwen2.5-1.5B-Instruct via the Model Manager (enter this repo name into the HuggingFace tab). The LLaVA vision model needed for image to prompt should already be installed.
In addition, this release corrects many small bugs and performance bottlenecks in image generation.
Image Gallery Improvements
You can now automatically create "virtual" boards that group images according to their characteristics. Right now you can only create virtual boards that group images by their date, but more options are on their way.
Canvas Improvements
There have been several new features added to the Canvas. One frequently-requested feature is the ability to hide and show the preview stage thumbnails that can obscure the bottom of the canvas. Another feature is the addition of a Lasso Tool with freehand and polygon modes for creating inpaint masks and selecting raster image regions.
Also among multiple new features is the ability to apply custom workflows directly to a Canvas image. First create a workflow with an image input field in the Form Builder and a Canvas Output node. Right click on the canvas and select Run Workflow. The workflow will run on the input image and create a new layer containing the output of the Canvas Output node.
Workflow Editor and Nodes
There is now a Custom Node Manager tab in the left-hand panel which allows you to install and manage related groups of nodes called node packs. Node packs can be installed from local files or URLs.
When adding a node to a workflow, you will find that the nodes are now organized by function with an easy to navigate node picker. Another convenient new feature is the workflow connector, a new type of edge that can be used to connect two or more input/output fields in a highly flexible fashion. To start using this feature, right click on the canvas and select Add Connector.
Internally the workflow execution engine has been overhauled and optimized to improve performance and stability. In addition, there are a number of new user-visible features. There are new generation nodes for running external models, new denoising nodes for the Qwen Image and Anima local models, and a node for decoding invisible watermarks in images. The Collect node now supports chaining to enable deterministic iteration over the items in a collection. A new If mode allows for conditional branching of workflows.
Multiuser Features
When running in Multiuser Mode each logged-in user can create private (only visible to the owner), shared (writable by the owner, readable by everyone else) and public (read/write for everyone) image boards. Similarly, users can now create private and shared workflows.
Documentation
The Documentation Site has been completely redesigned and overhauled. However, be aware that the documentation for some of the newer features has not yet been ported over to the new site, and can be found here.
What's Changed in Detail
Below is a log of all changes from the previous version. Many thanks to all the authors who contributed to this release.
Models and Model Manager
- External models (Gemini Nano Banana & OpenAI GPT Image) (#8633) by @CypherNaught-0x in #8884
- Feat[model support]: Qwen Image — full pipeline with edit, generate LoRA, GGUF, quantization, and UI by @lstein in #9000
- feat: add Anima model support by @4pointoh in #8961
- feat: add bulk reidentify action for models (#8951) by @Pfannkuchensack in #8952
- feat: add support for OneTrainer BFL Flux LoRA format by @Pfannkuchensack in #8984
- fix(mm): detect FLUX.2 Klein 9B Base variant via filename heuristic by @Pfannkuchensack in #9011
- fix: correct inaccurate download size estimates in starter models by @Pfannkuchensack in #8968
- Redesign Model Manager Installation Queue by @joshistoast in #8910
- fix(ui) Anima model encoder & vae auto-selection by @kappacommit in #9035
- feat:(model-manager) add sorting capabilities for models by @skunkworxdark in #9024
- fix: remove directory entry sizes from directory_size() to report accurate file totals by @kuishou68 in #9040
- fix(flux2-vae): support FLUX.2 small-decoder VAE variant by @Pfannkuchensack in #9032
- feat(ui): show progress bar animation while models are loading by @Pfannkuchensack in #9058
- feat(mm): add editable source_url field to model config by @Pfannkuchensack in #8997
Image Generation
- Feat(UI): Add LLM-powered prompt expansion and image-to-prompt features by @Pfannkuchensack in #8899
- Fix(frontend): Make ordering of multiple FLUX.2 reference images deterministic by @lstein in #8989
- Upgrade spandrel version by @JPPhoto in #8996
- fix: detect Z-Image LoRAs with transformer.layers prefix by @Pfannkuchensack in #8986
- Fix partial-load device recovery across CLIP, T5, and Qwen text encoders by @JPPhoto in #9034
- feat: add configurable shift parameter for Z-Image by @Pfannkuchensack in #9004
- Feature: Add optional setting to prune queue on startup by @DustyShoe in #8861
- Handle mixed-dtype mismatches in autocast linear and conv wrappers by @JPPhoto in #9006
- fix(ui): FLUX.2 Klein VAE/Qwen3 readiness checks and diffusers source auto-detection by @lstein in #9041
- Fix: Preserve reference image panel state and selection on recall by @ufuksarp in #9010
- Fix: Z-Image LoRA detection for Kohya and ComfyUI formats by @Pfannkuchensack in #9007
- [Fix] remove anima optimized denoise by @dunkeroni in #9072
- fix: SDXL DoRA LoRA fails with enable_partial_loading=true by @Pfannkuchensack in #9063
- fix(flux2): remove inert guidance UI, add Klein 4B Base variant, fix metadata recall by @Pfannkuchensack in #8995
- Align DyPE with the paper as much as possible by @JPPhoto in #8980
- feat(recall): support direct model reference images in recall API by @lstein in #9045
Image Gallery
- feat: Add virtual boards that dynamically group images by date by @Pfannkuchensack in #8971
Canvas
- Feat(Canvas): Add button to hide preview stage thumbnails by @DustyShoe in #8963
- feat(UI): add Inpaint Mask as drag & drop target on canvas by @Pfannkuchensack in #8942
- Fix(Canvas): Retain layer opacity on mode change. by @DustyShoe in #8879
- Feature(UI): Canvas Workflow Integration - Run Workflow on Raster Layer by @Pfannkuchensack in #8665
- Feat(canvas): Add Lasso Tool with Freehand and Polygon modes by @DustyShoe in #8908
- feat(ui): add canvas project save/load (.invk format) by @Pfannkuchensack in #8917
- perf(canvas): detach hidden layers from DOM, throttle onscreen check, increase cache by @Pfannkuchensack in #8939
- feat(canvas): add per-layer lock transparency option by @Pfannkuchensack in #8943
- feat(ui): add canvas snapshot save/restore functionality by @GGSSKK in #8978
Workflow Editor and Nodes
- feat: add Custom Node Manager by @Pfannkuchensack in #9047
- Add chaining to Collect node by @JPPhoto in #8933
- Feature (frontend): Add invisible watermark decoder node. by @lstein in #8967
- Fix workflows info copy focus by @JPPhoto in #9015
- Added
Ifnode and ability to link anAnyoutput to a node input if cardinality matches by @JPPhoto in #8869 graph.pyrefactoring andIfnode optimization by @JPPhoto in #9030- Added workflow connectors by @JPPhoto in #9027
- feat(ui): group nodes by category in add-node dialog by @Pfannkuchensack in #8912
- Prevent stale queue snapshots from regressing workflow completion state by @JPPhoto in #9043
- fix(ui): stabilize workflow node execution state updates by @JPPhoto in #9029
- Fix graph execution state resume after JSON round-trip by @JPPhoto in #9042
- Fix lazy If branch pruning and skipped-parent handling in graph runtime by @JPPhoto in #9079
Multiuser
- Fix (multiuser): Ask user to log back in when security token has expired by @lstein in #9017
- feat(multiuser): Shared/private workflows and image boards in multiuser mode by @lstein in #9018
- fix(multiuser): make preexisting workflows visible after migration by @lstein in #9049
Internationalizaton
- fix(ui): replace all hardcoded frontend strings with i18n translation keys by @Pfannkuchensack in #9013
- ui: translations update from weblate by @weblate in #8985
- ui: translations update from weblate by @weblate in #8992
- ui: translations update from weblate by @weblate in #9028
- ui: translations update from weblate by @weblate in #9036
- ui: translations update from weblate by @weblate in #9051
- ui: translations update from weblate by @weblate in #9075
- ui: translations update from weblate by @weblate in #9078
Thanks to Dmitry Warkentin and Riccardo Giovanetti (@Harvester62) for contributing Russian and Italian translations, respectively.
Documentation
- List Supported Models In Readme by @kappacommit in #9038
- Docs Overhaul by @joshistoast in #8896
- New Documentation Fixes by @joshistoast in #9061
- fix(docs): deployment paths mismatch by @joshistoast in #9067
- fix(docs): urls respect baseurl + better deployment scheduling by @joshistoast in #9069
- fix(docs): deploy targets by @joshistoast in #9070
- fix(docs): address bad links by @joshistoast in #9074
- fix(docs): anticipate more redirects and update more links by @joshistoast in #9076
Miscellaneous
- chore: bump version to 6.12.0.post1 by @lstein in #8990
- Fix SIGINT shutdown during active inference by @JPPhoto in #8993
- Run vitest during frontend build by @JPPhoto in #9022
- feat(frontend): suppress tooltips on touch devices by @lstein in #9001
- fix: patch openapi-typescript enum generation to match OpenAPI schema by @lstein in #9037
- Feat(UI): Replace prompt window resize handle with bottom edge drag handle. by @DustyShoe in #8975
- Feat(UI): Reintroduce middle click to open image in new tab by @DustyShoe in #8998
- chore(CI/CD): Prevent agents from running npm by @lstein in #8930
New Contributors
- @4pointoh made their first contribution in #8961
- @kappacommit made their first contribution in #9038
- @CypherNaught-0x made their first contribution in #8884
- @ufuksarp made their first contribution in #9010
- @kuishou68 made their first contribution in #9040
- @GGSSKK made their first contribution in #8978
Full Changelog: v6.12.0...v6.13.0.rc1