Sign In

LLM prompt helper (with FLUX.1 Kontext support)

94

2.2k

56

Updated: Nov 21, 2025

toolcomfyllm

Type

Workflows

Stats

175

0

Reviews

Published

Dec 22, 2024

Base Model

Flux.1 D

Hash

AutoV2
EE90B35DC2
default creator card background decoration
Style Fusion Contest Participant
TikFesku's Avatar

TikFesku

The FLUX.1 [dev] Model is licensed by Black Forest Labs. Inc. under the FLUX.1 [dev] Non-Commercial License. Copyright Black Forest Labs. Inc.

IN NO EVENT SHALL BLACK FOREST LABS, INC. BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH USE OF THIS MODEL.

Starting from version 2.0, the workflow support txt2img, img2img, Inpaint functionality and uses the built-in LLM node

https://github.com/AlexYez/comfyui-timesaver

instead of the external Ollama program. TS_Qwen3_Node node can describe images, translate prompts and enhance prompts.

If your operating system is Windows and you can't install Qwen3_Node dependencies (don’t have a compiler installed), try to download the .whl file from

https://github.com/boneylizard/llama-cpp-python-cu128-gemma3/releases

then close ComfyUI, open the python_embeded folder, type cmd in the address bar, and execute the following command.

.\python.exe -I -m pip install "path to downloaded.whl file"

after installing you can run ComfyUi and install missing custom nodes as normal way.

Edit: If .whl install fails, check your Python version and make sure that .whl was build for this version. If it is still fails, try to open .whl as archive and just extract all folders from archive to python_embeded\Lib\site-packages folder

===Old versions ===============================

This workflow combine power of LLM text models managed by Ollama with Flux image generation. It takes image or text as input, improve prompt or change it according to instructions.

Note: To refresh LLM models list you need to reload browser window by pressing F5 key.

Since 1.8 there is a blue switcher in Generate Image group to enable or disable context support.

Since 1.3 you need to switch blocks on and off and manually copy prompt text between blocks.

Information:

First of all you need to download and install Ollama from

https://ollama.com/

In current workfow we use 2 LLM models:

Img2Img use llava for image tagging and Mistral for manipulations

Combined 1.3 use llava and phi4

Txt2Img 1.2 use only phi4

Txt2Img 1.1 use only Mistral

Before running Comfy you need to download models:

open command prompt from Ollama folder (with ollama.exe) and say

ollama pull llava:7b (if you have 8-12 Vram)

or

ollama pull llava:13b (for 16+ Vram)

and wait for model download and say For img2img and Txt2Img v.1.1

ollama pull mistral-small

For Txt2Img v.1.2 and combined 1.3 use

ollama pull phi4

After download finished start ollama app.exe, wait for tray icon, start Comfy and install missing custom nodes.

If not set, select llava in Ollama Vision node and mistral in Translate and Ollama Generate Advance nodes.

If you plan to give IMG2IMG instructions in other language turn on and use Translate node.

TXT2IMG take as prompt any language

====================

For Redux IP Tools version you need to download 2 models:

Clip Vision -> models\clip_vision

Style model -> models\style_models