Files
ComfyUI/custom_nodes/x-flux-comfyui/Guide.md
jaidaken f09734b0ee
Some checks failed
Python Linting / Run Ruff (push) Has been cancelled
Python Linting / Run Pylint (push) Has been cancelled
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.10, [self-hosted Linux], stable) (push) Has been cancelled
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.11, [self-hosted Linux], stable) (push) Has been cancelled
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.12, [self-hosted Linux], stable) (push) Has been cancelled
Full Comfy CI Workflow Runs / test-unix-nightly (12.1, , linux, 3.11, [self-hosted Linux], nightly) (push) Has been cancelled
Execution Tests / test (macos-latest) (push) Has been cancelled
Execution Tests / test (ubuntu-latest) (push) Has been cancelled
Execution Tests / test (windows-latest) (push) Has been cancelled
Test server launches without errors / test (push) Has been cancelled
Unit Tests / test (macos-latest) (push) Has been cancelled
Unit Tests / test (ubuntu-latest) (push) Has been cancelled
Unit Tests / test (windows-2022) (push) Has been cancelled
Add custom nodes, Civitai loras (LFS), and vast.ai setup script
Includes 30 custom nodes committed directly, 7 Civitai-exclusive
loras stored via Git LFS, and a setup script that installs all
dependencies and downloads HuggingFace-hosted models on vast.ai.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-09 00:56:42 +00:00

122 lines
3.3 KiB
Markdown

# Guide
# Installing
First of all, you should install ComfyUI and [ComfyUI Manager](https://github.com/ltdrdata/ComfyUI-Manager).
After that, launch ComfUI.
In the right panel you can find “Manager” button, click on.
![manager.png](guide/manager.png)
You will see big window, click on “Custom Nodes Manager”.
![manager_menu.png](guide/manager_menu.png)
Go to search field, and start typing “x-flux-comfyui”,
![search.png](guide/search.png)
Click “install” button.
![download.png](guide/download.png)
Now, you have access to X-Labs nodes, you can find it in “XLabsNodes” category.
![nodes.png](guide/manager.png)
# Flux Installing
There is complete guide by Comfyanonimus [Guide](https://comfyanonymous.github.io/ComfyUI_examples/flux/)
Important! You should use flux dev, not schnell, if your computer is capable of this, it is better to use fp8 or bf16 (default).
# Nodes
## XLabs Sampler:
![sampler.png](guide/sampler.png)
Node inputs:
model: get FLUX diffusion model (from unet loader)
conditioning & neg_conditioning: input prompts after T5 and clip models (clip only allowed, but you should know, that you will not utilize about 40% of flux power, so use dual text node)
latent_image: latent input for flux, may be empty latent or encoded with FLUX AE (VAE Encode) image (for image-to-image using)
controlnet_condition: input for XLabs-AI ControlNet conditioning
Output:
latent: FLUX latent image, should be decoded with VAE Decoder to get image
Parameters:
noise_seed, control_after_generate: controls random generator
steps: how many denoise steps will diffusion go
timestep_to_start_cfg: how many steps diffusion will do before start to use negative sampling and cfg
true_gs: true cfg scale, will be used after first “**timestep_to_start_cfg” steps**
image_to_image_strength: how much will the original image affect to the output
denoise_strength: how many noise will remain
## Load Flux LoRA
![lora.png](guide/lora.png)
Gets flux model on input and outputs model modified with XLabs LoRA (strength of LoRA and name are selected by the parameters)
## Load Flux ControlNet
![loadcnet.png](guide/loadcnet.png)
Gets controlnet name and returns flux controlnet.
## Apply Flux ControlNet
![manager.png](guide/manager.png)
Gets controlnet, image and strength as parameter. Returns controlnet condition for XLabs Sampler.
## Load Flux IPAdapter
![loadip.png](guide/loadip.png)
Gets IP Adapter name, CLIP ViT model and on which device it will work. Choose CUDA only if you have enough VRAM. Return Flux IPAdapter.
## Apply Flux IPAdapter
![applyip.png](guide/applyip.png)
Gets FLUX model, IP Adapter and image. Returns modified model. Strength of IP Adapter comes from parameter.
## Apply Advanced IPAdatper
![advip.png](guide/advip.png)
Like common one, but has more strength parameters.
# Models
We use custom folder for LoRAs, ControlNets and IPAdapters, the folders contains in `models\xlabs`.
LoRAs goes to `ComfyUI\xlabs\loras`.
ControlNets goes to `ComfyUI\xlabs\controlnets`.
IPAdapters goes to `ComfyUI\xlabs\ipadapters`.
An IPAdapter requires a CLIP VIT. We currently use Open-AI Clip ViT Large. You can find it here.
[CLIP ViT model](https://huggingface.co/openai/clip-vit-large-patch14).
Download model.safetensors, rename it as you want to (but .safetensors is required). And put it to `ComfyUI\models\clip-vision\`.