flux ip adapter v2
XLabs-AIIntroduction
FLUX-IP-ADAPTER-V2 is an image-to-image model designed by XLABS AI, utilizing the FLUX.1-dev model as its base. It is designed for use with ComfyUI and supports image generation workflows.
Architecture
The IP Adapter employs a resolution of 512x512 for 150k training steps and 1024x1024 for 350k steps. It maintains the aspect ratio during training, enhancing its usability for image-to-image tasks. The model integrates with the diffusers
library and supports an English language interface.
Training
The model was trained using datasets such as CaptionEmporium/coyo-hd-11m-llavanext and CortexLM/midjourney-v6. The training process involved maintaining aspect ratios and varying resolutions to optimize the image generation capabilities.
Guide: Running Locally
To run the model locally, follow these steps:
-
Clone the Repository:
Navigate to theComfyUI/custom_nodes
directory and clone the repositoryx-flux-comfyui
. Ensure the file path isComfyUI/custom_nodes/x-flux-comfyui/*
. -
Setup Environment:
In the cloned directory, executepython setup.py
to set up the environment. -
Update Repository:
Usegit pull
to update or reinstall as necessary. -
Download Necessary Models:
Download the Clip-L model.safetensors from OpenAI's VIT CLIP large model and place it inComfyUI/models/clip_vision/*
.
Obtain the IPAdapter from Hugging Face and place it inComfyUI/models/xlabs/ipadapters/*
. -
Run Inference:
Use the Flux Load IPAdapter and Apply Flux IPAdapter nodes, select the appropriate CLIP model, and start generating images. -
Testing:
If results are suboptimal, adjust the IP strength settings and review example workflows provided in the repository.
For optimal performance, consider using cloud GPUs from providers like AWS, Google Cloud, or Azure.
License
The model is distributed under the FLUX.1-dev Non-Commercial License. More details can be found here.