Categories: Tutorials

KOLORS: Innovative Model by Kling Team

Kolors is a text-to-image diffusion-based model developed by Kuaishou Kolors Team and is also the official creator of the KlingAI project. This model has been trained on billions of parameters specifically with text-image pairs.

Source: Kolors Hugging Face Repository

This model is compatible with English and Chinese language released under Apache2.0 license. This means it can be used for research, educational, and commercial purposes but you need to contact their team.

For in-depth information, you can refer to their research paper.

Installation:

1. Install ComfyUI on your local machine.

2. Click “Update All” from the ComfyUI Manager to update ComfyUI.

3. There are two methods to create custom nodes.

 (a) Automatic Method:

Move to ComfyUI Manager and search for “ComfyUI-KwaiKolorsWrapper” labeled by author “Kijai” and click the install button to start custom nodes installation.

Then, just restart ComfyUI to take effect.

(b) Manual method:

Navigate into the “ComfyUI/custom_nodes” folder. Move to the folder address bar and type “cmd” to open the command prompt.

Clone the repository by copying and pasting the command into your command prompt provided below:

git clone https://github.com/kijai/ComfyUI-KwaiKolorsWrapper.git

4. Now, install the requirements files by typing these commands (For regular comfy users):

pip install -r requirements.txt

For ComfyUI portable users, use these commands:

python_embededpython.exe -m pip install -r ComfyUIcustom_nodesComfyUI-KwaiKolorsWrapperrequirements.txt

The respective Kolors model(fp16, 16.5GB) and ChatGLm3 get downloaded automatically into “ComfyUI/models/diffusers/Kolors“.

Workflow Explanation:

1. Get the workflow from your installation folder by navigating into the folder location “ComfyUI-KwaiKolorsWrapper/examples“. It has both the workflows included-

(a) text-to-image

(b) image-to-image

Just drag and drop the workflow into ComfyUI.

2. Recommended settings we used:

Sampling method: Euler

Steps:25

CFG: 5

Resolution:1024 by 1024

3. Load the Kolors model into the “Kolor model” node. 

Set adequate options from the “ChatGLM3 node“.  

Set the text encoder to the FP16 model if you have the minimum 13 GB VRAM, for lower 8-9 GB use quant8, and for quant4 use the lower end 4GBVRAM.

Put prompt into “Kolors Text Encode” node.

4. Click “Queue prompt” to generate images. At the first run, it will take some time to download the dependencies in the background.

Kolors Output1

Kolors Output2

Prompt used: 3d anime style, portrait photo of a girl, nightlife, raining, uhd, 8k

Here is the output with a really impressive result. Hmm, so the model is more intelligent in understanding the context. We only inputted “raining” and the model added a raincoat into the image.

admage

Share
Published by
admage

Recent Posts

1. Controlnet: Advanced Automation Solutions 2. IP Adapter: Seamless Network Integration 3. Inpainting: Enhanced Image Reconstruction 4. Outpainting: Innovative Design Techniques

Whether you are touching up photos, creating digital art, or developing innovative applications. FLUX.1 Tools…

4 hours ago

InstantIR: Restore Your Images

Figuring out the model that can fix your low quality pictures? Now, restoring your low…

1 week ago

Omnigen: Next-Gen Image Generation & Editing Tool

Traditional diffusion models uses various mechanisms for image modification like ControlNet, IP-Adapter, Inpainting, Face detection,…

2 weeks ago

Create Engaging Videos with Mochi1

Mochi 1, an open-source text-to-video diffusion model has been released by Genmo.  Trained with 10…

4 weeks ago

Local Installation of Stable Diffusion 3.5

So, it's finally here. Stable Diffusion 3.5 has been released by StabilityAI on October 22nd,…

1 month ago

Video Depth Mapper: Efficient 3D Depth Mapping Solution

Due to the extreme diversity of video content like motion, camera panning, and length, the…

1 month ago