Stable Diffusion

AI VideoGen: Create Stunning Videos with Wan 2.1 Phantom

If you are an AI video creator, animator, or just passionate about the latest in generative media, ByteDance’s new Phantom Subject2Video framework will make your work easier.  The model is built on top of the WAN 2.1 diffusion model trained…

Expand Video Length with Reduced VRAM Usage

Video generation has always been a resource-intensive task, often requiring powerful GPUs and significant processing time. But what if you could generate high-quality videos on a average level GPU? FramePack, a creative approach that’s changing how we think about next-frame…

Unfiltered Image Generation Tool: HiDream

HiDream another banger after Flux, developed by Vivago AI is making waves and for good reason. It’s a powerful, open-source, text-to-image diffusion model with 17 billion parameters, offering top-tier image quality and prompt adherence that rivals paid subscription models. It’s licensed…

ByteDance UNO: Instant Editing Power

 Whenever you want to do some kind of die hard editing to images, you of course need top level blending skills and patience to get satisfied results. ByteDance has recently introduced Uno (Unity and Novel Output) to solve this problem. An…

1. Motion Control & Style Transfer with Wan2.1 Fun ControlNet

Wan 2.1 Fun ControlNet is a cutting-edge AI model developed by Alibaba Pal, specifically designed for video generation by transferring instant style . It builds upon the Wan 2.1 framework and introduces two powerful models: Fun Control and Inpaint. These…

Flux: Developer vs Swift vs Professional (In-Depth Analysis)

If you have been following the latest developments in text-to-image models, you have probably heard about Flux – the revolutionize TextToImage model from Black Forest Labs that’s been making waves in the AI art community.  Here, we are going to…

WAN2.1 LoRa Model Training on Windows/Linux

Fine-tune your own LoRA model using WAN 2.1 locally is not that difficult. The process will be so similar than that of the other LoRA trainings. We will explain you the step by step process whether you are training locally…

Local Video Installation and Generation (Native/GGUFs)

  Again the new diffusion based video generation model released by AlibabaCloud. Wan2.1 an open-source suite of video foundation models licensed under Apache 2.0. It delivers state-of-the-art performance while remaining accessible on consumer hardware. You can read more information from their…

Empowering Human Connection Through SkyReels Videos

Another diffusion based video generation model is in the open source market. Skyreels, a Human-Centric Video framework fined tuned on HunyuanVideo that provides you Open-Source Leadership, Advanced Facial Animation and Cinematic Lighting and Aesthetics. Now, the problem is that the…

Lumina Image 2.0 – Innovative Text to Image Creation

Lumina Image 2.0 is a powerful text-to-image generation model with 2 billion parameters that leverages features compared with the earlier one. With its flow-based diffusion transformer and Gemma (from Google) as the text encoder, it generates high-quality images from text…

Deepkseek Janus Pro: Optimize Image/Video Generation

While working with Text-to-Image or Text-to-Video workflows, you already know the struggle of getting accurate and high-quality results is not always easy. Now, this can be achieved by  DeepSeek’s Janus Pro Model (an upgraded version of Janus) is here to…

TeaCache: Boost Your ComfyUI Speed by 2x

You have ever stuck slower inference speeds with various image/video generation models. Here, the solution is to use TeachCache. Timestep Embedding Aware Cache (TeaCache), a game-changing, training-free caching technique that optimizes performance across timesteps for diffusion models. Whether you are…