Built a tool for anyone drowning in huge image folders: HybridScorer
https://redd.it/1sg5paj
@rStableDiffusion
https://redd.it/1sg5paj
@rStableDiffusion
Anima Preview 3 is out and its better than illustrious or pony.
this is the biggest potential "best diffuser ever" for anime kind of diffusers. just take a look at it on civitai try it and you will never want to use illustrious or pony ever again.
https://redd.it/1sgfjbs
@rStableDiffusion
this is the biggest potential "best diffuser ever" for anime kind of diffusers. just take a look at it on civitai try it and you will never want to use illustrious or pony ever again.
https://redd.it/1sgfjbs
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Vibe Code Your First ComfyUI Custom Node Step by Step (Ep12)
https://www.youtube.com/watch?v=oiiCkrX8hq4
https://redd.it/1sfvnnz
@rStableDiffusion
https://www.youtube.com/watch?v=oiiCkrX8hq4
https://redd.it/1sfvnnz
@rStableDiffusion
YouTube
Vibe Code Your First ComfyUI Custom Node Step by Step (Ep12)
Learn how to create your first ComfyUI custom node step by step with AI, even if you have no coding experience. In this episode, I show how to vibe code a working custom node for ComfyUI using tools like Gemini and Claude, how custom nodes are structured…
ACE-Step 1.5 XL Turbo — BF16 version (converted from FP32)
I converted the ACE-Step 1.5 XL Turbo model from FP32 to BF16.
The original weights were \~18.8 GB in FP32, this version is \~9.97 GB — same quality, lower VRAM usage.
🤗 https://huggingface.co/marcorez8/acestep-v15-xl-turbo-bf16
https://redd.it/1sgiqg7
@rStableDiffusion
I converted the ACE-Step 1.5 XL Turbo model from FP32 to BF16.
The original weights were \~18.8 GB in FP32, this version is \~9.97 GB — same quality, lower VRAM usage.
🤗 https://huggingface.co/marcorez8/acestep-v15-xl-turbo-bf16
https://redd.it/1sgiqg7
@rStableDiffusion
huggingface.co
ACE-Step/acestep-v15-xl-turbo · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Qwen 2512 is so Underrated, prompt understanding is really great, only Flux 2 Dev is better. I'm using Q4KS with 4-6 steps and it is fast (20-30 sec per gen), almost as fast as Anima model. It just need that LoRA love from the community.
https://redd.it/1sgnfv0
@rStableDiffusion
https://redd.it/1sgnfv0
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: Qwen 2512 is so Underrated, prompt understanding is really great, only Flux 2 Dev…
Explore this post and more from the StableDiffusion community
Batch caption your entire image dataset locally (no API, no cost)
I was preparing datasets for LoRA / training and needed a fast way to caption a large number of images locally. Most tools I used were painfully slow either in generation or in editing captions.
So made few utily python scripts to caption images in bulk. It uses locally installed LM Studio in API mode with any vision LLM model i.e. Gemma 4, Qwen 3.5, etc.
GitHub: https://github.com/vizsumit/image-captioner
If you’re doing LoRA training dataset prep, this might save you some time.
https://redd.it/1sgk4yu
@rStableDiffusion
I was preparing datasets for LoRA / training and needed a fast way to caption a large number of images locally. Most tools I used were painfully slow either in generation or in editing captions.
So made few utily python scripts to caption images in bulk. It uses locally installed LM Studio in API mode with any vision LLM model i.e. Gemma 4, Qwen 3.5, etc.
GitHub: https://github.com/vizsumit/image-captioner
If you’re doing LoRA training dataset prep, this might save you some time.
https://redd.it/1sgk4yu
@rStableDiffusion
GitHub
GitHub - vizsumit/image-captioner: Image captioning utility scripts for preparing image-caption datasets. Uses LM Studio API with…
Image captioning utility scripts for preparing image-caption datasets. Uses LM Studio API with any vision model. - vizsumit/image-captioner
Outside of training a Lora what do people do to keep a face looking correct when making edits to an image?
Mostly been using Klein and Qwen. As per the title, if you change positions, angles of the person in the starting image too much, they lose the likeliness. I've tried using a close up of the face as a 2nd image reference, and tried inpainting on a second pass. Any other ideas?
There's also a Best Face Swap lora which I thought might work but with the same face, but nope.
https://redd.it/1sgvzed
@rStableDiffusion
Mostly been using Klein and Qwen. As per the title, if you change positions, angles of the person in the starting image too much, they lose the likeliness. I've tried using a close up of the face as a 2nd image reference, and tried inpainting on a second pass. Any other ideas?
There's also a Best Face Swap lora which I thought might work but with the same face, but nope.
https://redd.it/1sgvzed
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community