Anima preview3 was released
For those who has been following Anima, a new preview version was released around 2 hours ago.
Huggingface: https://huggingface.co/circlestone-labs/Anima
Civitai: https://civitai.com/models/2458426/anima-official?modelVersionId=2836417
The model is still in training. It is made by circlestone-labs.
The changes in preview3 (mentioned by the creator in the links above):
Highres training is in progress. Trained for much longer at 1024 resolution than preview2.
Expanded dataset to help learn less common artists (roughly 50-100 post count).
https://redd.it/1sf6w2x
@rStableDiffusion
For those who has been following Anima, a new preview version was released around 2 hours ago.
Huggingface: https://huggingface.co/circlestone-labs/Anima
Civitai: https://civitai.com/models/2458426/anima-official?modelVersionId=2836417
The model is still in training. It is made by circlestone-labs.
The changes in preview3 (mentioned by the creator in the links above):
Highres training is in progress. Trained for much longer at 1024 resolution than preview2.
Expanded dataset to help learn less common artists (roughly 50-100 post count).
https://redd.it/1sf6w2x
@rStableDiffusion
huggingface.co
circlestone-labs/Anima · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Last week in Generative Image & Video
I curate a weekly multimodal AI roundup, here are the open-source image & video highlights from the last week:
GEMS \- Closed-loop system for spatial logic and text rendering in image generation. Outperforms Nano Banana 2 on GenEval2. [GitHub](https://github.com/lcqysl/GEMS) | [Paper](https://arxiv.org/abs/2603.28088)
https://preview.redd.it/16r9ffhd9wtg1.png?width=1456&format=png&auto=webp&s=325ef8a75d23cfa625ac33dfd4d9727c690c11b0
ComfyUI Post-Processing Suite \- Photorealism suite by thezveroboy. Simulates sensor noise, analog artifacts, and camera metadata with base64 EXIF transfer and calibrated DNG writing. GitHub
https://preview.redd.it/mhs0fi5f9wtg1.png?width=990&format=png&auto=webp&s=716128b81d8dd091615d3ede8f0acbcb3d1327a6
CutClaw \- Open multi-agent video editing framework. Autonomously cuts hours of footage into narrative shorts. [Paper](https://arxiv.org/abs/2603.29664) | [GitHub](https://github.com/GVCLab/CutClaw) | [Hugging Face](https://huggingface.co/papers/2603.29664)
https://reddit.com/link/1sfj9dt/video/uw4oz84j9wtg1/player
Netflix VOID \- Video object deletion with physics simulation. Built on CogVideoX-5B and SAM 2. Project | Hugging Face Space
https://reddit.com/link/1sfj9dt/video/1vzz6zck9wtg1/player
Flux FaceIR \- Flux-2-klein LoRA for blind or reference-guided face restoration. [GitHub](https://github.com/cosmicrealm/ComfyUI-Flux-FaceIR)
https://preview.redd.it/05o2181m9wtg1.png?width=1456&format=png&auto=webp&s=691420332c1e42d9511c7d1cbecf305a5d885d67
Flux-restoration \- Unified face restoration LoRA on FLUX.2-klein-base-4B. GitHub
https://preview.redd.it/l69v7cfn9wtg1.png?width=1456&format=png&auto=webp&s=1711dc1321b997d4247e5db0ac8e13ec4e56180b
LTX2.3 Cameraman LoRA \- Transfers camera motion from reference videos to new scenes. No trigger words. [Hugging Face](https://huggingface.co/Cseti/LTX2.3-22B_IC-LoRA-Cameraman_v1)
https://reddit.com/link/1sfj9dt/video/v8jl2nlq9wtg1/player
Honorable Mentions:
Gen-Searcher \- Agentic search image generation across styles. Hugging Face | GitHub
https://preview.redd.it/suqsu3et9wtg1.png?width=1268&format=png&auto=webp&s=8008783b5d3e298703a8673b6a15c54f4d2155bd
OmniVoice \- 600+ language TTS with voice cloning. [Hugging Face](https://huggingface.co/k2-fsa/OmniVoice) | [ComfyUI](https://github.com/Saganaki22/ComfyUI-OmniVoice-TTS)
https://reddit.com/link/1sfj9dt/video/im1ywh7gcwtg1/player
DreamLite \- On-device 1024x1024 image gen and editing in under a second on a smartphone. (I couldnt find models on HF) GitHub
Checkout the full roundup for more demos, papers, and resources.
https://redd.it/1sfj9dt
@rStableDiffusion
I curate a weekly multimodal AI roundup, here are the open-source image & video highlights from the last week:
GEMS \- Closed-loop system for spatial logic and text rendering in image generation. Outperforms Nano Banana 2 on GenEval2. [GitHub](https://github.com/lcqysl/GEMS) | [Paper](https://arxiv.org/abs/2603.28088)
https://preview.redd.it/16r9ffhd9wtg1.png?width=1456&format=png&auto=webp&s=325ef8a75d23cfa625ac33dfd4d9727c690c11b0
ComfyUI Post-Processing Suite \- Photorealism suite by thezveroboy. Simulates sensor noise, analog artifacts, and camera metadata with base64 EXIF transfer and calibrated DNG writing. GitHub
https://preview.redd.it/mhs0fi5f9wtg1.png?width=990&format=png&auto=webp&s=716128b81d8dd091615d3ede8f0acbcb3d1327a6
CutClaw \- Open multi-agent video editing framework. Autonomously cuts hours of footage into narrative shorts. [Paper](https://arxiv.org/abs/2603.29664) | [GitHub](https://github.com/GVCLab/CutClaw) | [Hugging Face](https://huggingface.co/papers/2603.29664)
https://reddit.com/link/1sfj9dt/video/uw4oz84j9wtg1/player
Netflix VOID \- Video object deletion with physics simulation. Built on CogVideoX-5B and SAM 2. Project | Hugging Face Space
https://reddit.com/link/1sfj9dt/video/1vzz6zck9wtg1/player
Flux FaceIR \- Flux-2-klein LoRA for blind or reference-guided face restoration. [GitHub](https://github.com/cosmicrealm/ComfyUI-Flux-FaceIR)
https://preview.redd.it/05o2181m9wtg1.png?width=1456&format=png&auto=webp&s=691420332c1e42d9511c7d1cbecf305a5d885d67
Flux-restoration \- Unified face restoration LoRA on FLUX.2-klein-base-4B. GitHub
https://preview.redd.it/l69v7cfn9wtg1.png?width=1456&format=png&auto=webp&s=1711dc1321b997d4247e5db0ac8e13ec4e56180b
LTX2.3 Cameraman LoRA \- Transfers camera motion from reference videos to new scenes. No trigger words. [Hugging Face](https://huggingface.co/Cseti/LTX2.3-22B_IC-LoRA-Cameraman_v1)
https://reddit.com/link/1sfj9dt/video/v8jl2nlq9wtg1/player
Honorable Mentions:
Gen-Searcher \- Agentic search image generation across styles. Hugging Face | GitHub
https://preview.redd.it/suqsu3et9wtg1.png?width=1268&format=png&auto=webp&s=8008783b5d3e298703a8673b6a15c54f4d2155bd
OmniVoice \- 600+ language TTS with voice cloning. [Hugging Face](https://huggingface.co/k2-fsa/OmniVoice) | [ComfyUI](https://github.com/Saganaki22/ComfyUI-OmniVoice-TTS)
https://reddit.com/link/1sfj9dt/video/im1ywh7gcwtg1/player
DreamLite \- On-device 1024x1024 image gen and editing in under a second on a smartphone. (I couldnt find models on HF) GitHub
Checkout the full roundup for more demos, papers, and resources.
https://redd.it/1sfj9dt
@rStableDiffusion
GitHub
GitHub - lcqysl/GEMS: GEMS: Agent-Native Multimodal Generation with Memory and Skills
GEMS: Agent-Native Multimodal Generation with Memory and Skills - lcqysl/GEMS
A new SOTA local video model (HappyHorse 1.0) will be released in april 10th.
https://redd.it/1sfo3dq
@rStableDiffusion
https://redd.it/1sfo3dq
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: A new SOTA local video model (HappyHorse 1.0) will be released in april 10th.
Explore this post and more from the StableDiffusion community
Built a tool for anyone drowning in huge image folders: HybridScorer
https://redd.it/1sg5paj
@rStableDiffusion
https://redd.it/1sg5paj
@rStableDiffusion
Anima Preview 3 is out and its better than illustrious or pony.
this is the biggest potential "best diffuser ever" for anime kind of diffusers. just take a look at it on civitai try it and you will never want to use illustrious or pony ever again.
https://redd.it/1sgfjbs
@rStableDiffusion
this is the biggest potential "best diffuser ever" for anime kind of diffusers. just take a look at it on civitai try it and you will never want to use illustrious or pony ever again.
https://redd.it/1sgfjbs
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Vibe Code Your First ComfyUI Custom Node Step by Step (Ep12)
https://www.youtube.com/watch?v=oiiCkrX8hq4
https://redd.it/1sfvnnz
@rStableDiffusion
https://www.youtube.com/watch?v=oiiCkrX8hq4
https://redd.it/1sfvnnz
@rStableDiffusion
YouTube
Vibe Code Your First ComfyUI Custom Node Step by Step (Ep12)
Learn how to create your first ComfyUI custom node step by step with AI, even if you have no coding experience. In this episode, I show how to vibe code a working custom node for ComfyUI using tools like Gemini and Claude, how custom nodes are structured…
ACE-Step 1.5 XL Turbo — BF16 version (converted from FP32)
I converted the ACE-Step 1.5 XL Turbo model from FP32 to BF16.
The original weights were \~18.8 GB in FP32, this version is \~9.97 GB — same quality, lower VRAM usage.
🤗 https://huggingface.co/marcorez8/acestep-v15-xl-turbo-bf16
https://redd.it/1sgiqg7
@rStableDiffusion
I converted the ACE-Step 1.5 XL Turbo model from FP32 to BF16.
The original weights were \~18.8 GB in FP32, this version is \~9.97 GB — same quality, lower VRAM usage.
🤗 https://huggingface.co/marcorez8/acestep-v15-xl-turbo-bf16
https://redd.it/1sgiqg7
@rStableDiffusion
huggingface.co
ACE-Step/acestep-v15-xl-turbo · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Qwen 2512 is so Underrated, prompt understanding is really great, only Flux 2 Dev is better. I'm using Q4KS with 4-6 steps and it is fast (20-30 sec per gen), almost as fast as Anima model. It just need that LoRA love from the community.
https://redd.it/1sgnfv0
@rStableDiffusion
https://redd.it/1sgnfv0
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: Qwen 2512 is so Underrated, prompt understanding is really great, only Flux 2 Dev…
Explore this post and more from the StableDiffusion community