Media is too big
VIEW IN TELEGRAM
I guess this happened a Week after Riker Rick Rolled the ship. With a Special Ending. lol.
https://redd.it/1te17ky
@rStableDiffusion
https://redd.it/1te17ky
@rStableDiffusion
Anima is in process of being added to diffusers
https://github.com/huggingface/diffusers/pull/13732
Hopefully support on major trainers like OneTrainer is coming after this.
With all the respect to diffusion-pipe its bucketing is a headscratcher and I don't really trust all standalone trainers based on kohya-SS after issues reported and do not want a stack of those.
https://redd.it/1te2d2i
@rStableDiffusion
https://github.com/huggingface/diffusers/pull/13732
Hopefully support on major trainers like OneTrainer is coming after this.
With all the respect to diffusion-pipe its bucketing is a headscratcher and I don't really trust all standalone trainers based on kohya-SS after issues reported and do not want a stack of those.
https://redd.it/1te2d2i
@rStableDiffusion
GitHub
Add Anima modular pipeline by rmatif · Pull Request #13732 · huggingface/diffusers
What does this PR do?
Adds modular-only support for Anima, a text-to-image model built on top of the Cosmos Predict2 DiT architecture.
This PR adds:
AnimaModularPipeline and AnimaAutoBlocks
AnimaT...
Adds modular-only support for Anima, a text-to-image model built on top of the Cosmos Predict2 DiT architecture.
This PR adds:
AnimaModularPipeline and AnimaAutoBlocks
AnimaT...
Microsoft lens is less than 4B params. The tendency is less params...
Ok, they have retired it. It was 3.8B IIRC. In any case, it seems there´s this tendency to do smaller and smaller models but they manage to get better and better anyhow.
My 12GB card loves it. Lets keep the good work
https://redd.it/1te4ieu
@rStableDiffusion
Ok, they have retired it. It was 3.8B IIRC. In any case, it seems there´s this tendency to do smaller and smaller models but they manage to get better and better anyhow.
My 12GB card loves it. Lets keep the good work
https://redd.it/1te4ieu
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Pixal3D: Generate high-fidelity 3D assets from a single image. (TencentARC, locally runnable model)
https://huggingface.co/TencentARC/Pixal3D
"Pixal3D generates high-fidelity 3D assets from a single image. Unlike previous methods that loosely inject image features via attention, Pixal3D explicitly lifts pixel features into 3D through back-projection, establishing direct pixel-to-3D correspondences. This enables near-reconstruction-level fidelity with detailed geometry and PBR textures."
Looks like no one mentioned this in the sub, so here's everyone's notification.
Some fast points:
* It's a locally runnable model
* I got it working on an RTX 5090 by yelling "Fix it!" at Claude over and over like Philip J. Fry. (This works on most models by the way, I suggest you try it if you have Claude and want to try local models before Comfy's team gets around to it)
* To my eyes, this looks like a step up from Trellis.2 raw, but don't take my word on that. It has some online demo, give it a go.
Please note that it did take a good amount of time getting creative with the yelling-at-claude part, with me having to make some judgment calls and give it advice about how to proceed. But tenacity paid off for me, and I figure it will pay off for anyone else who cares to put in the effort, at least until someone makes a more broadly available guide.
https://redd.it/1te93yi
@rStableDiffusion
https://huggingface.co/TencentARC/Pixal3D
"Pixal3D generates high-fidelity 3D assets from a single image. Unlike previous methods that loosely inject image features via attention, Pixal3D explicitly lifts pixel features into 3D through back-projection, establishing direct pixel-to-3D correspondences. This enables near-reconstruction-level fidelity with detailed geometry and PBR textures."
Looks like no one mentioned this in the sub, so here's everyone's notification.
Some fast points:
* It's a locally runnable model
* I got it working on an RTX 5090 by yelling "Fix it!" at Claude over and over like Philip J. Fry. (This works on most models by the way, I suggest you try it if you have Claude and want to try local models before Comfy's team gets around to it)
* To my eyes, this looks like a step up from Trellis.2 raw, but don't take my word on that. It has some online demo, give it a go.
Please note that it did take a good amount of time getting creative with the yelling-at-claude part, with me having to make some judgment calls and give it advice about how to proceed. But tenacity paid off for me, and I figure it will pay off for anyone else who cares to put in the effort, at least until someone makes a more broadly available guide.
https://redd.it/1te93yi
@rStableDiffusion
huggingface.co
TencentARC/Pixal3D · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
stable-diffusion-webui-codex v0.3.0-beta is live (now with link 😅)
https://redd.it/1te4zvv
@rStableDiffusion
https://redd.it/1te4zvv
@rStableDiffusion
This media is not supported in your browser
VIEW IN TELEGRAM
Playing with Anima Base 1.0 + Flux.2 Klein 9b + Wan 2.2 (No Audio)
https://redd.it/1teatpv
@rStableDiffusion
https://redd.it/1teatpv
@rStableDiffusion
I built a custom NVENC encoder bridge to split FLUX 2 Models across two GPUs over Ethernet LAN (example: 5090 + laptop 4090 spreading model layers over two machines via Eth = 4.4s per image). Completely bypasses the need for NVLink. Multi GPU in one PC supported, Wifi 6 works very well also.
https://github.com/shootthesound/comfyui-mesh
https://redd.it/1tegs83
@rStableDiffusion
https://github.com/shootthesound/comfyui-mesh
https://redd.it/1tegs83
@rStableDiffusion
GitHub
GitHub - shootthesound/comfyui-mesh: Split FLUX.2 across two GPUs (LAN or same-machine) — NVENC compresses activations live on…
Split FLUX.2 across two GPUs (LAN or same-machine) — NVENC compresses activations live on the wire. Icarus (ComfyUI node) + Daedalus (back-half server). - shootthesound/comfyui-mesh
Sharing my experience with Anima (ComfyUI): great detail, but struggling with multiple characters
Hi everyone, I wanted to share my experience.
Lately I’ve started using the Anima model with ComfyUI, and I have to say I’m really enjoying the results so far. What stands out to me the most is the level of detail, which I’ve found to be particularly strong not only on the characters, but even more on backgrounds and environments. I wasn’t really able to reach the same quality with models like Illustrious or Pony.
Another thing I really like (and honestly find kind of genius) is the possibility to build prompts using a mix of Gelbooru-style tags and natural language descriptions. That hybrid approach works incredibly well for me and feels much more flexible compared to sticking to only one style.
That said, I’ve noticed a limitation: when Anima has to handle more than one character in the scene, the results seem noticeably worse compared to what I could get with Illustrious or Pony.
I’m curious if anyone else has run into the same issue, and if there are specific techniques to better handle multi-character compositions.
I’m also wondering whether there’s any kind of regional prompting or similar workflow that works well with Anima, or if there are alternative approaches to improve consistency when generating multiple characters.
Curious to hear your thoughts and tips!
https://redd.it/1tepgn4
@rStableDiffusion
Hi everyone, I wanted to share my experience.
Lately I’ve started using the Anima model with ComfyUI, and I have to say I’m really enjoying the results so far. What stands out to me the most is the level of detail, which I’ve found to be particularly strong not only on the characters, but even more on backgrounds and environments. I wasn’t really able to reach the same quality with models like Illustrious or Pony.
Another thing I really like (and honestly find kind of genius) is the possibility to build prompts using a mix of Gelbooru-style tags and natural language descriptions. That hybrid approach works incredibly well for me and feels much more flexible compared to sticking to only one style.
That said, I’ve noticed a limitation: when Anima has to handle more than one character in the scene, the results seem noticeably worse compared to what I could get with Illustrious or Pony.
I’m curious if anyone else has run into the same issue, and if there are specific techniques to better handle multi-character compositions.
I’m also wondering whether there’s any kind of regional prompting or similar workflow that works well with Anima, or if there are alternative approaches to improve consistency when generating multiple characters.
Curious to hear your thoughts and tips!
https://redd.it/1tepgn4
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
This media is not supported in your browser
VIEW IN TELEGRAM
Found this in the attic...morphing between unrelated images...
https://redd.it/1terefl
@rStableDiffusion
https://redd.it/1terefl
@rStableDiffusion
ComfyUI Tutorial : LTX 2.3 Style Enhancer LoRA For More Beautiful Cinematic Videos (Res: 1920x1080, Vram: 6 Gb, Gen Time: 20 min)
https://youtu.be/zEckV4j40x4
https://redd.it/1tetekz
@rStableDiffusion
https://youtu.be/zEckV4j40x4
https://redd.it/1tetekz
@rStableDiffusion
YouTube
ComfyUI Tutorial : LTX 2.3 Style Enhancer LoRA Beautiful Cinematic Videos #comfyui #comfyuitutorial
Hello everyone, in this tutorial we explore the style enhance lora for the LTX 2.3 model. This lora model is natural detail enhancer made for users who want a cleaner, more refined look. The cutom workflow helps in generating 5 seconds AI video at full…
What ai was used to make these images or does anyone know a certain prompt?
https://redd.it/1tes221
@rStableDiffusion
https://redd.it/1tes221
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: What ai was used to make these images or does anyone know a certain prompt?
Explore this post and more from the StableDiffusion community