Pixal3D: Generate high-fidelity 3D assets from a single image. (TencentARC, locally runnable model)

https://huggingface.co/TencentARC/Pixal3D

"Pixal3D generates high-fidelity 3D assets from a single image. Unlike previous methods that loosely inject image features via attention, Pixal3D explicitly lifts pixel features into 3D through back-projection, establishing direct pixel-to-3D correspondences. This enables near-reconstruction-level fidelity with detailed geometry and PBR textures."

Looks like no one mentioned this in the sub, so here's everyone's notification.

Some fast points:

* It's a locally runnable model

* I got it working on an RTX 5090 by yelling "Fix it!" at Claude over and over like Philip J. Fry. (This works on most models by the way, I suggest you try it if you have Claude and want to try local models before Comfy's team gets around to it)

* To my eyes, this looks like a step up from Trellis.2 raw, but don't take my word on that. It has some online demo, give it a go.

Please note that it did take a good amount of time getting creative with the yelling-at-claude part, with me having to make some judgment calls and give it advice about how to proceed. But tenacity paid off for me, and I figure it will pay off for anyone else who cares to put in the effort, at least until someone makes a more broadly available guide.

https://redd.it/1te93yi
@rStableDiffusion
stable-diffusion-webui-codex v0.3.0-beta is live (now with link 😅)
https://redd.it/1te4zvv
@rStableDiffusion
I built a custom NVENC encoder bridge to split FLUX 2 Models across two GPUs over Ethernet LAN (example: 5090 + laptop 4090 spreading model layers over two machines via Eth = 4.4s per image). Completely bypasses the need for NVLink. Multi GPU in one PC supported, Wifi 6 works very well also.
https://github.com/shootthesound/comfyui-mesh

https://redd.it/1tegs83
@rStableDiffusion
Sharing my experience with Anima (ComfyUI): great detail, but struggling with multiple characters

Hi everyone, I wanted to share my experience.

Lately I’ve started using the Anima model with ComfyUI, and I have to say I’m really enjoying the results so far. What stands out to me the most is the level of detail, which I’ve found to be particularly strong not only on the characters, but even more on backgrounds and environments. I wasn’t really able to reach the same quality with models like Illustrious or Pony.

Another thing I really like (and honestly find kind of genius) is the possibility to build prompts using a mix of Gelbooru-style tags and natural language descriptions. That hybrid approach works incredibly well for me and feels much more flexible compared to sticking to only one style.

That said, I’ve noticed a limitation: when Anima has to handle more than one character in the scene, the results seem noticeably worse compared to what I could get with Illustrious or Pony.

I’m curious if anyone else has run into the same issue, and if there are specific techniques to better handle multi-character compositions.

I’m also wondering whether there’s any kind of regional prompting or similar workflow that works well with Anima, or if there are alternative approaches to improve consistency when generating multiple characters.

Curious to hear your thoughts and tips!

https://redd.it/1tepgn4
@rStableDiffusion