Whats the verdict on Sage Attention 3 now? or stick with Sage 2.2?

I use Image Z Turbo, Wan 2.2 and LTX 2.3

I noticed that Sage Attention 3 altered the dress in a video of a dancing woman to a trousers when using LTX 2.3, I switched to Sage 2.2 and also tried disabling it and the issue was fixed

I actually thought it was the GGUF text encoder that causes the dress to turn into a pants but to my surprise it was Sage 3 that was causing it.

I went back to 2.2 only lost a few seconds speed by the quality was like if it' was disabled very good.

https://redd.it/1s73r4e
@rStableDiffusion
This media is not supported in your browser
VIEW IN TELEGRAM
I went from being a total dummy at ComfyUi to generating this I2V using LTX 2.3, I feel so proud of myself.

https://redd.it/1s76eod
@rStableDiffusion
What can you do if your hardware can generate 15,000 token/s?

[https://taalas.com/](https://taalas.com/)

Demo:

[https://chatjimmy.ai/](https://chatjimmy.ai/)

Saw this posted from r/Qwen_AI and r/LocalLLM today. I also remember seeing this from a few years ago when they first published their studies, but completely forgot about it.

Basically instead of inference on a graphics card where models are loaded onto memory, we burn the model into hardware. Remember CDs? It is cheap to build this compare to GPUs, they are using 6nm chips instead of the latest tech, no memories needed! The biggest downside is you can't swap models, there is no flexibility.

Thoughts? Would this making live streaming AI movies, games possible? You can have a MMO where every single npc have their own unique dialog with no delay for thousands of players.

What a crazy world we live in.

https://redd.it/1s77t1e
@rStableDiffusion