Ernie-image-edit still coming ?
I have looked for news about it everywhere but all i heard were roumours, some saying that it’s delayed but other saying it’s not coming anymore. - none of the sides giving any reason or source.
So i guess i’m asking if everybody heard anything on twitter or somewhere else about the model ?
Personally i’m super excited for it and i’ve been looking forward for it’s release. I like the quality of the t2i model lots.
Flux2 klein is pretty much the only edit model i can use and would like an alternative since qwen and fire red are too heavy, can’t get joy image and glm to run. Also there’s long cat image that i installed but seems a little too plastic on what it edits at times
https://redd.it/1t3jzt4
@rStableDiffusion
I have looked for news about it everywhere but all i heard were roumours, some saying that it’s delayed but other saying it’s not coming anymore. - none of the sides giving any reason or source.
So i guess i’m asking if everybody heard anything on twitter or somewhere else about the model ?
Personally i’m super excited for it and i’ve been looking forward for it’s release. I like the quality of the t2i model lots.
Flux2 klein is pretty much the only edit model i can use and would like an alternative since qwen and fire red are too heavy, can’t get joy image and glm to run. Also there’s long cat image that i installed but seems a little too plastic on what it edits at times
https://redd.it/1t3jzt4
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
GitHub: ComfyUI SenseNova U1 Released – Anyone Got It Working Yet for ComfyUI?
Github
https://github.com/smthemex/ComfyUI\_SenseNova\_U1
models/checkpoints
https://huggingface.co/smthem/SenseNova-U1-8B-MoT-Merger-gguf/tree/main
so we seem to have support, has anyone got it work yet for ComfyUI? i wonder do we need to wait for the Comfyui official patch that they implemented it
https://redd.it/1t3ns9g
@rStableDiffusion
Github
https://github.com/smthemex/ComfyUI\_SenseNova\_U1
models/checkpoints
https://huggingface.co/smthem/SenseNova-U1-8B-MoT-Merger-gguf/tree/main
so we seem to have support, has anyone got it work yet for ComfyUI? i wonder do we need to wait for the Comfyui official patch that they implemented it
https://redd.it/1t3ns9g
@rStableDiffusion
GitHub
GitHub - smthemex/ComfyUI_SenseNova_U1: enseNova-U1: Unifying Multimodal Understanding and Generation with NEO-Unify Architecture
enseNova-U1: Unifying Multimodal Understanding and Generation with NEO-Unify Architecture - smthemex/ComfyUI_SenseNova_U1
A new open weights image model appears in ArtificialAnalysis. Outperforming Flux.2 Pro and Z Image Turbo.
https://redd.it/1t3rd6m
@rStableDiffusion
https://redd.it/1t3rd6m
@rStableDiffusion
Mickmumpitz has knocked it out of the park with this LTX2.3 and Klein movie-making workflow
https://www.youtube.com/watch?v=0mT4p86ZxGQ&t
https://redd.it/1t3og0x
@rStableDiffusion
https://www.youtube.com/watch?v=0mT4p86ZxGQ&t
https://redd.it/1t3og0x
@rStableDiffusion
YouTube
Generate ENTIRE AI MOVIES with this NEW METHOD! [FREE & LOCAL]
I built a free, fully local AI movie pipeline that lets you create entire short films shot by shot.
If you like my work, please consider supporting me on Patreon:
https://www.patreon.com/Mickmumpitz
Follow me on Twitter: https://twitter.com/mickmumpitz
…
If you like my work, please consider supporting me on Patreon:
https://www.patreon.com/Mickmumpitz
Follow me on Twitter: https://twitter.com/mickmumpitz
…
OneTrainer now supports Ernie LoRA
It has presets for 16GB VRAM and 8GB VRAM. Hopefully they add full fine tuning support too.
Ernie Image Hugging Face repo.
OneTrainer Github repo.
https://redd.it/1t3u8fn
@rStableDiffusion
It has presets for 16GB VRAM and 8GB VRAM. Hopefully they add full fine tuning support too.
Ernie Image Hugging Face repo.
OneTrainer Github repo.
https://redd.it/1t3u8fn
@rStableDiffusion
huggingface.co
baidu/ERNIE-Image · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
This media is not supported in your browser
VIEW IN TELEGRAM
Load Video UI - Custom Node to Trim, Resize, and Preview Videos in Realtime
https://redd.it/1t3x1ec
@rStableDiffusion
https://redd.it/1t3x1ec
@rStableDiffusion
It's the 24th century. How is there still no actually good porn model?
https://redd.it/1t445jw
@rStableDiffusion
https://redd.it/1t445jw
@rStableDiffusion
Converting 2D animations to 3D with LTX 2.3 Lora
https://www.youtube.com/watch?v=g88lmbYmZWs
https://redd.it/1t4a0r5
@rStableDiffusion
https://www.youtube.com/watch?v=g88lmbYmZWs
https://redd.it/1t4a0r5
@rStableDiffusion
YouTube
2D to 3D Animations
This is my first attempt at converting 2D animations to 3D. These clips are the result of an on going experiment.
All of my AI demos:
https://www.youtube.com/playlist?list=PLe3OBqR7FeRhZM6SNoIWibQ1PA2JREYtL
All of my AI demos:
https://www.youtube.com/playlist?list=PLe3OBqR7FeRhZM6SNoIWibQ1PA2JREYtL
My LTX 2.3 LoRA Training Journey: Fighting for VRAM even with a 5090
I recently completed a training run for an LTX 2.3 LoRA and wanted to share my settings and findings for those working with similar hardware. I’m running an RTX 5090 with 32GB of VRAM.
1. Tooling & Troubleshooting
AI-Toolkit: I initially tried using AI-Toolkit, but it was a frustrating experience. It suffered from frequent, random freezes with no clear way to debug or recover.
Official Trainer: I eventually switched to the official Trainer scripts. Since the official scripts can be a bit finicky to set up, I used AI agents like Claude to help debug and refine the scripts. This made the transition much smoother and allowed me to get the environment running properly.
2. VRAM & Stability (Avoiding OOM)
To fit the training within 32GB VRAM, a few adjustments were necessary:
Disable Audio Module: This is a mandatory step to prevent Out of Memory (OOM) errors.
Resolution: I settled on 512x512x49. Anything beyond these dimensions proved unstable on my setup.
Other Settings: Followed the official recommended configurations.
3. Performance Metrics
Speed: \~0.58 steps/second.
Total Duration: 1500 steps took approximately 40 minutes.
https://preview.redd.it/ktmt9cljoazg1.png?width=1039&format=png&auto=webp&s=d2ac1f8234c5d822ffe0f479ca9937a1bf1ce3cd
4. Results & Conclusion
The primary goal of this LoRA was to capture specific repeating motions in 2D animation.
The results were very satisfying. While the base LTX model didn't naturally produce these specific movements, adding the LoRA successfully introduced the intended motion patterns. Interestingly, even though I trained at a lower resolution/frame count (512px, 49 frames), the LoRA generalized perfectly to high-resolution inference at 121 frames.
https://redd.it/1t4bbsi
@rStableDiffusion
I recently completed a training run for an LTX 2.3 LoRA and wanted to share my settings and findings for those working with similar hardware. I’m running an RTX 5090 with 32GB of VRAM.
1. Tooling & Troubleshooting
AI-Toolkit: I initially tried using AI-Toolkit, but it was a frustrating experience. It suffered from frequent, random freezes with no clear way to debug or recover.
Official Trainer: I eventually switched to the official Trainer scripts. Since the official scripts can be a bit finicky to set up, I used AI agents like Claude to help debug and refine the scripts. This made the transition much smoother and allowed me to get the environment running properly.
2. VRAM & Stability (Avoiding OOM)
To fit the training within 32GB VRAM, a few adjustments were necessary:
Disable Audio Module: This is a mandatory step to prevent Out of Memory (OOM) errors.
Resolution: I settled on 512x512x49. Anything beyond these dimensions proved unstable on my setup.
Other Settings: Followed the official recommended configurations.
3. Performance Metrics
Speed: \~0.58 steps/second.
Total Duration: 1500 steps took approximately 40 minutes.
https://preview.redd.it/ktmt9cljoazg1.png?width=1039&format=png&auto=webp&s=d2ac1f8234c5d822ffe0f479ca9937a1bf1ce3cd
4. Results & Conclusion
The primary goal of this LoRA was to capture specific repeating motions in 2D animation.
The results were very satisfying. While the base LTX model didn't naturally produce these specific movements, adding the LoRA successfully introduced the intended motion patterns. Interestingly, even though I trained at a lower resolution/frame count (512px, 49 frames), the LoRA generalized perfectly to high-resolution inference at 121 frames.
https://redd.it/1t4bbsi
@rStableDiffusion
Badass professional workflow - How High-Effort AI Usage Looks
https://youtu.be/--LJZeuN2PE?si=aps7FTS480hVcavu
The video shows how to create the initial and final frames of an animation, starting from the manual creation of an original robot to the creation of environments and 3D meshes to guide the various AI steps.
https://redd.it/1t49nyt
@rStableDiffusion
https://youtu.be/--LJZeuN2PE?si=aps7FTS480hVcavu
The video shows how to create the initial and final frames of an animation, starting from the manual creation of an original robot to the creation of environments and 3D meshes to guide the various AI steps.
https://redd.it/1t49nyt
@rStableDiffusion
YouTube
Gen AI Workflow: How to Actually Maintain Creative Control
In this video, I am going to be demonstrating a workflow that is an alternative to a lot of the AI videos you see online that make big claims, but in reality are just simple text-prompt-to-video generation that anyone with a keyboard can create.
I will show…
I will show…