MIDI: Multi-Instance Diffusion for Single Image to 3D Scene Generation
https://redd.it/1l7jf0z
@rStableDiffusion
https://redd.it/1l7jf0z
@rStableDiffusion
This media is not supported in your browser
VIEW IN TELEGRAM
Comparison Video between Wan 2.1 and Google Veo 2 of 2 female spies fighting a man enemy agent. This is the first time I have tried 2 against 1 in a fight. This a first generation for each. Prompt was basically describing the female agents by color of clothing for the fighting moves.
https://redd.it/1l7n8k2
@rStableDiffusion
https://redd.it/1l7n8k2
@rStableDiffusion
🚀 Everlyn.app – Fast Image/Video Gen with Motion Control, 30s Length, and Free Images (Now Live)
Hey folks!
We just launched [Everlyn.app](http://Everlyn.app) — a new platform for video generation that is fast, powered by our newly developed tech in collaboration with world-class professors, and built with an intuitive UI. You can generate high-quality images and videos up to 30 seconds, add optional image input, use our intelligent prompt enhancement, and control the motions.
Key Features:
* ⚡ Fast inference (typically under 30s)
* 🎬 Long videos (up to 30s, multi-paragraph prompts supported)
* 📸 Free image generation (unlimited, watermark-free)
* 🎯 Fine-grained motion control
* 🤖 AI-powered prompt enhancement
💬 Since I’ve learned so much from this community and friends here, I’d love to give back. If you leave your email in the comments, I’ll personally send you 50 free credits to try Everlyn.ai.
https://redd.it/1l7omhx
@rStableDiffusion
Hey folks!
We just launched [Everlyn.app](http://Everlyn.app) — a new platform for video generation that is fast, powered by our newly developed tech in collaboration with world-class professors, and built with an intuitive UI. You can generate high-quality images and videos up to 30 seconds, add optional image input, use our intelligent prompt enhancement, and control the motions.
Key Features:
* ⚡ Fast inference (typically under 30s)
* 🎬 Long videos (up to 30s, multi-paragraph prompts supported)
* 📸 Free image generation (unlimited, watermark-free)
* 🎯 Fine-grained motion control
* 🤖 AI-powered prompt enhancement
💬 Since I’ve learned so much from this community and friends here, I’d love to give back. If you leave your email in the comments, I’ll personally send you 50 free credits to try Everlyn.ai.
https://redd.it/1l7omhx
@rStableDiffusion
www.everlyn.app
Everlyn AI-Free Fast Unlimited AI Video and Image Generator
Everlyn AI,World's fastest cheapest leading unlimited AI video generator,free unlimited AI image generator.image to video,text to video,text to image supported
People who've trained LORA models on both Kohya and OneTrainer with the same datasets, what differences have you noticed between the two?
https://redd.it/1l7p387
@rStableDiffusion
https://redd.it/1l7p387
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Whats the best Virtual Try-On model today?
I know none of them are perfect at assigning patterns/textures/text. But from what you've researched, which do you think in today's age is the most accurate at them?
I tried Flux Kontext Pro on Fal and it wasnt very accurate in determining what to change and what not to, same with 4o Image Gen. I wanted to try the google "dressup" virtual try on, but I cant seem to find it anywhere.
OSS models would be ideal as I can tweak the workflow rather than just the prompt on ComfyUI.
https://redd.it/1l7roia
@rStableDiffusion
I know none of them are perfect at assigning patterns/textures/text. But from what you've researched, which do you think in today's age is the most accurate at them?
I tried Flux Kontext Pro on Fal and it wasnt very accurate in determining what to change and what not to, same with 4o Image Gen. I wanted to try the google "dressup" virtual try on, but I cant seem to find it anywhere.
OSS models would be ideal as I can tweak the workflow rather than just the prompt on ComfyUI.
https://redd.it/1l7roia
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Does anyone know what ai software and prompts this guy uses to make these kinds of morphs?
https://youtu.be/5Vh61jy9suo?si=kSGsjeSd42hK0L2o
https://redd.it/1l7s5ja
@rStableDiffusion
https://youtu.be/5Vh61jy9suo?si=kSGsjeSd42hK0L2o
https://redd.it/1l7s5ja
@rStableDiffusion
YouTube
Volkswagen Best Car Models Evolution from 1938 to 2045
Discover the fascinating history of Volkswagen cars from the legendary Beetle (Type 1) to the futuristic all-electric ID. Trinity in just 4 minutes. This video showcases 45 iconic Volkswagen models in a stunning digital transformation timeline, highlighting…
Self Forcing: The new Holy Grail for video generation?
https://self-forcing.github.io/
> Our model generates high-quality 480P videos with an initial latency of ~0.8 seconds, after which frames are generated in a streaming fashion at ~16 FPS on a single H100 GPU and ~10 FPS on a single 4090 with some optimizations.
> Our method has the same speed as CausVid but has much better video quality, free from over-saturation artifacts and having more natural motion. Compared to Wan, SkyReels, and MAGI, our approach is 150–400× faster in terms of latency, while achieving comparable or superior visual quality.
https://redd.it/1l7sxh3
@rStableDiffusion
https://self-forcing.github.io/
> Our model generates high-quality 480P videos with an initial latency of ~0.8 seconds, after which frames are generated in a streaming fashion at ~16 FPS on a single H100 GPU and ~10 FPS on a single 4090 with some optimizations.
> Our method has the same speed as CausVid but has much better video quality, free from over-saturation artifacts and having more natural motion. Compared to Wan, SkyReels, and MAGI, our approach is 150–400× faster in terms of latency, while achieving comparable or superior visual quality.
https://redd.it/1l7sxh3
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community