Revisiting WAN 2.2 for real-person realism, consented LoRA, retuned settings
https://redd.it/1t7cnaj
@rStableDiffusion
https://redd.it/1t7cnaj
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: Revisiting WAN 2.2 for real-person realism, consented LoRA, retuned settings
Explore this post and more from the StableDiffusion community
Z-Image Turbo for character LoRAs — honest comparison vs Flux after training the same character on both
https://redd.it/1t7g8de
@rStableDiffusion
https://redd.it/1t7g8de
@rStableDiffusion
FLUX, Open Research, and the Future of Visual AI — Stephen Batifol, Black Forest Labs
https://youtu.be/x8Yb4RidLgM?si=rRA-QvBXt4aUWu5k
https://redd.it/1t7ekxn
@rStableDiffusion
https://youtu.be/x8Yb4RidLgM?si=rRA-QvBXt4aUWu5k
https://redd.it/1t7ekxn
@rStableDiffusion
YouTube
FLUX, Open Research, and the Future of Visual AI — Stephen Batifol, Black Forest Labs
FLUX started as an image model story, but this talk makes the larger ambition clear: visual intelligence, not just image generation. From FLUX.1 through Kontext, FLUX.2, and FLUX.2 Klein, Black Forest Labs has been pushing fast, open releases while building…
AI tooling is starting to feel like PC modding culture
I think local AI setups are about to split into two completely different communities.
One side cares about actual production workflows:
* agents
* automation
* APIs
* inference efficiency
* data quality
* reproducibility
The other side mostly treats it like PC modding:
* model collecting
* benchmark screenshots
* “look how many params I run”
* endless UI tweaking
* generating the same test prompts forever
Not even judging either side honestly. I just think it explains why AI discussions online feel so weird lately. Two people can both be “into local AI” and barely even be talking about the same thing anymore.
https://redd.it/1t7fm79
@rStableDiffusion
I think local AI setups are about to split into two completely different communities.
One side cares about actual production workflows:
* agents
* automation
* APIs
* inference efficiency
* data quality
* reproducibility
The other side mostly treats it like PC modding:
* model collecting
* benchmark screenshots
* “look how many params I run”
* endless UI tweaking
* generating the same test prompts forever
Not even judging either side honestly. I just think it explains why AI discussions online feel so weird lately. Two people can both be “into local AI” and barely even be talking about the same thing anymore.
https://redd.it/1t7fm79
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Spent 3 training rounds trying to get a Jean-Léon Gérôme lora to retain fini surfaces
https://redd.it/1t7kpic
@rStableDiffusion
https://redd.it/1t7kpic
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: Spent 3 training rounds trying to get a Jean-Léon Gérôme lora to retain fini surfaces
Explore this post and more from the StableDiffusion community
LTX-2.3 PolarQuant Q5: 88% size reduction, near lossless quality (Cosine Similarity: 0.9986).
https://redd.it/1t7mhaw
@rStableDiffusion
https://redd.it/1t7mhaw
@rStableDiffusion
This media is not supported in your browser
VIEW IN TELEGRAM
Flux.2-Klein pipeline for real-time webcam stream processing in 30 FPS
https://redd.it/1t7nd7e
@rStableDiffusion
https://redd.it/1t7nd7e
@rStableDiffusion
3 years of training with AI tools finally put to use
I have learned so much from this community and I want to say thank you all who have contributed endlessly to this subreddit. Me and 2 other AI users teamed up to make children's music videos. Here are some of the clips that utilized WAN22. Not everything on the youtube channel is opensourced, so I won' t post the link here unless it's requested. These are all made with standard WAN22 FFLF workflow which I have tweaked over the years.
The one thing I realized along the way is that WAN can do some amazing things, it's all in the prompt. Such as block transition, crash zoom, pan, dolly, tilt, rotate. It can pretty much do it all.
Here is the workflow for the first video.
https://reddit.com/link/1t7nqgz/video/8dsi4qysuzzg1/player
https://reddit.com/link/1t7nqgz/video/01c16z8tuzzg1/player
https://reddit.com/link/1t7nqgz/video/0tz5363vuzzg1/player
https://reddit.com/link/1t7nqgz/video/n1guckfxuzzg1/player
https://reddit.com/link/1t7nqgz/video/plda65pxuzzg1/player
https://redd.it/1t7nqgz
@rStableDiffusion
I have learned so much from this community and I want to say thank you all who have contributed endlessly to this subreddit. Me and 2 other AI users teamed up to make children's music videos. Here are some of the clips that utilized WAN22. Not everything on the youtube channel is opensourced, so I won' t post the link here unless it's requested. These are all made with standard WAN22 FFLF workflow which I have tweaked over the years.
The one thing I realized along the way is that WAN can do some amazing things, it's all in the prompt. Such as block transition, crash zoom, pan, dolly, tilt, rotate. It can pretty much do it all.
Here is the workflow for the first video.
https://reddit.com/link/1t7nqgz/video/8dsi4qysuzzg1/player
https://reddit.com/link/1t7nqgz/video/01c16z8tuzzg1/player
https://reddit.com/link/1t7nqgz/video/0tz5363vuzzg1/player
https://reddit.com/link/1t7nqgz/video/n1guckfxuzzg1/player
https://reddit.com/link/1t7nqgz/video/plda65pxuzzg1/player
https://redd.it/1t7nqgz
@rStableDiffusion
Pastebin
{ "id": "f522eabf-4924-41b4-a7ce-ed9bcdcb53f4", "revision": 0, "last_no - Pastebin.com
Pastebin.com is the number one paste tool since 2002. Pastebin is a website where you can store text online for a set period of time.