Media is too big
VIEW IN TELEGRAM
Shooting Aliens - 100% Qwen Image Edit 2509 + NextScene LoRA + Wan 2.2 I2V
https://redd.it/1o6m23n
@rStableDiffusion
https://redd.it/1o6m23n
@rStableDiffusion
ByteDance FaceCLIP Model Taken Down
HuggingFace Repo (Now Removed): https://huggingface.co/ByteDance/FaceCLIP
Did anyone make a copy of the files? Not sure why this was removed, it was a brilliant model.
From the release:
"ByteDance just released FaceCLIP on Hugging Face!
A new vision-language model specializing in understanding and generating diverse human faces.
Dive into the future of facial AI."
They released both SDXL and Flux fine-tunes that worked with the FaceCLIP weights.
https://redd.it/1o6xiry
@rStableDiffusion
HuggingFace Repo (Now Removed): https://huggingface.co/ByteDance/FaceCLIP
Did anyone make a copy of the files? Not sure why this was removed, it was a brilliant model.
From the release:
"ByteDance just released FaceCLIP on Hugging Face!
A new vision-language model specializing in understanding and generating diverse human faces.
Dive into the future of facial AI."
They released both SDXL and Flux fine-tunes that worked with the FaceCLIP weights.
https://redd.it/1o6xiry
@rStableDiffusion
huggingface.co
ByteDance/FaceCLIP · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
FREE Face Dataset generation workflow for lora training (Qwen edit 2509)
https://redd.it/1o6xjwu
@rStableDiffusion
https://redd.it/1o6xjwu
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: FREE Face Dataset generation workflow for lora training (Qwen edit 2509)
Explore this post and more from the StableDiffusion community
Anyone else use their ai rig as a heater?
So, I recently moved my ai machine(RTX3090) into my bedroom and discovered the thing is literally a space heater. Woke up this morning sweating. My electric bill has been ridiculous but I just chalked it up to inflation and summer time running the air conditioner a lot.
https://redd.it/1o6nhly
@rStableDiffusion
So, I recently moved my ai machine(RTX3090) into my bedroom and discovered the thing is literally a space heater. Woke up this morning sweating. My electric bill has been ridiculous but I just chalked it up to inflation and summer time running the air conditioner a lot.
https://redd.it/1o6nhly
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Where to post music and other kinds of Lora’s?
Hey
Just wondering has anyone been trai ing any music models or other kinds of models and where do you guys post these.
I'm sitting on a lot of trained Loras for ace step and music gen and have no idea where to post.
Are people even training music Loras or other kinds of Loras? If so where are you posting them.
https://redd.it/1o72or8
@rStableDiffusion
Hey
Just wondering has anyone been trai ing any music models or other kinds of models and where do you guys post these.
I'm sitting on a lot of trained Loras for ace step and music gen and have no idea where to post.
Are people even training music Loras or other kinds of Loras? If so where are you posting them.
https://redd.it/1o72or8
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
I built a wheel of nunchaku for cuda130, reducing size by 57%.
whl (only windows): https://huggingface.co/X5R/nunchaku-cu130-wheel
it works with torch2.9.0+cu130. here to install
besides, torch of cu130 is also smaller than cu12x, reducing size more than 50%. I don't know why.
https://redd.it/1o741un
@rStableDiffusion
whl (only windows): https://huggingface.co/X5R/nunchaku-cu130-wheel
it works with torch2.9.0+cu130. here to install
pip install -U torch torchaudio torchvision --index-url https://download.pytorch.org/whl/test/cu130
besides, torch of cu130 is also smaller than cu12x, reducing size more than 50%. I don't know why.
https://redd.it/1o741un
@rStableDiffusion
huggingface.co
X5R/nunchaku-cu130-wheel · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Where do people train Qwen Image Edit 2509 LoRAs?
Hi, I trained a few small LoRAs with AI-Toolkit locally, and some bigger ones for Qwen Image Edit running AI-Toolkit on Runpod using Ostris guide. Is it possible to train 2509 LoRAs there already? Don't wanna rent a GPU just to check if it's available, and I cannot find the info with researches. Thanks!
https://redd.it/1o74lnd
@rStableDiffusion
Hi, I trained a few small LoRAs with AI-Toolkit locally, and some bigger ones for Qwen Image Edit running AI-Toolkit on Runpod using Ostris guide. Is it possible to train 2509 LoRAs there already? Don't wanna rent a GPU just to check if it's available, and I cannot find the info with researches. Thanks!
https://redd.it/1o74lnd
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Compile fp8 on RTX 30xx in triton-windows 3.5
I've merged the patch to let
I hoped this could be superseded by GGUF + better
This is the first feature in the 'core' part (rather than the Windows support code) that's deliberately different from the official Triton. It should also work on Linux but I'm not sure what's the best way to publish Linux wheels.
I'm not an expert on PTX. Welcome help in optimizing those PTX code.
https://redd.it/1o75zgt
@rStableDiffusion
I've merged the patch to let
torch.compile
work with fp8 on Ampere GPUs and let's see how it rolls out: https://github.com/woct0rdho/triton-windows/pull/140I hoped this could be superseded by GGUF + better
torch.compile
or Nunchaku, but as of PyTorch 2.9 I realized that fp8 + the block swap in ComfyUI-WanVideoWrapper (or ComfyUI-wanBlockswap for native workflows) runs faster and causes fewer recompilations than GGUF + the block swap in ComfyUI-GGUF on my machine.This is the first feature in the 'core' part (rather than the Windows support code) that's deliberately different from the official Triton. It should also work on Linux but I'm not sure what's the best way to publish Linux wheels.
I'm not an expert on PTX. Welcome help in optimizing those PTX code.
triton-windows 3.2.0.post21
is also released, which supports fp8 on RTX 20xx.https://redd.it/1o75zgt
@rStableDiffusion
GitHub
Enable F8E4M3 conversions on Nvidia GPUs with sm < 89, and fix F8E5M2 conversions by woct0rdho · Pull Request #140 · woct0rdho/triton…
Motivation
Nvidia GPUs with sm < 89 are still widely used, see e.g. Steam hardware survey. When running large AI models, a common usage is to store the parameters in fp8, and cast them to fp...
Nvidia GPUs with sm < 89 are still widely used, see e.g. Steam hardware survey. When running large AI models, a common usage is to store the parameters in fp8, and cast them to fp...
16 GB of VRAM: Is it worth leaving SDXL for Chroma, Flux, or WAN text-to-image?
Hello, I currently mainly use SDXL or its PONY variant. For 20 steps and a resolution of 896x1152, I can generate an image without LoRAs in 10 seconds using FORGE or its variants.
Like most people, I use the unscientific method of trial and error: I create an image, and 10 seconds is a comfortable waiting time to change parameters and try again.
However, I would like to be able to use the real text generation capabilities and the strong prompt adherence that other models like Chroma, Flux, or WAN have.
The problem is the waiting time for image generation with those models. In my case, it easily goes over 60 seconds, which obviously makes a trial-and-error-based creation method useless and impossible.
Basically, my question is: Is there any way to reduce the times to something close to SDXL's while maintaining image quality? I tried "Sagge Attention" in ComfyUI with WAN 2.2 and the times for generating one image were absolutely excessive.
https://redd.it/1o76sa4
@rStableDiffusion
Hello, I currently mainly use SDXL or its PONY variant. For 20 steps and a resolution of 896x1152, I can generate an image without LoRAs in 10 seconds using FORGE or its variants.
Like most people, I use the unscientific method of trial and error: I create an image, and 10 seconds is a comfortable waiting time to change parameters and try again.
However, I would like to be able to use the real text generation capabilities and the strong prompt adherence that other models like Chroma, Flux, or WAN have.
The problem is the waiting time for image generation with those models. In my case, it easily goes over 60 seconds, which obviously makes a trial-and-error-based creation method useless and impossible.
Basically, my question is: Is there any way to reduce the times to something close to SDXL's while maintaining image quality? I tried "Sagge Attention" in ComfyUI with WAN 2.2 and the times for generating one image were absolutely excessive.
https://redd.it/1o76sa4
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
Which one of you? | Man Stores AI-Generated ‘Robot Porn' on His Government Computer, Loses Access to Nuclear Secrets
https://www.404media.co/man-stores-ai-generated-robot-porn-on-his-government-computer-loses-access-to-nuclear-secrets/
https://redd.it/1o7c7fe
@rStableDiffusion
https://www.404media.co/man-stores-ai-generated-robot-porn-on-his-government-computer-loses-access-to-nuclear-secrets/
https://redd.it/1o7c7fe
@rStableDiffusion
404 Media
Man Stores AI-Generated ‘Robot Porn' on His Government Computer, Loses Access to Nuclear Secrets
A man who works for the people overseeing America’s nuclear stockpile has lost his security clearance after he uploaded 187,000 pornographic images to a Department of Energy (DOE) network.
I made Nunchaku SVDQuant for my current favorite model CenKreChro (Krea+Chroma merge)
https://huggingface.co/spooknik/CenKreChro-SVDQ
https://redd.it/1o7bk44
@rStableDiffusion
https://huggingface.co/spooknik/CenKreChro-SVDQ
https://redd.it/1o7bk44
@rStableDiffusion
huggingface.co
spooknik/CenKreChro-SVDQ · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
Hunyuanimage 3.0 vs Sora 2 frame caps refined with Wan2.2 low noise 2 step upscaler
https://redd.it/1o7cak1
@rStableDiffusion
https://redd.it/1o7cak1
@rStableDiffusion
Reddit
From the StableDiffusion community on Reddit: Hunyuanimage 3.0 vs Sora 2 frame caps refined with Wan2.2 low noise 2 step upscaler
Explore this post and more from the StableDiffusion community