Stable Diffusion

4217 readers
10 users here now

Discuss matters related to our favourite AI Art generation technology

Also see

Other communities

founded 1 year ago
MODERATORS
1
2
 
 

Text:

Emad@EMostaque

Delighted to announce the public open source release of #StableDiffusion!

Please see our release post and retweet! stability.ai/blog/stable-di...

Proud of everyone involved in releasing this tech that is the first of a series of models to activate the creative potential of humanity

11:07 AM • Aug 22, 2022

3
4
 
 

Has anyone of you stumbled upon any information on how to get it running on machines like mine, or does it just not have enough power?

5
3
submitted 2 days ago* (last edited 2 days ago) by [email protected] to c/[email protected]
6
7
8
9
10
 
 
11
 
 

V2 is quantized in a better way and is 0.5 GB larger than the previous version.

On Hugging Face: https://huggingface.co/lllyasviel/flux1-dev-bnb-nf4

12
13
14
15
16
17
18
19
 
 

Among other improvements, the new defaults set --flux_guidance_value=1, removing the need to use CFG nodes at inference, reducing generation time and improving image quality of LoRAs slightly.

Changelog: https://github.com/bghira/SimpleTuner/releases/tag/v0.9.8.1

Sample LoRA: https://huggingface.co/ptx0/flux-dreambooth-lora-r16-dev-cfg1/blob/main/pytorch_lora_weights.safetensors

20
 
 

I think this was here a while ago, but I lost the post. If I remember, there was an extension where someone had built up a repository of poses, and then in the UI you could go through them and pick out the one you wanted for your image. Does anyone remember that? I have controlnet and can do poses, but finding them is a bear right now.

21
22
23
24
 
 

From The Hugging Face Model Card:

Not Ready

This is a WIP and not ready for use. This is an early testing version for research and development. You may know what this is and how to use it, if so, feel free, but it will change as I continue to develop it. I plan to do many updates to it frequently. So you may want to set a revision if you intend to use it anyway.

What is this?

FLUX.1-schnell is an amazing distilled model with an apache 2.0 license. However, it is not finetunable. LoRAs, IP adapters, control nets, etc, cannot be trained on it because it is distilled. The goal of this project is to finetune a non-distilled version of it that can be used as a training base to train adapters for FLUX.1-schnell.

Current Issues

Since we are breaking the distillation, this model will need many steps and guidance to produce good results. Currently, this model, like the schnell version, does not have guidance embeddings. Because of this (and possible other factors) images generated with this model will not look great. However, this hopefully will not affect training, since guidance is not used during training. The things trained on this model are intended to be used on the schnell version anyway. I am working on training guidance embeddings for it, but hopefully it will work as a training base without them.

25
 
 

Quoted From Reddit:

Release: https://github.com/bghira/SimpleTuner/releases/tag/v0.9.8

It's here! Runs on 24G cards using Quanto's 8bit quantisation or down to 13G with a 2bit base model for the truly terrifying potato LoRA of your dreams!

If you're after accuracy, a 40G card will do Just Fine, with 80G cards being somewhat of a sweet spot for larger training efforts.

What you get:

  • LoRA, full tuning (but probably just don't do that)
  • Documentation to get you started fast
  • Probably better for just square crop training for now - might artifact for weird resolutions
  • Quantised base model unlocks the ability to safely use Adafactor, Prodigy, and other neat optimisers as a consolation prize for losing access to full bf16 training (AdamWBF16 just won't work with Quanto)

not a fine-tune, but, Flux-fast

frequently observed questions

  • 10k images isn't a requirement for training, that's just a healthy amount of regularisation data to have.

  • Regularisation data with text in it is needed to retain text while tuning Flux. It's sensitive to forgetting.

  • you can finetune either dev or schnell, and you probably don't even need special training dynamics for schnell. it seems to work just fine, but at lower quality than dev, because the base model is lower quality.

  • yes, multiple 4090s or 3090s can be used. no, it's probably not a good idea to try splitting the model across them - stick with quantising and LoRAs.

thank you

You all had a really good response to my work; as well as respect for the limitations of the progress at that point, and the optimism on what can happen next.

I'm not sure whether we can really "improve" this state of the art model - probably merely being able to change it without ruining it is good enough for me.

further work, help needed

If any of you would like to take on any of the items in this issue, we can implement them into SimpleTuner next and unlock another level of fine-tuning efficiency: https://github.com/huggingface/peft/issues/1935

The principle improvement for Flux here will be the ability to train quantised LoKr models, where even the weights of the LoRA itself become quantised in addition to the base model.

view more: next ›