drhead commited on
Commit
2df88a6
1 Parent(s): 3a51ff7

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: diffusers
3
+ ---
4
+ # Furception v1.0, by Project RedRocket.
5
+
6
+ This is a VAE decoder finetune, resumed from stabilityai/sd-vae-ft-mse using images from e621. It is trained with a mixture of MAE and MSE loss to maintain an acceptable balance between sharpness and smooth outputs, and loss is calculated in Oklab color space in order to prioritize image reconstruction based on which color channels are more perceptually significant.
7
+
8
+ Our testing has shown that the VAE is good at eliminating unwanted high-frequency noise when used on models trained on similar data. It may have some generalizability to a broader range of art styles due to the variety of different styles in the dataset.
9
+
10
+ #### Licensing:
11
+
12
+ You are free to use this model for personal, non-commercial use.
13
+
14
+ You are also free to distribute this model alongside other (non-commercial) models, as long as you give credit. Please include the version number as well in case future models are released.
15
+
16
+ #### Training details:
17
+ Overall training is fundamentally similar to LDM. We used the same relative base weights for MAE, MSE, and LPIPS as used in LDM and in sd-vae-ft-mse in the case of LPIPS. The discriminator's weight in the loss objective is dynamically set so that the gradient norm for the discriminator is half that of the reconstruction loss, just like LDM. We used a similar discriminator to what LDM uses, except reparameterized to Wasserstein loss with a gradient penalty and with its group norm layers replaced with layer norms.
18
+
19
+ Training for version 1.0 used random square crops at various levels of downscales (Lanczos with antialiasing), randomly rotated and flipped. Training ran for 150,000 steps at a batch size of 32. EMA weights were accumulated using a similar decay to sd-vae-ft-mse scaled for our batch size and are the release version of the model.
20
+
21
+ #### Credits:
22
+ Development and research lead by @drhead.
23
+
24
+ With research and development assistance by @RedHotTensors.
25
+
26
+ And additional research assistance by @lodestones and Thessalo.
27
+
28
+ Dataset curation by @lodestones and Bannanapuncakes, with additional curation by @RedHotTensors.
29
+
30
+ And thanks to dogarrowtype for system administration assistance.
31
+
32
+ #### Based on:
33
+ CompVis Latent Diffusion: https://github.com/CompVis/latent-diffusion/
34
+
35
+ StabilityAI sd-vae-ft-mse: https://huggingface.co/stabilityai/sd-vae-ft-mse
36
+
37
+ LPIPS by Richard Zhang, et al: https://github.com/richzhang/PerceptualSimilarity
38
+
39
+ OkLab by Björn Ottosson: https://bottosson.github.io/posts/oklab/
40
+
41
+ fine-tune-models by Jonathan Chang: https://github.com/cccntu/fine-tune-models/
42
+
43
+ #### Built on:
44
+ Flax by Google Brain: https://github.com/google/flax
45
+
46
+ And Huggingface Diffusers: https://github.com/huggingface/diffusers
47
+
48
+ With deep thanks to the innumerable artists who released their works to the public for fair use in this non-commercial research project.