SD 2.1 Text Encoder in 256 conv LyCORIS #565
idlebg
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Enhance your model's quality and sharpness using your own pre-trained Unet.
Trained on "121361" images.
How is it working for your 2.1? :)
Just uploaded it to Civit https://civitai.com/models/83622/difusionai-text-encoder-sd-21-lycoris
The text encoder (without UNET) is wrapped in LyCORIS. Optimizer: torch.optim.adamw.AdamW(weight_decay=0.01, betas=(0.9, 0.99))
Network dimension/rank: 768.0 Alpha: 768.0 Module: lycoris.kohya {'conv_dim': '256', 'conv_alpha': '256', 'algo': 'loha'}
Large size due to Lyco CONV 256 .. and saved in Float :D
For a1111
Install https://github.com/KohakuBlueleaf/a1111-sd-webui-lycoris
Download di.ffusion.ai-tXe-FXAA to /models/Lycoris
Option1:
Insert lyco:[di.FFUSION.ai](http://di.ffusion.ai/)-tXe-FXAA:1.0 to prompt
No need to split Unet and Text Enc as its only TX encoder there.
You can go up to 2x weights
Option2: If you need it always ON (ex run a batch from txt file) then you can go to settings / Quicksettings list
add sd_lyco
restart and you should have a drop-down now 🤟 🥃
![image](https://private-user-images.githubusercontent.com/50985923/243155372-e4f0a71d-c626-40ad-9585-177ba0fe63c0.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzg3NzY0MTIsIm5iZiI6MTczODc3NjExMiwicGF0aCI6Ii81MDk4NTkyMy8yNDMxNTUzNzItZTRmMGE3MWQtYzYyNi00MGFkLTk1ODUtMTc3YmEwZmU2M2MwLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMDUlMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjA1VDE3MjE1MlomWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPTI4YWExYjFmNmI5ZWUyNzM3YWI5YmIzNjM2MGM0MmVlMWZkNDM3YmM4Nzg2ODgwMTZiZjUwNjQ4MjFkZjI4YjgmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.TtauXV7zMushf_alXKsdlwd8zQc58TjKYzTxlYAlLTw)
More info:
"ss_text_encoder_lr": "1e-07",
"ss_keep_tokens": "3",
"ss_network_args": {
"conv_dim": "256",
"conv_alpha": "256",
"algo": "loha"
},
"img_count": 121361
}
"ss_total_batch_size": "100",
"ss_network_dim": "768",
"ss_max_bucket_reso": "1024",
"ss_network_alpha": "768.0",
"ss_steps": "2444",
"sshs_legacy_hash": "539b2745",
"ss_batch_size_per_device": "20",
"ss_max_train_steps": "2444",
"ss_network_module": "lycoris.kohya",
This is a heavy experimental version we used to test even with sloppy captions (quick WD tags and terrible clip), yet the results were satisfying.
Note: This is not the text encoder used in the official FFUSION AI model.
Beta Was this translation helpful? Give feedback.
All reactions