This commit is contained in:
Andrew Z 2023-05-01 02:04:23 -04:00
parent 60932f8234
commit 1974a9e0ef
4 changed files with 7 additions and 11 deletions

View File

@ -1,3 +1,3 @@
{
"last_updated_timestamp": 1682906665
"last_updated_timestamp": 1682921061
}

View File

@ -274607,13 +274607,13 @@
{
"Name": "Sygil-webui",
"Repository": "hlky/sd-webui:latest",
"Registry": "https://hub.docker.com/r/hlky/sd-webui",
"Registry": "https://hub.docker.com/r/tukirito/sygil-webui",
"Network": "bridge",
"Shell": "sh",
"Privileged": "false",
"Support": "https://github.com/Sygil-Dev/sygil-webui/discussions",
"Project": "https://github.com/Sygil-Dev/sygil-webui/",
"Overview": "/**/ THIS DOCKER TAKES A LONG TIME TO BOOT THE FIRST TIME SEE LOG FOR OUTPUT /**/\r\n\r\nStreamlit ui for Stable Diffusion: https://github.com/sd-webui/stable-diffusion-webui/\r\nThis now includes all models:\r\n\r\nSD v1-4 (ckpt)\r\nWaifu Diffusion (ckpt)\r\nTrinArt (ckpt)\r\nViT-L-14 (pytorch.bin, used by inference models, is usually ~1.8gb autodownload to ~/.cache/huggingface/transformers)\r\nSD v1-4 (diffusers, used by txt2vid and textual inversion training)\r\nBLIP\r\nCLIP models: ViT-L-14, ViT-H-14, ViT-g-14, used by img2txt\r\nsd-concepts-library: (nearly) 600 pre-trained concepts to use in prompts, autoloaded when the placeholder is used in the prompt.\r\nGFPGAN v1.4 (and other weights that would normally be autodownloaded, detection_Resnet50_Final.pth and parsing_parsenet.pth, ~185MB)\r\nRealESRGAN: RealESRGAN_x4plus and RealESRGAN_x4plus_anime_6B\r\nLatent Diffusion Super Resolution (LDSR)\r\nNotes:\r\n\r\nThere is also hlky/sd-webui:runpod (vs :latest under the Repository tag) which includes a lot of models already\nMost models will unload when you use something else, atm there is a config option for img2txt.keep_all_models_loaded: False if you set this to True img2txt models (the CLIP ones, BLIP stays loaded unless you have optimized mode available) will all stay loaded, by default if a model that is loaded is no longer selected it will be unloaded\r\nvolume path now mirrors the repo so is /sd/outputs instead of /sd/user_data/outputs",
"Overview": "/**/ THIS DOCKER TAKES A LONG TIME TO BOOT THE FIRST TIME SEE LOG FOR OUTPUT /**/\r\n\r\nStreamlit ui for Stable Diffusion: https://github.com/sd-webui/stable-diffusion-webui/\r\nThis now includes all models:\r\n\r\nSD v1-4 (ckpt)\r\nWaifu Diffusion (ckpt)\r\nTrinArt (ckpt)\r\nViT-L-14 (pytorch.bin, used by inference models, is usually ~1.8gb autodownload to ~/.cache/huggingface/transformers)\r\nSD v1-4 (diffusers, used by txt2vid and textual inversion training)\r\nBLIP\r\nCLIP models: ViT-L-14, ViT-H-14, ViT-g-14, used by img2txt\r\nsd-concepts-library: (nearly) 600 pre-trained concepts to use in prompts, autoloaded when the placeholder is used in the prompt.\r\nGFPGAN v1.4 (and other weights that would normally be autodownloaded, detection_Resnet50_Final.pth and parsing_parsenet.pth, ~185MB)\r\nRealESRGAN: RealESRGAN_x4plus and RealESRGAN_x4plus_anime_6B\r\nLatent Diffusion Super Resolution (LDSR)\r\nNotes:\r\n\r\nMost models will unload when you use something else, atm there is a config option for img2txt.keep_all_models_loaded: False if you set this to True img2txt models (the CLIP ones, BLIP stays loaded unless you have optimized mode available) will all stay loaded, by default if a model that is loaded is no longer selected it will be unloaded\r\nvolume path now mirrors the repo so is /sd/outputs instead of /sd/user_data/outputs",
"WebUI": "http://[IP]:[PORT:8501]/",
"TemplateURL": "https://raw.githubusercontent.com/pyrater/docker-templates/master/sd-webui.xml",
"Icon": "https://github.com/pyrater/docker-templates/raw/master/hlky.png",

View File

@ -1,7 +1,7 @@
{
"apps": 2066,
"last_updated_timestamp": 1682906665,
"last_updated": "2023-04-30 22:04",
"last_updated_timestamp": 1682921061,
"last_updated": "2023-05-01 02:04",
"categories": [
{
"Cat": "Backup:",
@ -274769,13 +274769,13 @@
{
"Name": "Sygil-webui",
"Repository": "hlky/sd-webui:latest",
"Registry": "https://hub.docker.com/r/hlky/sd-webui",
"Registry": "https://hub.docker.com/r/tukirito/sygil-webui",
"Network": "bridge",
"Shell": "sh",
"Privileged": "false",
"Support": "https://github.com/Sygil-Dev/sygil-webui/discussions",
"Project": "https://github.com/Sygil-Dev/sygil-webui/",
"Overview": "/**/ THIS DOCKER TAKES A LONG TIME TO BOOT THE FIRST TIME SEE LOG FOR OUTPUT /**/\r\n\r\nStreamlit ui for Stable Diffusion: https://github.com/sd-webui/stable-diffusion-webui/\r\nThis now includes all models:\r\n\r\nSD v1-4 (ckpt)\r\nWaifu Diffusion (ckpt)\r\nTrinArt (ckpt)\r\nViT-L-14 (pytorch.bin, used by inference models, is usually ~1.8gb autodownload to ~/.cache/huggingface/transformers)\r\nSD v1-4 (diffusers, used by txt2vid and textual inversion training)\r\nBLIP\r\nCLIP models: ViT-L-14, ViT-H-14, ViT-g-14, used by img2txt\r\nsd-concepts-library: (nearly) 600 pre-trained concepts to use in prompts, autoloaded when the placeholder is used in the prompt.\r\nGFPGAN v1.4 (and other weights that would normally be autodownloaded, detection_Resnet50_Final.pth and parsing_parsenet.pth, ~185MB)\r\nRealESRGAN: RealESRGAN_x4plus and RealESRGAN_x4plus_anime_6B\r\nLatent Diffusion Super Resolution (LDSR)\r\nNotes:\r\n\r\nThere is also hlky/sd-webui:runpod (vs :latest under the Repository tag) which includes a lot of models already\nMost models will unload when you use something else, atm there is a config option for img2txt.keep_all_models_loaded: False if you set this to True img2txt models (the CLIP ones, BLIP stays loaded unless you have optimized mode available) will all stay loaded, by default if a model that is loaded is no longer selected it will be unloaded\r\nvolume path now mirrors the repo so is /sd/outputs instead of /sd/user_data/outputs",
"Overview": "/**/ THIS DOCKER TAKES A LONG TIME TO BOOT THE FIRST TIME SEE LOG FOR OUTPUT /**/\r\n\r\nStreamlit ui for Stable Diffusion: https://github.com/sd-webui/stable-diffusion-webui/\r\nThis now includes all models:\r\n\r\nSD v1-4 (ckpt)\r\nWaifu Diffusion (ckpt)\r\nTrinArt (ckpt)\r\nViT-L-14 (pytorch.bin, used by inference models, is usually ~1.8gb autodownload to ~/.cache/huggingface/transformers)\r\nSD v1-4 (diffusers, used by txt2vid and textual inversion training)\r\nBLIP\r\nCLIP models: ViT-L-14, ViT-H-14, ViT-g-14, used by img2txt\r\nsd-concepts-library: (nearly) 600 pre-trained concepts to use in prompts, autoloaded when the placeholder is used in the prompt.\r\nGFPGAN v1.4 (and other weights that would normally be autodownloaded, detection_Resnet50_Final.pth and parsing_parsenet.pth, ~185MB)\r\nRealESRGAN: RealESRGAN_x4plus and RealESRGAN_x4plus_anime_6B\r\nLatent Diffusion Super Resolution (LDSR)\r\nNotes:\r\n\r\nMost models will unload when you use something else, atm there is a config option for img2txt.keep_all_models_loaded: False if you set this to True img2txt models (the CLIP ones, BLIP stays loaded unless you have optimized mode available) will all stay loaded, by default if a model that is loaded is no longer selected it will be unloaded\r\nvolume path now mirrors the repo so is /sd/outputs instead of /sd/user_data/outputs",
"WebUI": "http://[IP]:[PORT:8501]/",
"TemplateURL": "https://raw.githubusercontent.com/pyrater/docker-templates/master/sd-webui.xml",
"Icon": "https://github.com/pyrater/docker-templates/raw/master/hlky.png",

View File

@ -3565,9 +3565,5 @@ Updating stats on containers
Privileged: 104
No change to feed
Updating GitHub