mirror of
https://github.com/sd-webui/stable-diffusion-webui.git
synced 2024-12-13 18:02:31 +03:00
added cloudflared tunnel.
This commit is contained in:
parent
5432b73fb8
commit
70e6d12451
@ -49,7 +49,7 @@
|
||||
"\n",
|
||||
"## Installation instructions for:\n",
|
||||
"\n",
|
||||
"- **[Windows](https://sygil-dev.github.io/sygil-webui/docs/1.windows-installation.html)** \n",
|
||||
"- **[Windows](https://sygil-dev.github.io/sygil-webui/docs/1.windows-installation.html)**\n",
|
||||
"- **[Linux](https://sygil-dev.github.io/sygil-webui/docs/2.linux-installation.html)**\n",
|
||||
"\n",
|
||||
"### Want to ask a question or request a feature?\n",
|
||||
@ -172,7 +172,7 @@
|
||||
"\n",
|
||||
"If you want to use GFPGAN to improve generated faces, you need to install it separately.\n",
|
||||
"Download [GFPGANv1.4.pth](https://github.com/TencentARC/GFPGAN/releases/download/v1.3.4/GFPGANv1.4.pth) and put it\n",
|
||||
"into the `/sygil-webui/models/gfpgan` directory. \n",
|
||||
"into the `/sygil-webui/models/gfpgan` directory.\n",
|
||||
"\n",
|
||||
"### RealESRGAN\n",
|
||||
"\n",
|
||||
@ -182,7 +182,7 @@
|
||||
"There is also a separate tab for using RealESRGAN on any picture.\n",
|
||||
"\n",
|
||||
"Download [RealESRGAN_x4plus.pth](https://github.com/xinntao/Real-ESRGAN/releases/download/v0.1.0/RealESRGAN_x4plus.pth) and [RealESRGAN_x4plus_anime_6B.pth](https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.2.4/RealESRGAN_x4plus_anime_6B.pth).\n",
|
||||
"Put them into the `sygil-webui/models/realesrgan` directory. \n",
|
||||
"Put them into the `sygil-webui/models/realesrgan` directory.\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"\n",
|
||||
@ -219,8 +219,8 @@
|
||||
"\n",
|
||||
"[Stable Diffusion](#stable-diffusion-v1) is a latent text-to-image diffusion\n",
|
||||
"model.\n",
|
||||
"Thanks to a generous compute donation from [Stability AI](https://stability.ai/) and support from [LAION](https://laion.ai/), we were able to train a Latent Diffusion Model on 512x512 images from a subset of the [LAION-5B](https://laion.ai/blog/laion-5b/) database. \n",
|
||||
"Similar to Google's [Imagen](https://arxiv.org/abs/2205.11487), \n",
|
||||
"Thanks to a generous compute donation from [Stability AI](https://stability.ai/) and support from [LAION](https://laion.ai/), we were able to train a Latent Diffusion Model on 512x512 images from a subset of the [LAION-5B](https://laion.ai/blog/laion-5b/) database.\n",
|
||||
"Similar to Google's [Imagen](https://arxiv.org/abs/2205.11487),\n",
|
||||
"this model uses a frozen CLIP ViT-L/14 text encoder to condition the model on text prompts.\n",
|
||||
"With its 860M UNet and 123M text encoder, the model is relatively lightweight and runs on a GPU with at least 10GB VRAM.\n",
|
||||
"See [this section](#stable-diffusion-v1) below and the [model card](https://huggingface.co/CompVis/stable-diffusion).\n",
|
||||
@ -229,26 +229,26 @@
|
||||
"\n",
|
||||
"Stable Diffusion v1 refers to a specific configuration of the model\n",
|
||||
"architecture that uses a downsampling-factor 8 autoencoder with an 860M UNet\n",
|
||||
"and CLIP ViT-L/14 text encoder for the diffusion model. The model was pretrained on 256x256 images and \n",
|
||||
"and CLIP ViT-L/14 text encoder for the diffusion model. The model was pretrained on 256x256 images and\n",
|
||||
"then finetuned on 512x512 images.\n",
|
||||
"\n",
|
||||
"*Note: Stable Diffusion v1 is a general text-to-image diffusion model and therefore mirrors biases and (mis-)conceptions that are present\n",
|
||||
"in its training data. \n",
|
||||
"in its training data.\n",
|
||||
"Details on the training procedure and data, as well as the intended use of the model can be found in the corresponding [model card](https://huggingface.co/CompVis/stable-diffusion).\n",
|
||||
"\n",
|
||||
"## Comments\n",
|
||||
"\n",
|
||||
"- Our codebase for the diffusion models builds heavily on [OpenAI's ADM codebase](https://github.com/openai/guided-diffusion)\n",
|
||||
" and [https://github.com/lucidrains/denoising-diffusion-pytorch](https://github.com/lucidrains/denoising-diffusion-pytorch). \n",
|
||||
" and [https://github.com/lucidrains/denoising-diffusion-pytorch](https://github.com/lucidrains/denoising-diffusion-pytorch).\n",
|
||||
" Thanks for open-sourcing!\n",
|
||||
"\n",
|
||||
"- The implementation of the transformer encoder is from [x-transformers](https://github.com/lucidrains/x-transformers) by [lucidrains](https://github.com/lucidrains?tab=repositories). \n",
|
||||
"- The implementation of the transformer encoder is from [x-transformers](https://github.com/lucidrains/x-transformers) by [lucidrains](https://github.com/lucidrains?tab=repositories).\n",
|
||||
"\n",
|
||||
"## BibTeX\n",
|
||||
"\n",
|
||||
"```\n",
|
||||
"@misc{rombach2021highresolution,\n",
|
||||
" title={High-Resolution Image Synthesis with Latent Diffusion Models}, \n",
|
||||
" title={High-Resolution Image Synthesis with Latent Diffusion Models},\n",
|
||||
" author={Robin Rombach and Andreas Blattmann and Dominik Lorenz and Patrick Esser and Björn Ommer},\n",
|
||||
" year={2021},\n",
|
||||
" eprint={2112.10752},\n",
|
||||
@ -502,7 +502,7 @@
|
||||
" file_url = file_info[file]['download_link']\n",
|
||||
" if 'save_location' in file_info[file]:\n",
|
||||
" file_path = file_info[file]['save_location']\n",
|
||||
" else: \n",
|
||||
" else:\n",
|
||||
" file_path = models[model]['save_location']\n",
|
||||
" download_file(file_name, file_path, file_url)\n",
|
||||
"\n",
|
||||
@ -580,6 +580,54 @@
|
||||
},
|
||||
"execution_count": null,
|
||||
"outputs": []
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"source": [
|
||||
"Run Streamlit through cloudflare."
|
||||
],
|
||||
"metadata": {
|
||||
"id": "QhazvrFG97zX"
|
||||
}
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"metadata": {
|
||||
"id": "jjjjjjjjjjjjjj"
|
||||
},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"#@title Run Streamlit through cloudflare.\n",
|
||||
"!wget https://github.com/cloudflare/cloudflared/releases/latest/download/cloudflared-linux-amd64.deb\n",
|
||||
"!dpkg -i cloudflared-linux-amd64.deb\n",
|
||||
"\n",
|
||||
"import subprocess\n",
|
||||
"import threading\n",
|
||||
"import time\n",
|
||||
"import socket\n",
|
||||
"import urllib.request\n",
|
||||
"\n",
|
||||
"def iframe_thread(port):\n",
|
||||
" while True:\n",
|
||||
" time.sleep(0.5)\n",
|
||||
" sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n",
|
||||
" result = sock.connect_ex(('127.0.0.1', port))\n",
|
||||
" if result == 0:\n",
|
||||
" break\n",
|
||||
" sock.close()\n",
|
||||
"\n",
|
||||
" p = subprocess.Popen([\"cloudflared\", \"tunnel\", \"--url\", \"http://127.0.0.1:{}\".format(port)], stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n",
|
||||
" for line in p.stderr:\n",
|
||||
" l = line.decode()\n",
|
||||
" if \"trycloudflare.com \" in l:\n",
|
||||
" print(\"This is the URL to access Sygil WebUI:\", l[l.find(\"http\"):], end='')\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"threading.Thread(target=iframe_thread, daemon=True, args=(8501)).start()\n",
|
||||
"\n",
|
||||
"!streamlit run scripts/webui_streamlit.py --theme.base dark --server.headless true"
|
||||
]
|
||||
}
|
||||
]
|
||||
}
|
||||
}
|
Loading…
Reference in New Issue
Block a user