2022-11-01 16:13:59 +00:00
|
|
|
import base64
|
|
|
|
import io
|
2022-10-29 19:45:29 +00:00
|
|
|
import time
|
2022-10-17 06:58:42 +00:00
|
|
|
import uvicorn
|
2022-11-03 03:51:22 +00:00
|
|
|
from threading import Lock
|
2022-11-23 09:43:58 +00:00
|
|
|
from io import BytesIO
|
2022-11-24 05:10:40 +00:00
|
|
|
from gradio.processing_utils import decode_base64_to_file
|
2022-11-03 03:51:22 +00:00
|
|
|
from fastapi import APIRouter, Depends, FastAPI, HTTPException
|
2022-11-15 08:12:34 +00:00
|
|
|
from fastapi.security import HTTPBasic, HTTPBasicCredentials
|
|
|
|
from secrets import compare_digest
|
|
|
|
|
2022-10-17 06:58:42 +00:00
|
|
|
import modules.shared as shared
|
2022-12-24 23:02:22 +00:00
|
|
|
from modules import sd_samplers, deepbooru, sd_hijack
|
2022-10-22 23:24:04 +00:00
|
|
|
from modules.api.models import *
|
2022-10-21 23:27:40 +00:00
|
|
|
from modules.processing import StableDiffusionProcessingTxt2Img, StableDiffusionProcessingImg2Img, process_images
|
2022-10-29 19:09:19 +00:00
|
|
|
from modules.extras import run_extras, run_pnginfo
|
2022-12-24 23:02:22 +00:00
|
|
|
from modules.textual_inversion.textual_inversion import create_embedding, train_embedding
|
|
|
|
from modules.textual_inversion.preprocess import preprocess
|
|
|
|
from modules.hypernetworks.hypernetwork import create_hypernetwork, train_hypernetwork
|
2022-11-23 09:43:58 +00:00
|
|
|
from PIL import PngImagePlugin,Image
|
2022-11-03 03:51:22 +00:00
|
|
|
from modules.sd_models import checkpoints_list
|
|
|
|
from modules.realesrgan_model import get_realesrgan_models
|
2022-12-24 23:02:22 +00:00
|
|
|
from modules import devices
|
2022-11-03 03:51:22 +00:00
|
|
|
from typing import List
|
2022-10-18 19:04:56 +00:00
|
|
|
|
2022-10-23 02:13:32 +00:00
|
|
|
def upscaler_to_index(name: str):
|
|
|
|
try:
|
|
|
|
return [x.name.lower() for x in shared.sd_upscalers].index(name.lower())
|
|
|
|
except:
|
2022-10-23 19:01:16 +00:00
|
|
|
raise HTTPException(status_code=400, detail=f"Invalid upscaler, needs to be on of these: {' , '.join([x.name for x in sd_upscalers])}")
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-30 06:10:22 +00:00
|
|
|
|
2022-11-19 09:01:51 +00:00
|
|
|
def validate_sampler_name(name):
|
|
|
|
config = sd_samplers.all_samplers_map.get(name, None)
|
|
|
|
if config is None:
|
|
|
|
raise HTTPException(status_code=404, detail="Sampler not found")
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-11-19 09:01:51 +00:00
|
|
|
return name
|
2022-10-21 23:27:40 +00:00
|
|
|
|
2022-10-23 19:01:16 +00:00
|
|
|
def setUpscalers(req: dict):
|
|
|
|
reqDict = vars(req)
|
|
|
|
reqDict['extras_upscaler_1'] = upscaler_to_index(req.upscaler_1)
|
|
|
|
reqDict['extras_upscaler_2'] = upscaler_to_index(req.upscaler_2)
|
|
|
|
reqDict.pop('upscaler_1')
|
|
|
|
reqDict.pop('upscaler_2')
|
|
|
|
return reqDict
|
2022-10-27 19:20:15 +00:00
|
|
|
|
2022-11-24 05:10:40 +00:00
|
|
|
def decode_base64_to_image(encoding):
|
|
|
|
if encoding.startswith("data:image/"):
|
|
|
|
encoding = encoding.split(";")[1].split(",")[1]
|
|
|
|
return Image.open(BytesIO(base64.b64decode(encoding)))
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-11-01 16:13:59 +00:00
|
|
|
def encode_pil_to_base64(image):
|
2022-11-02 14:37:45 +00:00
|
|
|
with io.BytesIO() as output_bytes:
|
|
|
|
|
|
|
|
# Copy any text-only metadata
|
|
|
|
use_metadata = False
|
|
|
|
metadata = PngImagePlugin.PngInfo()
|
|
|
|
for key, value in image.info.items():
|
|
|
|
if isinstance(key, str) and isinstance(value, str):
|
|
|
|
metadata.add_text(key, value)
|
|
|
|
use_metadata = True
|
|
|
|
|
|
|
|
image.save(
|
|
|
|
output_bytes, "PNG", pnginfo=(metadata if use_metadata else None)
|
|
|
|
)
|
|
|
|
bytes_data = output_bytes.getvalue()
|
|
|
|
return base64.b64encode(bytes_data)
|
2022-11-01 16:13:59 +00:00
|
|
|
|
|
|
|
|
2022-10-17 06:58:42 +00:00
|
|
|
class Api:
|
2022-11-03 03:51:22 +00:00
|
|
|
def __init__(self, app: FastAPI, queue_lock: Lock):
|
2022-11-15 08:12:34 +00:00
|
|
|
if shared.cmd_opts.api_auth:
|
2022-12-15 02:01:32 +00:00
|
|
|
self.credentials = dict()
|
2022-11-15 08:12:34 +00:00
|
|
|
for auth in shared.cmd_opts.api_auth.split(","):
|
|
|
|
user, password = auth.split(":")
|
2022-12-15 02:01:32 +00:00
|
|
|
self.credentials[user] = password
|
2022-11-15 08:12:34 +00:00
|
|
|
|
2022-10-17 06:58:42 +00:00
|
|
|
self.router = APIRouter()
|
2022-10-18 06:51:53 +00:00
|
|
|
self.app = app
|
|
|
|
self.queue_lock = queue_lock
|
2022-11-15 08:12:34 +00:00
|
|
|
self.add_api_route("/sdapi/v1/txt2img", self.text2imgapi, methods=["POST"], response_model=TextToImageResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/img2img", self.img2imgapi, methods=["POST"], response_model=ImageToImageResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/extra-single-image", self.extras_single_image_api, methods=["POST"], response_model=ExtrasSingleImageResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/extra-batch-images", self.extras_batch_images_api, methods=["POST"], response_model=ExtrasBatchImagesResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/png-info", self.pnginfoapi, methods=["POST"], response_model=PNGInfoResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/progress", self.progressapi, methods=["GET"], response_model=ProgressResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/interrogate", self.interrogateapi, methods=["POST"])
|
|
|
|
self.add_api_route("/sdapi/v1/interrupt", self.interruptapi, methods=["POST"])
|
2022-11-19 12:13:07 +00:00
|
|
|
self.add_api_route("/sdapi/v1/skip", self.skip, methods=["POST"])
|
2022-11-15 08:12:34 +00:00
|
|
|
self.add_api_route("/sdapi/v1/options", self.get_config, methods=["GET"], response_model=OptionsModel)
|
|
|
|
self.add_api_route("/sdapi/v1/options", self.set_config, methods=["POST"])
|
|
|
|
self.add_api_route("/sdapi/v1/cmd-flags", self.get_cmd_flags, methods=["GET"], response_model=FlagsModel)
|
|
|
|
self.add_api_route("/sdapi/v1/samplers", self.get_samplers, methods=["GET"], response_model=List[SamplerItem])
|
|
|
|
self.add_api_route("/sdapi/v1/upscalers", self.get_upscalers, methods=["GET"], response_model=List[UpscalerItem])
|
|
|
|
self.add_api_route("/sdapi/v1/sd-models", self.get_sd_models, methods=["GET"], response_model=List[SDModelItem])
|
|
|
|
self.add_api_route("/sdapi/v1/hypernetworks", self.get_hypernetworks, methods=["GET"], response_model=List[HypernetworkItem])
|
|
|
|
self.add_api_route("/sdapi/v1/face-restorers", self.get_face_restorers, methods=["GET"], response_model=List[FaceRestorerItem])
|
|
|
|
self.add_api_route("/sdapi/v1/realesrgan-models", self.get_realesrgan_models, methods=["GET"], response_model=List[RealesrganItem])
|
2022-12-15 02:01:32 +00:00
|
|
|
self.add_api_route("/sdapi/v1/prompt-styles", self.get_prompt_styles, methods=["GET"], response_model=List[PromptStyleItem])
|
2022-11-15 08:12:34 +00:00
|
|
|
self.add_api_route("/sdapi/v1/artist-categories", self.get_artists_categories, methods=["GET"], response_model=List[str])
|
|
|
|
self.add_api_route("/sdapi/v1/artists", self.get_artists, methods=["GET"], response_model=List[ArtistItem])
|
2022-12-11 19:16:44 +00:00
|
|
|
self.add_api_route("/sdapi/v1/refresh-checkpoints", self.refresh_checkpoints, methods=["POST"])
|
2022-12-24 23:02:22 +00:00
|
|
|
self.add_api_route("/sdapi/v1/create/embedding", self.create_embedding, methods=["POST"], response_model=CreateResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/create/hypernetwork", self.create_hypernetwork, methods=["POST"], response_model=CreateResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/preprocess", self.preprocess, methods=["POST"], response_model=PreprocessResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/train/embedding", self.train_embedding, methods=["POST"], response_model=TrainResponse)
|
|
|
|
self.add_api_route("/sdapi/v1/train/hypernetwork", self.train_hypernetwork, methods=["POST"], response_model=TrainResponse)
|
2022-11-15 08:12:34 +00:00
|
|
|
|
|
|
|
def add_api_route(self, path: str, endpoint, **kwargs):
|
|
|
|
if shared.cmd_opts.api_auth:
|
|
|
|
return self.app.add_api_route(path, endpoint, dependencies=[Depends(self.auth)], **kwargs)
|
|
|
|
return self.app.add_api_route(path, endpoint, **kwargs)
|
|
|
|
|
2022-12-15 02:01:32 +00:00
|
|
|
def auth(self, credentials: HTTPBasicCredentials = Depends(HTTPBasic())):
|
|
|
|
if credentials.username in self.credentials:
|
|
|
|
if compare_digest(credentials.password, self.credentials[credentials.username]):
|
2022-11-15 08:12:34 +00:00
|
|
|
return True
|
|
|
|
|
|
|
|
raise HTTPException(status_code=401, detail="Incorrect username or password", headers={"WWW-Authenticate": "Basic"})
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-21 23:27:40 +00:00
|
|
|
def text2imgapi(self, txt2imgreq: StableDiffusionTxt2ImgProcessingAPI):
|
2022-10-17 19:10:36 +00:00
|
|
|
populate = txt2imgreq.copy(update={ # Override __init__ params
|
2022-11-27 13:12:37 +00:00
|
|
|
"sampler_name": validate_sampler_name(txt2imgreq.sampler_name or txt2imgreq.sampler_index),
|
2022-10-17 20:36:14 +00:00
|
|
|
"do_not_save_samples": True,
|
|
|
|
"do_not_save_grid": True
|
2022-10-17 19:10:36 +00:00
|
|
|
}
|
|
|
|
)
|
2022-11-27 13:19:47 +00:00
|
|
|
if populate.sampler_name:
|
|
|
|
populate.sampler_index = None # prevent a warning later on
|
2022-10-30 06:10:22 +00:00
|
|
|
|
2022-10-18 06:51:53 +00:00
|
|
|
with self.queue_lock:
|
2022-12-31 09:44:26 +00:00
|
|
|
p = StableDiffusionProcessingTxt2Img(sd_model=shared.sd_model, **vars(populate))
|
|
|
|
|
2022-12-25 10:45:44 +00:00
|
|
|
shared.state.begin()
|
2022-10-18 06:51:53 +00:00
|
|
|
processed = process_images(p)
|
2022-12-25 10:45:44 +00:00
|
|
|
shared.state.end()
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-26 14:33:45 +00:00
|
|
|
|
2022-10-23 19:01:16 +00:00
|
|
|
b64images = list(map(encode_pil_to_base64, processed.images))
|
2022-10-26 14:33:45 +00:00
|
|
|
|
2022-10-26 12:50:26 +00:00
|
|
|
return TextToImageResponse(images=b64images, parameters=vars(txt2imgreq), info=processed.js())
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-21 23:27:40 +00:00
|
|
|
def img2imgapi(self, img2imgreq: StableDiffusionImg2ImgProcessingAPI):
|
|
|
|
init_images = img2imgreq.init_images
|
|
|
|
if init_images is None:
|
2022-10-26 14:33:45 +00:00
|
|
|
raise HTTPException(status_code=404, detail="Init image not found")
|
2022-10-21 23:27:40 +00:00
|
|
|
|
2022-10-22 19:42:00 +00:00
|
|
|
mask = img2imgreq.mask
|
|
|
|
if mask:
|
2022-11-24 05:10:40 +00:00
|
|
|
mask = decode_base64_to_image(mask)
|
2022-10-22 19:42:00 +00:00
|
|
|
|
2022-10-21 23:27:40 +00:00
|
|
|
populate = img2imgreq.copy(update={ # Override __init__ params
|
2022-11-27 13:12:37 +00:00
|
|
|
"sampler_name": validate_sampler_name(img2imgreq.sampler_name or img2imgreq.sampler_index),
|
2022-10-21 23:27:40 +00:00
|
|
|
"do_not_save_samples": True,
|
2022-10-26 14:33:45 +00:00
|
|
|
"do_not_save_grid": True,
|
2022-10-22 21:10:28 +00:00
|
|
|
"mask": mask
|
2022-10-21 23:27:40 +00:00
|
|
|
}
|
|
|
|
)
|
2022-11-27 13:19:47 +00:00
|
|
|
if populate.sampler_name:
|
|
|
|
populate.sampler_index = None # prevent a warning later on
|
2022-12-03 06:15:24 +00:00
|
|
|
|
|
|
|
args = vars(populate)
|
|
|
|
args.pop('include_init_images', None) # this is meant to be done by "exclude": True in model, but it's for a reason that I cannot determine.
|
2022-10-30 06:10:22 +00:00
|
|
|
|
2022-10-21 23:27:40 +00:00
|
|
|
with self.queue_lock:
|
2022-12-31 09:44:26 +00:00
|
|
|
p = StableDiffusionProcessingImg2Img(sd_model=shared.sd_model, **args)
|
|
|
|
p.init_images = [decode_base64_to_image(x) for x in init_images]
|
|
|
|
|
2022-12-25 10:45:44 +00:00
|
|
|
shared.state.begin()
|
2022-10-21 23:27:40 +00:00
|
|
|
processed = process_images(p)
|
2022-12-25 10:45:44 +00:00
|
|
|
shared.state.end()
|
2022-10-26 14:33:45 +00:00
|
|
|
|
2022-10-23 19:01:16 +00:00
|
|
|
b64images = list(map(encode_pil_to_base64, processed.images))
|
2022-10-21 23:27:40 +00:00
|
|
|
|
2022-12-03 06:15:24 +00:00
|
|
|
if not img2imgreq.include_init_images:
|
2022-10-24 15:16:07 +00:00
|
|
|
img2imgreq.init_images = None
|
|
|
|
img2imgreq.mask = None
|
|
|
|
|
2022-10-26 12:50:26 +00:00
|
|
|
return ImageToImageResponse(images=b64images, parameters=vars(img2imgreq), info=processed.js())
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-23 02:13:32 +00:00
|
|
|
def extras_single_image_api(self, req: ExtrasSingleImageRequest):
|
2022-10-23 19:01:16 +00:00
|
|
|
reqDict = setUpscalers(req)
|
2022-10-23 02:13:32 +00:00
|
|
|
|
2022-10-23 19:01:16 +00:00
|
|
|
reqDict['image'] = decode_base64_to_image(reqDict['image'])
|
2022-10-23 02:13:32 +00:00
|
|
|
|
|
|
|
with self.queue_lock:
|
2022-12-24 11:04:53 +00:00
|
|
|
result = run_extras(extras_mode=0, image_folder="", input_dir="", output_dir="", save_output=False, **reqDict)
|
2022-10-23 02:13:32 +00:00
|
|
|
|
2022-10-23 19:03:30 +00:00
|
|
|
return ExtrasSingleImageResponse(image=encode_pil_to_base64(result[0][0]), html_info=result[1])
|
2022-10-23 16:07:59 +00:00
|
|
|
|
|
|
|
def extras_batch_images_api(self, req: ExtrasBatchImagesRequest):
|
2022-10-23 19:01:16 +00:00
|
|
|
reqDict = setUpscalers(req)
|
2022-10-23 16:07:59 +00:00
|
|
|
|
2022-10-24 11:32:18 +00:00
|
|
|
def prepareFiles(file):
|
|
|
|
file = decode_base64_to_file(file.data, file_path=file.name)
|
|
|
|
file.orig_name = file.name
|
|
|
|
return file
|
|
|
|
|
|
|
|
reqDict['image_folder'] = list(map(prepareFiles, reqDict['imageList']))
|
2022-10-23 16:07:59 +00:00
|
|
|
reqDict.pop('imageList')
|
|
|
|
|
|
|
|
with self.queue_lock:
|
2022-12-24 11:04:53 +00:00
|
|
|
result = run_extras(extras_mode=1, image="", input_dir="", output_dir="", save_output=False, **reqDict)
|
2022-10-23 16:07:59 +00:00
|
|
|
|
2022-10-23 19:03:30 +00:00
|
|
|
return ExtrasBatchImagesResponse(images=list(map(encode_pil_to_base64, result[0])), html_info=result[1])
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-29 19:10:00 +00:00
|
|
|
def pnginfoapi(self, req: PNGInfoRequest):
|
2022-10-29 19:09:19 +00:00
|
|
|
if(not req.image.strip()):
|
|
|
|
return PNGInfoResponse(info="")
|
|
|
|
|
|
|
|
result = run_pnginfo(decode_base64_to_image(req.image.strip()))
|
|
|
|
|
|
|
|
return PNGInfoResponse(info=result[1])
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-29 22:03:32 +00:00
|
|
|
def progressapi(self, req: ProgressRequest = Depends()):
|
2022-10-26 14:33:45 +00:00
|
|
|
# copy from check_progress_call of ui.py
|
|
|
|
|
|
|
|
if shared.state.job_count == 0:
|
2022-10-29 21:04:29 +00:00
|
|
|
return ProgressResponse(progress=0, eta_relative=0, state=shared.state.dict())
|
2022-10-26 14:33:45 +00:00
|
|
|
|
|
|
|
# avoid dividing zero
|
|
|
|
progress = 0.01
|
|
|
|
|
|
|
|
if shared.state.job_count > 0:
|
|
|
|
progress += shared.state.job_no / shared.state.job_count
|
|
|
|
if shared.state.sampling_steps > 0:
|
|
|
|
progress += 1 / shared.state.job_count * shared.state.sampling_step / shared.state.sampling_steps
|
|
|
|
|
|
|
|
time_since_start = time.time() - shared.state.time_start
|
|
|
|
eta = (time_since_start/progress)
|
|
|
|
eta_relative = eta-time_since_start
|
|
|
|
|
|
|
|
progress = min(progress, 1)
|
|
|
|
|
2022-11-02 09:12:32 +00:00
|
|
|
shared.state.set_current_image()
|
2022-10-30 09:02:47 +00:00
|
|
|
|
2022-10-29 21:19:17 +00:00
|
|
|
current_image = None
|
2022-10-29 22:03:32 +00:00
|
|
|
if shared.state.current_image and not req.skip_current_image:
|
2022-10-29 21:19:17 +00:00
|
|
|
current_image = encode_pil_to_base64(shared.state.current_image)
|
|
|
|
|
|
|
|
return ProgressResponse(progress=progress, eta_relative=eta_relative, state=shared.state.dict(), current_image=current_image)
|
2022-10-26 14:33:45 +00:00
|
|
|
|
2022-10-31 15:45:52 +00:00
|
|
|
def interrogateapi(self, interrogatereq: InterrogateRequest):
|
2022-10-27 19:20:15 +00:00
|
|
|
image_b64 = interrogatereq.image
|
|
|
|
if image_b64 is None:
|
2022-12-15 02:01:32 +00:00
|
|
|
raise HTTPException(status_code=404, detail="Image not found")
|
2022-10-27 19:20:15 +00:00
|
|
|
|
2022-11-06 18:32:06 +00:00
|
|
|
img = decode_base64_to_image(image_b64)
|
|
|
|
img = img.convert('RGB')
|
2022-10-27 19:20:15 +00:00
|
|
|
|
|
|
|
# Override object param
|
|
|
|
with self.queue_lock:
|
2022-11-06 18:32:06 +00:00
|
|
|
if interrogatereq.model == "clip":
|
|
|
|
processed = shared.interrogator.interrogate(img)
|
|
|
|
elif interrogatereq.model == "deepdanbooru":
|
2022-11-20 13:39:20 +00:00
|
|
|
processed = deepbooru.model.tag(img)
|
2022-11-06 18:32:06 +00:00
|
|
|
else:
|
|
|
|
raise HTTPException(status_code=404, detail="Model not found")
|
2022-12-15 02:01:32 +00:00
|
|
|
|
2022-10-31 15:45:52 +00:00
|
|
|
return InterrogateResponse(caption=processed)
|
2022-10-17 06:58:42 +00:00
|
|
|
|
2022-10-30 10:08:40 +00:00
|
|
|
def interruptapi(self):
|
|
|
|
shared.state.interrupt()
|
|
|
|
|
|
|
|
return {}
|
|
|
|
|
2022-11-05 22:05:15 +00:00
|
|
|
def skip(self):
|
|
|
|
shared.state.skip()
|
|
|
|
|
2022-11-03 03:51:22 +00:00
|
|
|
def get_config(self):
|
|
|
|
options = {}
|
|
|
|
for key in shared.opts.data.keys():
|
|
|
|
metadata = shared.opts.data_labels.get(key)
|
|
|
|
if(metadata is not None):
|
|
|
|
options.update({key: shared.opts.data.get(key, shared.opts.data_labels.get(key).default)})
|
|
|
|
else:
|
|
|
|
options.update({key: shared.opts.data.get(key, None)})
|
2022-11-04 17:43:02 +00:00
|
|
|
|
2022-11-03 03:51:22 +00:00
|
|
|
return options
|
2022-11-04 17:43:02 +00:00
|
|
|
|
2022-11-05 21:58:19 +00:00
|
|
|
def set_config(self, req: Dict[str, Any]):
|
2022-11-19 12:15:24 +00:00
|
|
|
for k, v in req.items():
|
|
|
|
shared.opts.set(k, v)
|
2022-11-03 03:51:22 +00:00
|
|
|
|
|
|
|
shared.opts.save(shared.config_filename)
|
|
|
|
return
|
|
|
|
|
|
|
|
def get_cmd_flags(self):
|
|
|
|
return vars(shared.cmd_opts)
|
|
|
|
|
|
|
|
def get_samplers(self):
|
2022-11-19 12:15:24 +00:00
|
|
|
return [{"name": sampler[0], "aliases":sampler[2], "options":sampler[3]} for sampler in sd_samplers.all_samplers]
|
2022-11-03 03:51:22 +00:00
|
|
|
|
|
|
|
def get_upscalers(self):
|
|
|
|
upscalers = []
|
2022-11-04 17:43:02 +00:00
|
|
|
|
2022-11-03 03:51:22 +00:00
|
|
|
for upscaler in shared.sd_upscalers:
|
|
|
|
u = upscaler.scaler
|
|
|
|
upscalers.append({"name":u.name, "model_name":u.model_name, "model_path":u.model_path, "model_url":u.model_url})
|
2022-11-04 17:43:02 +00:00
|
|
|
|
2022-11-03 03:51:22 +00:00
|
|
|
return upscalers
|
2022-11-04 17:43:02 +00:00
|
|
|
|
2022-11-03 03:51:22 +00:00
|
|
|
def get_sd_models(self):
|
|
|
|
return [{"title":x.title, "model_name":x.model_name, "hash":x.hash, "filename": x.filename, "config": x.config} for x in checkpoints_list.values()]
|
|
|
|
|
|
|
|
def get_hypernetworks(self):
|
|
|
|
return [{"name": name, "path": shared.hypernetworks[name]} for name in shared.hypernetworks]
|
|
|
|
|
|
|
|
def get_face_restorers(self):
|
|
|
|
return [{"name":x.name(), "cmd_dir": getattr(x, "cmd_dir", None)} for x in shared.face_restorers]
|
|
|
|
|
|
|
|
def get_realesrgan_models(self):
|
|
|
|
return [{"name":x.name,"path":x.data_path, "scale":x.scale} for x in get_realesrgan_models(None)]
|
2022-11-04 17:43:02 +00:00
|
|
|
|
2022-12-15 02:01:32 +00:00
|
|
|
def get_prompt_styles(self):
|
2022-11-03 03:51:22 +00:00
|
|
|
styleList = []
|
|
|
|
for k in shared.prompt_styles.styles:
|
2022-11-04 17:43:02 +00:00
|
|
|
style = shared.prompt_styles.styles[k]
|
2022-11-22 14:02:59 +00:00
|
|
|
styleList.append({"name":style[0], "prompt": style[1], "negative_prompt": style[2]})
|
2022-11-03 03:51:22 +00:00
|
|
|
|
|
|
|
return styleList
|
|
|
|
|
|
|
|
def get_artists_categories(self):
|
|
|
|
return shared.artist_db.cats
|
|
|
|
|
|
|
|
def get_artists(self):
|
|
|
|
return [{"name":x[0], "score":x[1], "category":x[2]} for x in shared.artist_db.artists]
|
2022-12-24 11:04:53 +00:00
|
|
|
|
2022-12-11 19:16:44 +00:00
|
|
|
def refresh_checkpoints(self):
|
|
|
|
shared.refresh_checkpoints()
|
2022-10-30 10:08:40 +00:00
|
|
|
|
2022-12-24 23:02:22 +00:00
|
|
|
def create_embedding(self, args: dict):
|
|
|
|
try:
|
|
|
|
shared.state.begin()
|
|
|
|
filename = create_embedding(**args) # create empty embedding
|
|
|
|
sd_hijack.model_hijack.embedding_db.load_textual_inversion_embeddings() # reload embeddings so new one can be immediately used
|
|
|
|
shared.state.end()
|
|
|
|
return CreateResponse(info = "create embedding filename: {filename}".format(filename = filename))
|
|
|
|
except AssertionError as e:
|
|
|
|
shared.state.end()
|
|
|
|
return TrainResponse(info = "create embedding error: {error}".format(error = e))
|
|
|
|
|
|
|
|
def create_hypernetwork(self, args: dict):
|
|
|
|
try:
|
|
|
|
shared.state.begin()
|
|
|
|
filename = create_hypernetwork(**args) # create empty embedding
|
|
|
|
shared.state.end()
|
|
|
|
return CreateResponse(info = "create hypernetwork filename: {filename}".format(filename = filename))
|
|
|
|
except AssertionError as e:
|
|
|
|
shared.state.end()
|
|
|
|
return TrainResponse(info = "create hypernetwork error: {error}".format(error = e))
|
|
|
|
|
|
|
|
def preprocess(self, args: dict):
|
|
|
|
try:
|
|
|
|
shared.state.begin()
|
|
|
|
preprocess(**args) # quick operation unless blip/booru interrogation is enabled
|
|
|
|
shared.state.end()
|
|
|
|
return PreprocessResponse(info = 'preprocess complete')
|
|
|
|
except KeyError as e:
|
|
|
|
shared.state.end()
|
|
|
|
return PreprocessResponse(info = "preprocess error: invalid token: {error}".format(error = e))
|
|
|
|
except AssertionError as e:
|
|
|
|
shared.state.end()
|
|
|
|
return PreprocessResponse(info = "preprocess error: {error}".format(error = e))
|
|
|
|
except FileNotFoundError as e:
|
|
|
|
shared.state.end()
|
|
|
|
return PreprocessResponse(info = 'preprocess error: {error}'.format(error = e))
|
|
|
|
|
|
|
|
def train_embedding(self, args: dict):
|
|
|
|
try:
|
|
|
|
shared.state.begin()
|
|
|
|
apply_optimizations = shared.opts.training_xattention_optimizations
|
|
|
|
error = None
|
|
|
|
filename = ''
|
|
|
|
if not apply_optimizations:
|
|
|
|
sd_hijack.undo_optimizations()
|
|
|
|
try:
|
|
|
|
embedding, filename = train_embedding(**args) # can take a long time to complete
|
|
|
|
except Exception as e:
|
|
|
|
error = e
|
|
|
|
finally:
|
|
|
|
if not apply_optimizations:
|
|
|
|
sd_hijack.apply_optimizations()
|
|
|
|
shared.state.end()
|
|
|
|
return TrainResponse(info = "train embedding complete: filename: {filename} error: {error}".format(filename = filename, error = error))
|
|
|
|
except AssertionError as msg:
|
|
|
|
shared.state.end()
|
|
|
|
return TrainResponse(info = "train embedding error: {msg}".format(msg = msg))
|
|
|
|
|
|
|
|
def train_hypernetwork(self, args: dict):
|
|
|
|
try:
|
|
|
|
shared.state.begin()
|
|
|
|
initial_hypernetwork = shared.loaded_hypernetwork
|
|
|
|
apply_optimizations = shared.opts.training_xattention_optimizations
|
|
|
|
error = None
|
|
|
|
filename = ''
|
|
|
|
if not apply_optimizations:
|
|
|
|
sd_hijack.undo_optimizations()
|
|
|
|
try:
|
|
|
|
hypernetwork, filename = train_hypernetwork(*args)
|
|
|
|
except Exception as e:
|
|
|
|
error = e
|
|
|
|
finally:
|
|
|
|
shared.loaded_hypernetwork = initial_hypernetwork
|
|
|
|
shared.sd_model.cond_stage_model.to(devices.device)
|
|
|
|
shared.sd_model.first_stage_model.to(devices.device)
|
|
|
|
if not apply_optimizations:
|
|
|
|
sd_hijack.apply_optimizations()
|
|
|
|
shared.state.end()
|
|
|
|
return TrainResponse(info = "train embedding complete: filename: {filename} error: {error}".format(filename = filename, error = error))
|
|
|
|
except AssertionError as msg:
|
|
|
|
shared.state.end()
|
|
|
|
return TrainResponse(info = "train embedding error: {error}".format(error = error))
|
|
|
|
|
2022-10-17 06:58:42 +00:00
|
|
|
def launch(self, server_name, port):
|
2022-10-18 06:51:53 +00:00
|
|
|
self.app.include_router(self.router)
|
|
|
|
uvicorn.run(self.app, host=server_name, port=port)
|