From f714c0db32dc7391adf18b7dec01cbcf9796e10e Mon Sep 17 00:00:00 2001 From: Last Date: Sun, 19 Feb 2023 05:48:35 +0800 Subject: [PATCH] =?UTF-8?q?=E8=BF=9B=E5=BA=A6=E6=9D=A1?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- server.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/server.py b/server.py index eda9cb1..99efca4 100644 --- a/server.py +++ b/server.py @@ -246,7 +246,7 @@ def main_dev(opt): # 如果模型不同,重新加载模型(注意释放内存) if task['ckpt'] != model_name: # 修改状态为加载模型 - update_task_status(task, "loading", 0) + update_task_status(task, "init", 0) # 获取环境配置 model_name = task['ckpt'] opt.config = f'/data/{model_name}.yaml' @@ -323,6 +323,7 @@ def main_dev(opt): prompts = list(prompts) c = model.get_learned_conditioning(prompts) shape = [opt.C, opt.H // opt.f, opt.W // opt.f] + update_task_status(task=task, status='diffusing', progress=0.5) # 修改任务状态 samples, _ = sampler.sample(S=opt.steps, conditioning=c, batch_size=task['number'], @@ -332,6 +333,7 @@ def main_dev(opt): unconditional_conditioning=uc, eta=opt.ddim_eta, x_T=start_code) + update_task_status(task=task, status='build', progress=0.8) # 修改任务状态 x_samples = model.decode_first_stage(samples) x_samples = torch.clamp((x_samples + 1.0) / 2.0, min=0.0, max=1.0) for x_sample in x_samples: