|
353 | 353 | ("Damo_XR_Lab/cv_rife_video-frame-interpolation", "flownet.pkl", "models/RIFE"), |
354 | 354 | ], |
355 | 355 | # Qwen Prompt |
356 | | - "QwenPrompt": [ |
357 | | - ("qwen/Qwen2-1.5B-Instruct", "config.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
358 | | - ("qwen/Qwen2-1.5B-Instruct", "generation_config.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
359 | | - ("qwen/Qwen2-1.5B-Instruct", "model.safetensors", "models/QwenPrompt/qwen2-1.5b-instruct"), |
360 | | - ("qwen/Qwen2-1.5B-Instruct", "special_tokens_map.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
361 | | - ("qwen/Qwen2-1.5B-Instruct", "tokenizer.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
362 | | - ("qwen/Qwen2-1.5B-Instruct", "tokenizer_config.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
363 | | - ("qwen/Qwen2-1.5B-Instruct", "merges.txt", "models/QwenPrompt/qwen2-1.5b-instruct"), |
364 | | - ("qwen/Qwen2-1.5B-Instruct", "vocab.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
365 | | - ], |
| 356 | + "QwenPrompt": { |
| 357 | + "file_list": [ |
| 358 | + ("qwen/Qwen2-1.5B-Instruct", "config.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 359 | + ("qwen/Qwen2-1.5B-Instruct", "generation_config.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 360 | + ("qwen/Qwen2-1.5B-Instruct", "model.safetensors", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 361 | + ("qwen/Qwen2-1.5B-Instruct", "special_tokens_map.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 362 | + ("qwen/Qwen2-1.5B-Instruct", "tokenizer.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 363 | + ("qwen/Qwen2-1.5B-Instruct", "tokenizer_config.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 364 | + ("qwen/Qwen2-1.5B-Instruct", "merges.txt", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 365 | + ("qwen/Qwen2-1.5B-Instruct", "vocab.json", "models/QwenPrompt/qwen2-1.5b-instruct"), |
| 366 | + ], |
| 367 | + "load_path": [ |
| 368 | + "models/QwenPrompt/qwen2-1.5b-instruct", |
| 369 | + ], |
| 370 | + }, |
366 | 371 | # Beautiful Prompt |
367 | | - "BeautifulPrompt": [ |
368 | | - ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "config.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
369 | | - ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "generation_config.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
370 | | - ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "model.safetensors", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
371 | | - ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "special_tokens_map.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
372 | | - ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "tokenizer.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
373 | | - ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "tokenizer_config.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
374 | | - ], |
| 372 | + "BeautifulPrompt": { |
| 373 | + "file_list": [ |
| 374 | + ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "config.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
| 375 | + ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "generation_config.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
| 376 | + ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "model.safetensors", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
| 377 | + ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "special_tokens_map.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
| 378 | + ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "tokenizer.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
| 379 | + ("AI-ModelScope/pai-bloom-1b1-text2prompt-sd", "tokenizer_config.json", "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd"), |
| 380 | + ], |
| 381 | + "load_path": [ |
| 382 | + "models/BeautifulPrompt/pai-bloom-1b1-text2prompt-sd", |
| 383 | + ], |
| 384 | + }, |
375 | 385 | # Omost prompt |
376 | | - "OmostPrompt":[ |
377 | | - ("Omost/omost-llama-3-8b-4bits", "model-00001-of-00002.safetensors", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
378 | | - ("Omost/omost-llama-3-8b-4bits", "model-00002-of-00002.safetensors", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
379 | | - ("Omost/omost-llama-3-8b-4bits", "tokenizer.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
380 | | - ("Omost/omost-llama-3-8b-4bits", "tokenizer_config.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
381 | | - ("Omost/omost-llama-3-8b-4bits", "config.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
382 | | - ("Omost/omost-llama-3-8b-4bits", "generation_config.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
383 | | - ("Omost/omost-llama-3-8b-4bits", "model.safetensors.index.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
384 | | - ("Omost/omost-llama-3-8b-4bits", "special_tokens_map.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
385 | | - ], |
| 386 | + "OmostPrompt": { |
| 387 | + "file_list": [ |
| 388 | + ("Omost/omost-llama-3-8b-4bits", "model-00001-of-00002.safetensors", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 389 | + ("Omost/omost-llama-3-8b-4bits", "model-00002-of-00002.safetensors", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 390 | + ("Omost/omost-llama-3-8b-4bits", "tokenizer.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 391 | + ("Omost/omost-llama-3-8b-4bits", "tokenizer_config.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 392 | + ("Omost/omost-llama-3-8b-4bits", "config.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 393 | + ("Omost/omost-llama-3-8b-4bits", "generation_config.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 394 | + ("Omost/omost-llama-3-8b-4bits", "model.safetensors.index.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 395 | + ("Omost/omost-llama-3-8b-4bits", "special_tokens_map.json", "models/OmostPrompt/omost-llama-3-8b-4bits"), |
| 396 | + ], |
| 397 | + "load_path": [ |
| 398 | + "models/OmostPrompt/omost-llama-3-8b-4bits", |
| 399 | + ], |
| 400 | + }, |
386 | 401 | # Translator |
387 | | - "opus-mt-zh-en": [ |
388 | | - ("moxying/opus-mt-zh-en", "config.json", "models/translator/opus-mt-zh-en"), |
389 | | - ("moxying/opus-mt-zh-en", "generation_config.json", "models/translator/opus-mt-zh-en"), |
390 | | - ("moxying/opus-mt-zh-en", "metadata.json", "models/translator/opus-mt-zh-en"), |
391 | | - ("moxying/opus-mt-zh-en", "pytorch_model.bin", "models/translator/opus-mt-zh-en"), |
392 | | - ("moxying/opus-mt-zh-en", "source.spm", "models/translator/opus-mt-zh-en"), |
393 | | - ("moxying/opus-mt-zh-en", "target.spm", "models/translator/opus-mt-zh-en"), |
394 | | - ("moxying/opus-mt-zh-en", "tokenizer_config.json", "models/translator/opus-mt-zh-en"), |
395 | | - ("moxying/opus-mt-zh-en", "vocab.json", "models/translator/opus-mt-zh-en"), |
396 | | - ], |
| 402 | + "opus-mt-zh-en": { |
| 403 | + "file_list": [ |
| 404 | + ("moxying/opus-mt-zh-en", "config.json", "models/translator/opus-mt-zh-en"), |
| 405 | + ("moxying/opus-mt-zh-en", "generation_config.json", "models/translator/opus-mt-zh-en"), |
| 406 | + ("moxying/opus-mt-zh-en", "metadata.json", "models/translator/opus-mt-zh-en"), |
| 407 | + ("moxying/opus-mt-zh-en", "pytorch_model.bin", "models/translator/opus-mt-zh-en"), |
| 408 | + ("moxying/opus-mt-zh-en", "source.spm", "models/translator/opus-mt-zh-en"), |
| 409 | + ("moxying/opus-mt-zh-en", "target.spm", "models/translator/opus-mt-zh-en"), |
| 410 | + ("moxying/opus-mt-zh-en", "tokenizer_config.json", "models/translator/opus-mt-zh-en"), |
| 411 | + ("moxying/opus-mt-zh-en", "vocab.json", "models/translator/opus-mt-zh-en"), |
| 412 | + ], |
| 413 | + "load_path": [ |
| 414 | + "models/translator/opus-mt-zh-en", |
| 415 | + ], |
| 416 | + }, |
397 | 417 | # IP-Adapter |
398 | 418 | "IP-Adapter-SD": [ |
399 | 419 | ("AI-ModelScope/IP-Adapter", "models/image_encoder/model.safetensors", "models/IpAdapter/stable_diffusion/image_encoder"), |
|
404 | 424 | ("AI-ModelScope/IP-Adapter", "sdxl_models/ip-adapter_sdxl.bin", "models/IpAdapter/stable_diffusion_xl"), |
405 | 425 | ], |
406 | 426 | # Kolors |
407 | | - "Kolors": [ |
408 | | - ("Kwai-Kolors/Kolors", "text_encoder/config.json", "models/kolors/Kolors/text_encoder"), |
409 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model.bin.index.json", "models/kolors/Kolors/text_encoder"), |
410 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00001-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
411 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00002-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
412 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00003-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
413 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00004-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
414 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00005-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
415 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00006-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
416 | | - ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00007-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
417 | | - ("Kwai-Kolors/Kolors", "unet/diffusion_pytorch_model.safetensors", "models/kolors/Kolors/unet"), |
418 | | - ("Kwai-Kolors/Kolors", "vae/diffusion_pytorch_model.safetensors", "models/kolors/Kolors/vae"), |
419 | | - ], |
| 427 | + "Kolors": { |
| 428 | + "file_list": [ |
| 429 | + ("Kwai-Kolors/Kolors", "text_encoder/config.json", "models/kolors/Kolors/text_encoder"), |
| 430 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model.bin.index.json", "models/kolors/Kolors/text_encoder"), |
| 431 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00001-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 432 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00002-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 433 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00003-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 434 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00004-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 435 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00005-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 436 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00006-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 437 | + ("Kwai-Kolors/Kolors", "text_encoder/pytorch_model-00007-of-00007.bin", "models/kolors/Kolors/text_encoder"), |
| 438 | + ("Kwai-Kolors/Kolors", "unet/diffusion_pytorch_model.safetensors", "models/kolors/Kolors/unet"), |
| 439 | + ("Kwai-Kolors/Kolors", "vae/diffusion_pytorch_model.safetensors", "models/kolors/Kolors/vae"), |
| 440 | + ], |
| 441 | + "load_path": [ |
| 442 | + "models/kolors/Kolors/text_encoder", |
| 443 | + "models/kolors/Kolors/unet/diffusion_pytorch_model.safetensors", |
| 444 | + "models/kolors/Kolors/vae/diffusion_pytorch_model.safetensors", |
| 445 | + ], |
| 446 | + }, |
420 | 447 | "SDXL-vae-fp16-fix": [ |
421 | 448 | ("AI-ModelScope/sdxl-vae-fp16-fix", "diffusion_pytorch_model.safetensors", "models/sdxl-vae-fp16-fix") |
422 | 449 | ], |
423 | 450 | # FLUX |
424 | | - "FLUX.1-dev": [ |
425 | | - ("AI-ModelScope/FLUX.1-dev", "text_encoder/model.safetensors", "models/FLUX/FLUX.1-dev/text_encoder"), |
426 | | - ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/config.json", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
427 | | - ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model-00001-of-00002.safetensors", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
428 | | - ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model-00002-of-00002.safetensors", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
429 | | - ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model.safetensors.index.json", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
430 | | - ("AI-ModelScope/FLUX.1-dev", "ae.safetensors", "models/FLUX/FLUX.1-dev"), |
431 | | - ("AI-ModelScope/FLUX.1-dev", "flux1-dev.safetensors", "models/FLUX/FLUX.1-dev"), |
432 | | - ], |
| 451 | + "FLUX.1-dev": { |
| 452 | + "file_list": [ |
| 453 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder/model.safetensors", "models/FLUX/FLUX.1-dev/text_encoder"), |
| 454 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/config.json", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 455 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model-00001-of-00002.safetensors", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 456 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model-00002-of-00002.safetensors", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 457 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model.safetensors.index.json", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 458 | + ("AI-ModelScope/FLUX.1-dev", "ae.safetensors", "models/FLUX/FLUX.1-dev"), |
| 459 | + ("AI-ModelScope/FLUX.1-dev", "flux1-dev.safetensors", "models/FLUX/FLUX.1-dev"), |
| 460 | + ], |
| 461 | + "load_path": [ |
| 462 | + "models/FLUX/FLUX.1-dev/text_encoder/model.safetensors", |
| 463 | + "models/FLUX/FLUX.1-dev/text_encoder_2", |
| 464 | + "models/FLUX/FLUX.1-dev/ae.safetensors", |
| 465 | + "models/FLUX/FLUX.1-dev/flux1-dev.safetensors" |
| 466 | + ], |
| 467 | + }, |
| 468 | + "FLUX.1-schnell": { |
| 469 | + "file_list": [ |
| 470 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder/model.safetensors", "models/FLUX/FLUX.1-dev/text_encoder"), |
| 471 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/config.json", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 472 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model-00001-of-00002.safetensors", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 473 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model-00002-of-00002.safetensors", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 474 | + ("AI-ModelScope/FLUX.1-dev", "text_encoder_2/model.safetensors.index.json", "models/FLUX/FLUX.1-dev/text_encoder_2"), |
| 475 | + ("AI-ModelScope/FLUX.1-dev", "ae.safetensors", "models/FLUX/FLUX.1-dev"), |
| 476 | + ("AI-ModelScope/FLUX.1-schnell", "flux1-schnell.safetensors", "models/FLUX/FLUX.1-schnell"), |
| 477 | + ], |
| 478 | + "load_path": [ |
| 479 | + "models/FLUX/FLUX.1-dev/text_encoder/model.safetensors", |
| 480 | + "models/FLUX/FLUX.1-dev/text_encoder_2", |
| 481 | + "models/FLUX/FLUX.1-dev/ae.safetensors", |
| 482 | + "models/FLUX/FLUX.1-schnell/flux1-schnell.safetensors" |
| 483 | + ], |
| 484 | + }, |
433 | 485 | # ESRGAN |
434 | 486 | "ESRGAN_x4": [ |
435 | 487 | ("AI-ModelScope/Real-ESRGAN", "RealESRGAN_x4.pth", "models/ESRGAN"), |
|
439 | 491 | ("AI-ModelScope/RIFE", "flownet.pkl", "models/RIFE"), |
440 | 492 | ], |
441 | 493 | # CogVideo |
442 | | - "CogVideoX-5B": [ |
443 | | - ("ZhipuAI/CogVideoX-5b", "text_encoder/config.json", "models/CogVideo/CogVideoX-5b/text_encoder"), |
444 | | - ("ZhipuAI/CogVideoX-5b", "text_encoder/model.safetensors.index.json", "models/CogVideo/CogVideoX-5b/text_encoder"), |
445 | | - ("ZhipuAI/CogVideoX-5b", "text_encoder/model-00001-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/text_encoder"), |
446 | | - ("ZhipuAI/CogVideoX-5b", "text_encoder/model-00002-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/text_encoder"), |
447 | | - ("ZhipuAI/CogVideoX-5b", "transformer/config.json", "models/CogVideo/CogVideoX-5b/transformer"), |
448 | | - ("ZhipuAI/CogVideoX-5b", "transformer/diffusion_pytorch_model.safetensors.index.json", "models/CogVideo/CogVideoX-5b/transformer"), |
449 | | - ("ZhipuAI/CogVideoX-5b", "transformer/diffusion_pytorch_model-00001-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/transformer"), |
450 | | - ("ZhipuAI/CogVideoX-5b", "transformer/diffusion_pytorch_model-00002-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/transformer"), |
451 | | - ("ZhipuAI/CogVideoX-5b", "vae/diffusion_pytorch_model.safetensors", "models/CogVideo/CogVideoX-5b/vae"), |
452 | | - ], |
| 494 | + "CogVideoX-5B": { |
| 495 | + "file_list": [ |
| 496 | + ("ZhipuAI/CogVideoX-5b", "text_encoder/config.json", "models/CogVideo/CogVideoX-5b/text_encoder"), |
| 497 | + ("ZhipuAI/CogVideoX-5b", "text_encoder/model.safetensors.index.json", "models/CogVideo/CogVideoX-5b/text_encoder"), |
| 498 | + ("ZhipuAI/CogVideoX-5b", "text_encoder/model-00001-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/text_encoder"), |
| 499 | + ("ZhipuAI/CogVideoX-5b", "text_encoder/model-00002-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/text_encoder"), |
| 500 | + ("ZhipuAI/CogVideoX-5b", "transformer/config.json", "models/CogVideo/CogVideoX-5b/transformer"), |
| 501 | + ("ZhipuAI/CogVideoX-5b", "transformer/diffusion_pytorch_model.safetensors.index.json", "models/CogVideo/CogVideoX-5b/transformer"), |
| 502 | + ("ZhipuAI/CogVideoX-5b", "transformer/diffusion_pytorch_model-00001-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/transformer"), |
| 503 | + ("ZhipuAI/CogVideoX-5b", "transformer/diffusion_pytorch_model-00002-of-00002.safetensors", "models/CogVideo/CogVideoX-5b/transformer"), |
| 504 | + ("ZhipuAI/CogVideoX-5b", "vae/diffusion_pytorch_model.safetensors", "models/CogVideo/CogVideoX-5b/vae"), |
| 505 | + ], |
| 506 | + "load_path": [ |
| 507 | + "models/CogVideo/CogVideoX-5b/text_encoder", |
| 508 | + "models/CogVideo/CogVideoX-5b/transformer", |
| 509 | + "models/CogVideo/CogVideoX-5b/vae/diffusion_pytorch_model.safetensors", |
| 510 | + ], |
| 511 | + }, |
453 | 512 | } |
454 | 513 | Preset_model_id: TypeAlias = Literal[ |
455 | 514 | "HunyuanDiT", |
|
481 | 540 | "SDXL-vae-fp16-fix", |
482 | 541 | "ControlNet_union_sdxl_promax", |
483 | 542 | "FLUX.1-dev", |
| 543 | + "FLUX.1-schnell", |
484 | 544 | "SDXL_lora_zyd232_ChineseInkStyle_SDXL_v1_0", |
485 | 545 | "QwenPrompt", |
486 | 546 | "OmostPrompt", |
|
0 commit comments