Upload model_updater.json
Browse files- model_updater.json +28 -1
model_updater.json
CHANGED
|
@@ -2,6 +2,7 @@
|
|
| 2 |
"models": [
|
| 3 |
{
|
| 4 |
"name": "VN Character Sheet",
|
|
|
|
| 5 |
"hf_path": "models/loras/vn_character_sheet_v4.safetensors",
|
| 6 |
"local_path": "models/loras/vn_character_sheet_v4.safetensors",
|
| 7 |
"version": "4.0.0",
|
|
@@ -9,6 +10,7 @@
|
|
| 9 |
},
|
| 10 |
{
|
| 11 |
"name": "VN Character Sheet",
|
|
|
|
| 12 |
"hf_path": "models/loras/vn_character_sheet.safetensors",
|
| 13 |
"local_path": "models/loras/vn_character_sheet.safetensors",
|
| 14 |
"version": "1.0.0",
|
|
@@ -16,6 +18,7 @@
|
|
| 16 |
},
|
| 17 |
{
|
| 18 |
"name": "VNCCS Clothes Helper",
|
|
|
|
| 19 |
"hf_path": "models/loras/qwen/VNCCS/ClothesHelperUltimateV1_000005100.safetensors",
|
| 20 |
"local_path": "models/loras/VNCCS/ClothesHelperUltimateV1_000005100.safetensors",
|
| 21 |
"version": "1.0.0",
|
|
@@ -23,6 +26,7 @@
|
|
| 23 |
},
|
| 24 |
{
|
| 25 |
"name": "VNCCS Emotion Core",
|
|
|
|
| 26 |
"hf_path": "models/loras/qwen/VNCCS/EmotionCoreV1_000003000.safetensors",
|
| 27 |
"local_path": "models/loras/VNCCS/EmotionCoreV1_000003000.safetensors",
|
| 28 |
"version": "1.0.0",
|
|
@@ -30,6 +34,7 @@
|
|
| 30 |
},
|
| 31 |
{
|
| 32 |
"name": "VNCCS Transfer Clothes",
|
|
|
|
| 33 |
"hf_path": "models/loras/qwen/VNCCS/TransferClothes_000006700.safetensors",
|
| 34 |
"local_path": "models/loras/VNCCS/TransferClothes_000006700.safetensors",
|
| 35 |
"version": "1.0.0",
|
|
@@ -37,6 +42,7 @@
|
|
| 37 |
},
|
| 38 |
{
|
| 39 |
"name": "VNCCS Poser Helper",
|
|
|
|
| 40 |
"hf_path": "models/loras/qwen/VNCCS/poser_helper_v2_000004200.safetensors",
|
| 41 |
"local_path": "models/loras/VNCCS/poser_helper_v2_000004200.safetensors",
|
| 42 |
"version": "2.0.0",
|
|
@@ -44,6 +50,7 @@
|
|
| 44 |
},
|
| 45 |
{
|
| 46 |
"name": "ControlNet AnyTest",
|
|
|
|
| 47 |
"hf_path": "models/controlnet/SDXL/AnytestV4.safetensors",
|
| 48 |
"local_path": "models/controlnet/AnytestV4.safetensors",
|
| 49 |
"version": "4.0.0",
|
|
@@ -51,6 +58,7 @@
|
|
| 51 |
},
|
| 52 |
{
|
| 53 |
"name": "ControlNet OpenPose Illustrious",
|
|
|
|
| 54 |
"hf_path": "models/controlnet/SDXL/IllustriousXL_openpose.safetensors",
|
| 55 |
"local_path": "models/controlnet/IllustriousXL_openpose.safetensors",
|
| 56 |
"version": "1.0.0",
|
|
@@ -58,6 +66,7 @@
|
|
| 58 |
},
|
| 59 |
{
|
| 60 |
"name": "DMD2 SDXL Lightning",
|
|
|
|
| 61 |
"hf_path": "models/loras/DMD2/dmd2_sdxl_4step_lora_fp16.safetensors",
|
| 62 |
"local_path": "models/loras/dmd2_sdxl_4step_lora_fp16.safetensors",
|
| 63 |
"version": "1.0.0",
|
|
@@ -65,6 +74,7 @@
|
|
| 65 |
},
|
| 66 |
{
|
| 67 |
"name": "Mimimeter",
|
|
|
|
| 68 |
"hf_path": "models/loras/IL/mimimeter.safetensors",
|
| 69 |
"local_path": "models/loras/mimimeter.safetensors",
|
| 70 |
"version": "1.0.0",
|
|
@@ -72,6 +82,7 @@
|
|
| 72 |
},
|
| 73 |
{
|
| 74 |
"name": "YOLOv8 Face (Medium)",
|
|
|
|
| 75 |
"hf_path": "models/ultralytics/bbox/face_yolov8m.pt",
|
| 76 |
"local_path": "models/ultralytics/bbox/face_yolov8m.pt",
|
| 77 |
"version": "8.0.0",
|
|
@@ -79,6 +90,7 @@
|
|
| 79 |
},
|
| 80 |
{
|
| 81 |
"name": "YOLOv9 Face (Compact)",
|
|
|
|
| 82 |
"hf_path": "models/ultralytics/bbox/face_yolov9c.pt",
|
| 83 |
"local_path": "models/ultralytics/bbox/face_yolov9c.pt",
|
| 84 |
"version": "9.0.0",
|
|
@@ -86,6 +98,7 @@
|
|
| 86 |
},
|
| 87 |
{
|
| 88 |
"name": "YOLOv8 Hand (Small)",
|
|
|
|
| 89 |
"hf_path": "models/ultralytics/bbox/hand_yolov8s.pt",
|
| 90 |
"local_path": "models/ultralytics/bbox/hand_yolov8s.pt",
|
| 91 |
"version": "8.0.0",
|
|
@@ -93,6 +106,7 @@
|
|
| 93 |
},
|
| 94 |
{
|
| 95 |
"name": "YOLOv8 Foot (Large)",
|
|
|
|
| 96 |
"hf_path": "models/ultralytics/bbox/foot-yolov8l.pt",
|
| 97 |
"local_path": "models/ultralytics/bbox/foot-yolov8l.pt",
|
| 98 |
"version": "8.0.0",
|
|
@@ -100,6 +114,7 @@
|
|
| 100 |
},
|
| 101 |
{
|
| 102 |
"name": "YOLOv8 Seg Face (Medium)",
|
|
|
|
| 103 |
"hf_path": "models/ultralytics/segm/face_yolov8m-seg_60.pt",
|
| 104 |
"local_path": "models/ultralytics/segm/face_yolov8m-seg_60.pt",
|
| 105 |
"version": "8.0.0",
|
|
@@ -107,6 +122,7 @@
|
|
| 107 |
},
|
| 108 |
{
|
| 109 |
"name": "YOLOv8 Seg Hair (Nano)",
|
|
|
|
| 110 |
"hf_path": "models/ultralytics/segm/hair_yolov8n-seg_60.pt",
|
| 111 |
"local_path": "models/ultralytics/segm/hair_yolov8n-seg_60.pt",
|
| 112 |
"version": "8.0.0",
|
|
@@ -114,6 +130,7 @@
|
|
| 114 |
},
|
| 115 |
{
|
| 116 |
"name": "YOLOv8 Seg Person (Medium)",
|
|
|
|
| 117 |
"hf_path": "models/ultralytics/segm/person_yolov8m-seg.pt",
|
| 118 |
"local_path": "models/ultralytics/segm/person_yolov8m-seg.pt",
|
| 119 |
"version": "8.0.0",
|
|
@@ -121,6 +138,7 @@
|
|
| 121 |
},
|
| 122 |
{
|
| 123 |
"name": "SAM ViT-B (Segment Anything)",
|
|
|
|
| 124 |
"hf_path": "models/sams/sam_vit_b_01ec64.pth",
|
| 125 |
"local_path": "models/sams/sam_vit_b_01ec64.pth",
|
| 126 |
"version": "1.0.0",
|
|
@@ -128,6 +146,7 @@
|
|
| 128 |
},
|
| 129 |
{
|
| 130 |
"name": "APISR 2x RRDB",
|
|
|
|
| 131 |
"hf_path": "models/upscale_models/2x_APISR_RRDB_GAN_generator.pth",
|
| 132 |
"local_path": "models/upscale_models/2x_APISR_RRDB_GAN_generator.pth",
|
| 133 |
"version": "1.0.0",
|
|
@@ -135,11 +154,19 @@
|
|
| 135 |
},
|
| 136 |
{
|
| 137 |
"name": "APISR 4x GRL",
|
|
|
|
| 138 |
"hf_path": "models/upscale_models/4x_APISR_GRL_GAN_generator.pth",
|
| 139 |
"local_path": "models/upscale_models/4x_APISR_GRL_GAN_generator.pth",
|
| 140 |
"version": "1.0.0",
|
| 141 |
"description": "Upscaler: 4x Anime Production Oriented ISR (GRL)."
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 142 |
}
|
| 143 |
],
|
| 144 |
"config_version": "1.0"
|
| 145 |
-
}
|
|
|
|
| 2 |
"models": [
|
| 3 |
{
|
| 4 |
"name": "VN Character Sheet",
|
| 5 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 6 |
"hf_path": "models/loras/vn_character_sheet_v4.safetensors",
|
| 7 |
"local_path": "models/loras/vn_character_sheet_v4.safetensors",
|
| 8 |
"version": "4.0.0",
|
|
|
|
| 10 |
},
|
| 11 |
{
|
| 12 |
"name": "VN Character Sheet",
|
| 13 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 14 |
"hf_path": "models/loras/vn_character_sheet.safetensors",
|
| 15 |
"local_path": "models/loras/vn_character_sheet.safetensors",
|
| 16 |
"version": "1.0.0",
|
|
|
|
| 18 |
},
|
| 19 |
{
|
| 20 |
"name": "VNCCS Clothes Helper",
|
| 21 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 22 |
"hf_path": "models/loras/qwen/VNCCS/ClothesHelperUltimateV1_000005100.safetensors",
|
| 23 |
"local_path": "models/loras/VNCCS/ClothesHelperUltimateV1_000005100.safetensors",
|
| 24 |
"version": "1.0.0",
|
|
|
|
| 26 |
},
|
| 27 |
{
|
| 28 |
"name": "VNCCS Emotion Core",
|
| 29 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 30 |
"hf_path": "models/loras/qwen/VNCCS/EmotionCoreV1_000003000.safetensors",
|
| 31 |
"local_path": "models/loras/VNCCS/EmotionCoreV1_000003000.safetensors",
|
| 32 |
"version": "1.0.0",
|
|
|
|
| 34 |
},
|
| 35 |
{
|
| 36 |
"name": "VNCCS Transfer Clothes",
|
| 37 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 38 |
"hf_path": "models/loras/qwen/VNCCS/TransferClothes_000006700.safetensors",
|
| 39 |
"local_path": "models/loras/VNCCS/TransferClothes_000006700.safetensors",
|
| 40 |
"version": "1.0.0",
|
|
|
|
| 42 |
},
|
| 43 |
{
|
| 44 |
"name": "VNCCS Poser Helper",
|
| 45 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 46 |
"hf_path": "models/loras/qwen/VNCCS/poser_helper_v2_000004200.safetensors",
|
| 47 |
"local_path": "models/loras/VNCCS/poser_helper_v2_000004200.safetensors",
|
| 48 |
"version": "2.0.0",
|
|
|
|
| 50 |
},
|
| 51 |
{
|
| 52 |
"name": "ControlNet AnyTest",
|
| 53 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 54 |
"hf_path": "models/controlnet/SDXL/AnytestV4.safetensors",
|
| 55 |
"local_path": "models/controlnet/AnytestV4.safetensors",
|
| 56 |
"version": "4.0.0",
|
|
|
|
| 58 |
},
|
| 59 |
{
|
| 60 |
"name": "ControlNet OpenPose Illustrious",
|
| 61 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 62 |
"hf_path": "models/controlnet/SDXL/IllustriousXL_openpose.safetensors",
|
| 63 |
"local_path": "models/controlnet/IllustriousXL_openpose.safetensors",
|
| 64 |
"version": "1.0.0",
|
|
|
|
| 66 |
},
|
| 67 |
{
|
| 68 |
"name": "DMD2 SDXL Lightning",
|
| 69 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 70 |
"hf_path": "models/loras/DMD2/dmd2_sdxl_4step_lora_fp16.safetensors",
|
| 71 |
"local_path": "models/loras/dmd2_sdxl_4step_lora_fp16.safetensors",
|
| 72 |
"version": "1.0.0",
|
|
|
|
| 74 |
},
|
| 75 |
{
|
| 76 |
"name": "Mimimeter",
|
| 77 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 78 |
"hf_path": "models/loras/IL/mimimeter.safetensors",
|
| 79 |
"local_path": "models/loras/mimimeter.safetensors",
|
| 80 |
"version": "1.0.0",
|
|
|
|
| 82 |
},
|
| 83 |
{
|
| 84 |
"name": "YOLOv8 Face (Medium)",
|
| 85 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 86 |
"hf_path": "models/ultralytics/bbox/face_yolov8m.pt",
|
| 87 |
"local_path": "models/ultralytics/bbox/face_yolov8m.pt",
|
| 88 |
"version": "8.0.0",
|
|
|
|
| 90 |
},
|
| 91 |
{
|
| 92 |
"name": "YOLOv9 Face (Compact)",
|
| 93 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 94 |
"hf_path": "models/ultralytics/bbox/face_yolov9c.pt",
|
| 95 |
"local_path": "models/ultralytics/bbox/face_yolov9c.pt",
|
| 96 |
"version": "9.0.0",
|
|
|
|
| 98 |
},
|
| 99 |
{
|
| 100 |
"name": "YOLOv8 Hand (Small)",
|
| 101 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 102 |
"hf_path": "models/ultralytics/bbox/hand_yolov8s.pt",
|
| 103 |
"local_path": "models/ultralytics/bbox/hand_yolov8s.pt",
|
| 104 |
"version": "8.0.0",
|
|
|
|
| 106 |
},
|
| 107 |
{
|
| 108 |
"name": "YOLOv8 Foot (Large)",
|
| 109 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 110 |
"hf_path": "models/ultralytics/bbox/foot-yolov8l.pt",
|
| 111 |
"local_path": "models/ultralytics/bbox/foot-yolov8l.pt",
|
| 112 |
"version": "8.0.0",
|
|
|
|
| 114 |
},
|
| 115 |
{
|
| 116 |
"name": "YOLOv8 Seg Face (Medium)",
|
| 117 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 118 |
"hf_path": "models/ultralytics/segm/face_yolov8m-seg_60.pt",
|
| 119 |
"local_path": "models/ultralytics/segm/face_yolov8m-seg_60.pt",
|
| 120 |
"version": "8.0.0",
|
|
|
|
| 122 |
},
|
| 123 |
{
|
| 124 |
"name": "YOLOv8 Seg Hair (Nano)",
|
| 125 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 126 |
"hf_path": "models/ultralytics/segm/hair_yolov8n-seg_60.pt",
|
| 127 |
"local_path": "models/ultralytics/segm/hair_yolov8n-seg_60.pt",
|
| 128 |
"version": "8.0.0",
|
|
|
|
| 130 |
},
|
| 131 |
{
|
| 132 |
"name": "YOLOv8 Seg Person (Medium)",
|
| 133 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 134 |
"hf_path": "models/ultralytics/segm/person_yolov8m-seg.pt",
|
| 135 |
"local_path": "models/ultralytics/segm/person_yolov8m-seg.pt",
|
| 136 |
"version": "8.0.0",
|
|
|
|
| 138 |
},
|
| 139 |
{
|
| 140 |
"name": "SAM ViT-B (Segment Anything)",
|
| 141 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 142 |
"hf_path": "models/sams/sam_vit_b_01ec64.pth",
|
| 143 |
"local_path": "models/sams/sam_vit_b_01ec64.pth",
|
| 144 |
"version": "1.0.0",
|
|
|
|
| 146 |
},
|
| 147 |
{
|
| 148 |
"name": "APISR 2x RRDB",
|
| 149 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 150 |
"hf_path": "models/upscale_models/2x_APISR_RRDB_GAN_generator.pth",
|
| 151 |
"local_path": "models/upscale_models/2x_APISR_RRDB_GAN_generator.pth",
|
| 152 |
"version": "1.0.0",
|
|
|
|
| 154 |
},
|
| 155 |
{
|
| 156 |
"name": "APISR 4x GRL",
|
| 157 |
+
"hf_repo": "MIUProject/VNCCS",
|
| 158 |
"hf_path": "models/upscale_models/4x_APISR_GRL_GAN_generator.pth",
|
| 159 |
"local_path": "models/upscale_models/4x_APISR_GRL_GAN_generator.pth",
|
| 160 |
"version": "1.0.0",
|
| 161 |
"description": "Upscaler: 4x Anime Production Oriented ISR (GRL)."
|
| 162 |
+
},
|
| 163 |
+
{
|
| 164 |
+
"name": "Test Civitai Model",
|
| 165 |
+
"url": "https://civitai.com/models/929497?modelVersionId=2241189",
|
| 166 |
+
"local_path": "models/loras/qwen/aesthetic.safetensors",
|
| 167 |
+
"version": "1.0.0",
|
| 168 |
+
"description": "Downloaded from Civitai"
|
| 169 |
}
|
| 170 |
],
|
| 171 |
"config_version": "1.0"
|
| 172 |
+
}
|