Add/Update backbone checkpoints (count=6)
Browse files- ds_proc.py +11 -5
- manifest_20260210_142559.json +55 -0
- models/google__efficientnet-b0/config.json +1 -1
- models/google__efficientnet-b0/ds_proc.py +11 -5
- models/google__efficientnet-b0/model.safetensors +1 -1
- models/google__vit-base-patch16-224/config.json +1 -1
- models/google__vit-base-patch16-224/ds_proc.py +11 -5
- models/google__vit-base-patch16-224/model.safetensors +1 -1
- models/microsoft__resnet-50/config.json +1 -1
- models/microsoft__resnet-50/ds_proc.py +11 -5
- models/microsoft__resnet-50/model.safetensors +1 -1
- models/microsoft__swin-tiny-patch4-window7-224/config.json +1 -1
- models/microsoft__swin-tiny-patch4-window7-224/ds_proc.py +11 -5
- models/microsoft__swin-tiny-patch4-window7-224/model.safetensors +1 -1
- models/timm__densenet121.tv_in1k/config.json +1 -1
- models/timm__densenet121.tv_in1k/ds_proc.py +11 -5
- models/timm__densenet121.tv_in1k/model.safetensors +1 -1
- models/torchvision__densenet121/config.json +1 -1
- models/torchvision__densenet121/ds_proc.py +11 -5
- models/torchvision__densenet121/model.safetensors +1 -1
ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
manifest_20260210_142559.json
ADDED
|
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"timestamp": "20260210_142559",
|
| 3 |
+
"repo_id": "dsaint31/bb_mlp_224",
|
| 4 |
+
"revision": "main",
|
| 5 |
+
"tag": null,
|
| 6 |
+
"num_labels": 3,
|
| 7 |
+
"build_device": "mps",
|
| 8 |
+
"count": 6,
|
| 9 |
+
"items": [
|
| 10 |
+
{
|
| 11 |
+
"backbone": "google/vit-base-patch16-224",
|
| 12 |
+
"subdir": "models/google__vit-base-patch16-224",
|
| 13 |
+
"dirname": "google__vit-base-patch16-224"
|
| 14 |
+
},
|
| 15 |
+
{
|
| 16 |
+
"backbone": "microsoft/swin-tiny-patch4-window7-224",
|
| 17 |
+
"subdir": "models/microsoft__swin-tiny-patch4-window7-224",
|
| 18 |
+
"dirname": "microsoft__swin-tiny-patch4-window7-224"
|
| 19 |
+
},
|
| 20 |
+
{
|
| 21 |
+
"backbone": "microsoft/resnet-50",
|
| 22 |
+
"subdir": "models/microsoft__resnet-50",
|
| 23 |
+
"dirname": "microsoft__resnet-50"
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"backbone": "google/efficientnet-b0",
|
| 27 |
+
"subdir": "models/google__efficientnet-b0",
|
| 28 |
+
"dirname": "google__efficientnet-b0"
|
| 29 |
+
},
|
| 30 |
+
{
|
| 31 |
+
"backbone": "timm/densenet121.tv_in1k",
|
| 32 |
+
"subdir": "models/timm__densenet121.tv_in1k",
|
| 33 |
+
"dirname": "timm__densenet121.tv_in1k"
|
| 34 |
+
},
|
| 35 |
+
{
|
| 36 |
+
"backbone": "torchvision/densenet121",
|
| 37 |
+
"subdir": "models/torchvision__densenet121",
|
| 38 |
+
"dirname": "torchvision__densenet121"
|
| 39 |
+
}
|
| 40 |
+
],
|
| 41 |
+
"root_code_included": true,
|
| 42 |
+
"root_code_files": [
|
| 43 |
+
"ds_proc.py",
|
| 44 |
+
"ds_model.py",
|
| 45 |
+
"ds_cfg.py",
|
| 46 |
+
"ds_meta.py"
|
| 47 |
+
],
|
| 48 |
+
"subfolder_code_included": true,
|
| 49 |
+
"subfolder_code_files": [
|
| 50 |
+
"ds_proc.py",
|
| 51 |
+
"ds_model.py",
|
| 52 |
+
"ds_cfg.py",
|
| 53 |
+
"ds_meta.py"
|
| 54 |
+
]
|
| 55 |
+
}
|
models/google__efficientnet-b0/config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
-
"created_at": "
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/google__efficientnet-b0",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
|
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
+
"created_at": "20260210_142559",
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/google__efficientnet-b0",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
models/google__efficientnet-b0/ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
models/google__efficientnet-b0/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 17558436
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7020f515e549776c727f92374975faf8bb9878444809463d1d1069e08f68d735
|
| 3 |
size 17558436
|
models/google__vit-base-patch16-224/config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
-
"created_at": "
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/google__vit-base-patch16-224",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
|
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
+
"created_at": "20260210_142559",
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/google__vit-base-patch16-224",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
models/google__vit-base-patch16-224/ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
models/google__vit-base-patch16-224/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 346372132
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c051b46d9d05d5ee0182d84ec6a01a6b2f03e4a3197e1989dbb2119b1c1554a9
|
| 3 |
size 346372132
|
models/microsoft__resnet-50/config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
-
"created_at": "
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/microsoft__resnet-50",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
|
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
+
"created_at": "20260210_142559",
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/microsoft__resnet-50",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
models/microsoft__resnet-50/ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
models/microsoft__resnet-50/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 96388660
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:672f7ce423ab8366b6988ab17a5946d3a9191438e7b2831ee0ac03e9ee9b28f0
|
| 3 |
size 96388660
|
models/microsoft__swin-tiny-patch4-window7-224/config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
-
"created_at": "
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/microsoft__swin-tiny-patch4-window7-224",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
|
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
+
"created_at": "20260210_142559",
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/microsoft__swin-tiny-patch4-window7-224",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
models/microsoft__swin-tiny-patch4-window7-224/ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
models/microsoft__swin-tiny-patch4-window7-224/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 111128348
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6a66b3ed3993cd9e93967288f9019ef2c355abdeacb5db056bd9cc5192b2624
|
| 3 |
size 111128348
|
models/timm__densenet121.tv_in1k/config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
-
"created_at": "
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/timm__densenet121.tv_in1k",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
|
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
+
"created_at": "20260210_142559",
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/timm__densenet121.tv_in1k",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
models/timm__densenet121.tv_in1k/ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
models/timm__densenet121.tv_in1k/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 29293620
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eda54a9ead50e30b80b8b0e60e9024149fd0cdeada25ea7023aa27333235090f
|
| 3 |
size 29293620
|
models/torchvision__densenet121/config.json
CHANGED
|
@@ -24,7 +24,7 @@
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
-
"created_at": "
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/torchvision__densenet121",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
|
|
|
| 24 |
"num_labels": 3,
|
| 25 |
"transformers_version": "5.1.0",
|
| 26 |
"ds_provenance": {
|
| 27 |
+
"created_at": "20260210_142559",
|
| 28 |
"repo_id": "dsaint31/bb_mlp_224",
|
| 29 |
"subdir": "models/torchvision__densenet121",
|
| 30 |
"wrapper_class": "BackboneWithMLPHeadForImageClassification",
|
models/torchvision__densenet121/ds_proc.py
CHANGED
|
@@ -128,6 +128,7 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
|
|
|
| 131 |
)
|
| 132 |
|
| 133 |
@staticmethod
|
|
@@ -257,17 +258,22 @@ class BackboneMLPHead224ImageProcessor(ImageProcessingMixin):
|
|
| 257 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 258 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 259 |
"""
|
| 260 |
-
|
| 261 |
-
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 262 |
-
if backbone is None:
|
| 263 |
-
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 264 |
-
|
| 265 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 266 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 267 |
#
|
| 268 |
# IMPORTANT:
|
| 269 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 270 |
use_fast = bool(kwargs.pop("use_fast", False))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 271 |
|
| 272 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 273 |
|
|
|
|
| 128 |
self._delegate = AutoImageProcessor.from_pretrained(
|
| 129 |
self.backbone_name_or_path,
|
| 130 |
use_fast=self.use_fast,
|
| 131 |
+
# trust_remote_code = True,
|
| 132 |
)
|
| 133 |
|
| 134 |
@staticmethod
|
|
|
|
| 258 |
- Read config.json via AutoConfig and recover backbone_name_or_path.
|
| 259 |
AutoConfig로 config.json을 읽고 backbone_name_or_path를 복구합니다.
|
| 260 |
"""
|
| 261 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 262 |
# is_training is runtime-only and should default to False for inference/serving.
|
| 263 |
# is_training은 런타임 전용이며 추론/서빙 기본값은 False가 맞습니다.
|
| 264 |
#
|
| 265 |
# IMPORTANT:
|
| 266 |
# - use_fast는 kwargs로 전달될 수 있으므로, 있으면 반영합니다.
|
| 267 |
use_fast = bool(kwargs.pop("use_fast", False))
|
| 268 |
+
|
| 269 |
+
kwargs.pop("trust_remote_code", None)
|
| 270 |
+
cfg = AutoConfig.from_pretrained(
|
| 271 |
+
pretrained_model_name_or_path,
|
| 272 |
+
trust_remote_code =True,
|
| 273 |
+
**kwargs)
|
| 274 |
+
backbone = getattr(cfg, "backbone_name_or_path", None)
|
| 275 |
+
if backbone is None:
|
| 276 |
+
raise ValueError("Cannot build processor: backbone_name_or_path not found in config.json")
|
| 277 |
|
| 278 |
return cls(backbone_name_or_path=backbone, is_training=False, use_fast=use_fast)
|
| 279 |
|
models/torchvision__densenet121/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 33394052
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e8f2bb78b2e777c1612bca3678fd638acbcba9ca4ff460616987ad3ad94dab19
|
| 3 |
size 33394052
|