Moyao001 commited on
Commit
4f4324f
·
verified ·
1 Parent(s): 6f3f85b

Add files using upload-large-folder tool

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/__init__.py +24 -0
  2. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/compat.py +229 -0
  3. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/config.py +265 -0
  4. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/defaults.py +650 -0
  5. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/instantiate.py +88 -0
  6. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/lazy.py +435 -0
  7. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/__init__.py +19 -0
  8. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/benchmark.py +225 -0
  9. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/build.py +556 -0
  10. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/catalog.py +236 -0
  11. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/common.py +301 -0
  12. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/dataset_mapper.py +191 -0
  13. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/README.md +9 -0
  14. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/__init__.py +9 -0
  15. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/builtin.py +259 -0
  16. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/builtin_meta.py +350 -0
  17. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/cityscapes.py +329 -0
  18. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/cityscapes_panoptic.py +187 -0
  19. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/coco.py +539 -0
  20. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/coco_panoptic.py +228 -0
  21. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis.py +241 -0
  22. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis_v0_5_categories.py +0 -0
  23. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis_v1_categories.py +0 -0
  24. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis_v1_category_image_count.py +20 -0
  25. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/pascal_voc.py +82 -0
  26. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/register_coco.py +3 -0
  27. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/detection_utils.py +659 -0
  28. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/samplers/grouped_batch_sampler.py +47 -0
  29. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/__init__.py +14 -0
  30. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/augmentation.py +380 -0
  31. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/augmentation_impl.py +736 -0
  32. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/transform.py +351 -0
  33. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/__init__.py +12 -0
  34. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/defaults.py +715 -0
  35. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/hooks.py +690 -0
  36. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/launch.py +123 -0
  37. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/train_loop.py +469 -0
  38. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/__init__.py +12 -0
  39. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/cityscapes_evaluation.py +197 -0
  40. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/coco_evaluation.py +722 -0
  41. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/evaluator.py +224 -0
  42. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/fast_eval_api.py +121 -0
  43. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/lvis_evaluation.py +380 -0
  44. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/panoptic_evaluation.py +199 -0
  45. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/pascal_voc_evaluation.py +300 -0
  46. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/rotated_coco_evaluation.py +207 -0
  47. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/sem_seg_evaluation.py +265 -0
  48. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/testing.py +85 -0
  49. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/model_zoo/__init__.py +10 -0
  50. CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/model_zoo/model_zoo.py +213 -0
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/__init__.py ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from .compat import downgrade_config, upgrade_config
3
+ from .config import CfgNode, get_cfg, global_cfg, set_global_cfg, configurable
4
+ from .instantiate import instantiate
5
+ from .lazy import LazyCall, LazyConfig
6
+
7
+ __all__ = [
8
+ "CfgNode",
9
+ "get_cfg",
10
+ "global_cfg",
11
+ "set_global_cfg",
12
+ "downgrade_config",
13
+ "upgrade_config",
14
+ "configurable",
15
+ "instantiate",
16
+ "LazyCall",
17
+ "LazyConfig",
18
+ ]
19
+
20
+
21
+ from annotator.oneformer.detectron2.utils.env import fixup_module_metadata
22
+
23
+ fixup_module_metadata(__name__, globals(), __all__)
24
+ del fixup_module_metadata
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/compat.py ADDED
@@ -0,0 +1,229 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ """
3
+ Backward compatibility of configs.
4
+
5
+ Instructions to bump version:
6
+ + It's not needed to bump version if new keys are added.
7
+ It's only needed when backward-incompatible changes happen
8
+ (i.e., some existing keys disappear, or the meaning of a key changes)
9
+ + To bump version, do the following:
10
+ 1. Increment _C.VERSION in defaults.py
11
+ 2. Add a converter in this file.
12
+
13
+ Each ConverterVX has a function "upgrade" which in-place upgrades config from X-1 to X,
14
+ and a function "downgrade" which in-place downgrades config from X to X-1
15
+
16
+ In each function, VERSION is left unchanged.
17
+
18
+ Each converter assumes that its input has the relevant keys
19
+ (i.e., the input is not a partial config).
20
+ 3. Run the tests (test_config.py) to make sure the upgrade & downgrade
21
+ functions are consistent.
22
+ """
23
+
24
+ import logging
25
+ from typing import List, Optional, Tuple
26
+
27
+ from .config import CfgNode as CN
28
+ from .defaults import _C
29
+
30
+ __all__ = ["upgrade_config", "downgrade_config"]
31
+
32
+
33
+ def upgrade_config(cfg: CN, to_version: Optional[int] = None) -> CN:
34
+ """
35
+ Upgrade a config from its current version to a newer version.
36
+
37
+ Args:
38
+ cfg (CfgNode):
39
+ to_version (int): defaults to the latest version.
40
+ """
41
+ cfg = cfg.clone()
42
+ if to_version is None:
43
+ to_version = _C.VERSION
44
+
45
+ assert cfg.VERSION <= to_version, "Cannot upgrade from v{} to v{}!".format(
46
+ cfg.VERSION, to_version
47
+ )
48
+ for k in range(cfg.VERSION, to_version):
49
+ converter = globals()["ConverterV" + str(k + 1)]
50
+ converter.upgrade(cfg)
51
+ cfg.VERSION = k + 1
52
+ return cfg
53
+
54
+
55
+ def downgrade_config(cfg: CN, to_version: int) -> CN:
56
+ """
57
+ Downgrade a config from its current version to an older version.
58
+
59
+ Args:
60
+ cfg (CfgNode):
61
+ to_version (int):
62
+
63
+ Note:
64
+ A general downgrade of arbitrary configs is not always possible due to the
65
+ different functionalities in different versions.
66
+ The purpose of downgrade is only to recover the defaults in old versions,
67
+ allowing it to load an old partial yaml config.
68
+ Therefore, the implementation only needs to fill in the default values
69
+ in the old version when a general downgrade is not possible.
70
+ """
71
+ cfg = cfg.clone()
72
+ assert cfg.VERSION >= to_version, "Cannot downgrade from v{} to v{}!".format(
73
+ cfg.VERSION, to_version
74
+ )
75
+ for k in range(cfg.VERSION, to_version, -1):
76
+ converter = globals()["ConverterV" + str(k)]
77
+ converter.downgrade(cfg)
78
+ cfg.VERSION = k - 1
79
+ return cfg
80
+
81
+
82
+ def guess_version(cfg: CN, filename: str) -> int:
83
+ """
84
+ Guess the version of a partial config where the VERSION field is not specified.
85
+ Returns the version, or the latest if cannot make a guess.
86
+
87
+ This makes it easier for users to migrate.
88
+ """
89
+ logger = logging.getLogger(__name__)
90
+
91
+ def _has(name: str) -> bool:
92
+ cur = cfg
93
+ for n in name.split("."):
94
+ if n not in cur:
95
+ return False
96
+ cur = cur[n]
97
+ return True
98
+
99
+ # Most users' partial configs have "MODEL.WEIGHT", so guess on it
100
+ ret = None
101
+ if _has("MODEL.WEIGHT") or _has("TEST.AUG_ON"):
102
+ ret = 1
103
+
104
+ if ret is not None:
105
+ logger.warning("Config '{}' has no VERSION. Assuming it to be v{}.".format(filename, ret))
106
+ else:
107
+ ret = _C.VERSION
108
+ logger.warning(
109
+ "Config '{}' has no VERSION. Assuming it to be compatible with latest v{}.".format(
110
+ filename, ret
111
+ )
112
+ )
113
+ return ret
114
+
115
+
116
+ def _rename(cfg: CN, old: str, new: str) -> None:
117
+ old_keys = old.split(".")
118
+ new_keys = new.split(".")
119
+
120
+ def _set(key_seq: List[str], val: str) -> None:
121
+ cur = cfg
122
+ for k in key_seq[:-1]:
123
+ if k not in cur:
124
+ cur[k] = CN()
125
+ cur = cur[k]
126
+ cur[key_seq[-1]] = val
127
+
128
+ def _get(key_seq: List[str]) -> CN:
129
+ cur = cfg
130
+ for k in key_seq:
131
+ cur = cur[k]
132
+ return cur
133
+
134
+ def _del(key_seq: List[str]) -> None:
135
+ cur = cfg
136
+ for k in key_seq[:-1]:
137
+ cur = cur[k]
138
+ del cur[key_seq[-1]]
139
+ if len(cur) == 0 and len(key_seq) > 1:
140
+ _del(key_seq[:-1])
141
+
142
+ _set(new_keys, _get(old_keys))
143
+ _del(old_keys)
144
+
145
+
146
+ class _RenameConverter:
147
+ """
148
+ A converter that handles simple rename.
149
+ """
150
+
151
+ RENAME: List[Tuple[str, str]] = [] # list of tuples of (old name, new name)
152
+
153
+ @classmethod
154
+ def upgrade(cls, cfg: CN) -> None:
155
+ for old, new in cls.RENAME:
156
+ _rename(cfg, old, new)
157
+
158
+ @classmethod
159
+ def downgrade(cls, cfg: CN) -> None:
160
+ for old, new in cls.RENAME[::-1]:
161
+ _rename(cfg, new, old)
162
+
163
+
164
+ class ConverterV1(_RenameConverter):
165
+ RENAME = [("MODEL.RPN_HEAD.NAME", "MODEL.RPN.HEAD_NAME")]
166
+
167
+
168
+ class ConverterV2(_RenameConverter):
169
+ """
170
+ A large bulk of rename, before public release.
171
+ """
172
+
173
+ RENAME = [
174
+ ("MODEL.WEIGHT", "MODEL.WEIGHTS"),
175
+ ("MODEL.PANOPTIC_FPN.SEMANTIC_LOSS_SCALE", "MODEL.SEM_SEG_HEAD.LOSS_WEIGHT"),
176
+ ("MODEL.PANOPTIC_FPN.RPN_LOSS_SCALE", "MODEL.RPN.LOSS_WEIGHT"),
177
+ ("MODEL.PANOPTIC_FPN.INSTANCE_LOSS_SCALE", "MODEL.PANOPTIC_FPN.INSTANCE_LOSS_WEIGHT"),
178
+ ("MODEL.PANOPTIC_FPN.COMBINE_ON", "MODEL.PANOPTIC_FPN.COMBINE.ENABLED"),
179
+ (
180
+ "MODEL.PANOPTIC_FPN.COMBINE_OVERLAP_THRESHOLD",
181
+ "MODEL.PANOPTIC_FPN.COMBINE.OVERLAP_THRESH",
182
+ ),
183
+ (
184
+ "MODEL.PANOPTIC_FPN.COMBINE_STUFF_AREA_LIMIT",
185
+ "MODEL.PANOPTIC_FPN.COMBINE.STUFF_AREA_LIMIT",
186
+ ),
187
+ (
188
+ "MODEL.PANOPTIC_FPN.COMBINE_INSTANCES_CONFIDENCE_THRESHOLD",
189
+ "MODEL.PANOPTIC_FPN.COMBINE.INSTANCES_CONFIDENCE_THRESH",
190
+ ),
191
+ ("MODEL.ROI_HEADS.SCORE_THRESH", "MODEL.ROI_HEADS.SCORE_THRESH_TEST"),
192
+ ("MODEL.ROI_HEADS.NMS", "MODEL.ROI_HEADS.NMS_THRESH_TEST"),
193
+ ("MODEL.RETINANET.INFERENCE_SCORE_THRESHOLD", "MODEL.RETINANET.SCORE_THRESH_TEST"),
194
+ ("MODEL.RETINANET.INFERENCE_TOPK_CANDIDATES", "MODEL.RETINANET.TOPK_CANDIDATES_TEST"),
195
+ ("MODEL.RETINANET.INFERENCE_NMS_THRESHOLD", "MODEL.RETINANET.NMS_THRESH_TEST"),
196
+ ("TEST.DETECTIONS_PER_IMG", "TEST.DETECTIONS_PER_IMAGE"),
197
+ ("TEST.AUG_ON", "TEST.AUG.ENABLED"),
198
+ ("TEST.AUG_MIN_SIZES", "TEST.AUG.MIN_SIZES"),
199
+ ("TEST.AUG_MAX_SIZE", "TEST.AUG.MAX_SIZE"),
200
+ ("TEST.AUG_FLIP", "TEST.AUG.FLIP"),
201
+ ]
202
+
203
+ @classmethod
204
+ def upgrade(cls, cfg: CN) -> None:
205
+ super().upgrade(cfg)
206
+
207
+ if cfg.MODEL.META_ARCHITECTURE == "RetinaNet":
208
+ _rename(
209
+ cfg, "MODEL.RETINANET.ANCHOR_ASPECT_RATIOS", "MODEL.ANCHOR_GENERATOR.ASPECT_RATIOS"
210
+ )
211
+ _rename(cfg, "MODEL.RETINANET.ANCHOR_SIZES", "MODEL.ANCHOR_GENERATOR.SIZES")
212
+ del cfg["MODEL"]["RPN"]["ANCHOR_SIZES"]
213
+ del cfg["MODEL"]["RPN"]["ANCHOR_ASPECT_RATIOS"]
214
+ else:
215
+ _rename(cfg, "MODEL.RPN.ANCHOR_ASPECT_RATIOS", "MODEL.ANCHOR_GENERATOR.ASPECT_RATIOS")
216
+ _rename(cfg, "MODEL.RPN.ANCHOR_SIZES", "MODEL.ANCHOR_GENERATOR.SIZES")
217
+ del cfg["MODEL"]["RETINANET"]["ANCHOR_SIZES"]
218
+ del cfg["MODEL"]["RETINANET"]["ANCHOR_ASPECT_RATIOS"]
219
+ del cfg["MODEL"]["RETINANET"]["ANCHOR_STRIDES"]
220
+
221
+ @classmethod
222
+ def downgrade(cls, cfg: CN) -> None:
223
+ super().downgrade(cfg)
224
+
225
+ _rename(cfg, "MODEL.ANCHOR_GENERATOR.ASPECT_RATIOS", "MODEL.RPN.ANCHOR_ASPECT_RATIOS")
226
+ _rename(cfg, "MODEL.ANCHOR_GENERATOR.SIZES", "MODEL.RPN.ANCHOR_SIZES")
227
+ cfg.MODEL.RETINANET.ANCHOR_ASPECT_RATIOS = cfg.MODEL.RPN.ANCHOR_ASPECT_RATIOS
228
+ cfg.MODEL.RETINANET.ANCHOR_SIZES = cfg.MODEL.RPN.ANCHOR_SIZES
229
+ cfg.MODEL.RETINANET.ANCHOR_STRIDES = [] # this is not used anywhere in any version
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/config.py ADDED
@@ -0,0 +1,265 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ import functools
5
+ import inspect
6
+ import logging
7
+ from fvcore.common.config import CfgNode as _CfgNode
8
+
9
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
10
+
11
+
12
+ class CfgNode(_CfgNode):
13
+ """
14
+ The same as `fvcore.common.config.CfgNode`, but different in:
15
+
16
+ 1. Use unsafe yaml loading by default.
17
+ Note that this may lead to arbitrary code execution: you must not
18
+ load a config file from untrusted sources before manually inspecting
19
+ the content of the file.
20
+ 2. Support config versioning.
21
+ When attempting to merge an old config, it will convert the old config automatically.
22
+
23
+ .. automethod:: clone
24
+ .. automethod:: freeze
25
+ .. automethod:: defrost
26
+ .. automethod:: is_frozen
27
+ .. automethod:: load_yaml_with_base
28
+ .. automethod:: merge_from_list
29
+ .. automethod:: merge_from_other_cfg
30
+ """
31
+
32
+ @classmethod
33
+ def _open_cfg(cls, filename):
34
+ return PathManager.open(filename, "r")
35
+
36
+ # Note that the default value of allow_unsafe is changed to True
37
+ def merge_from_file(self, cfg_filename: str, allow_unsafe: bool = True) -> None:
38
+ """
39
+ Load content from the given config file and merge it into self.
40
+
41
+ Args:
42
+ cfg_filename: config filename
43
+ allow_unsafe: allow unsafe yaml syntax
44
+ """
45
+ assert PathManager.isfile(cfg_filename), f"Config file '{cfg_filename}' does not exist!"
46
+ loaded_cfg = self.load_yaml_with_base(cfg_filename, allow_unsafe=allow_unsafe)
47
+ loaded_cfg = type(self)(loaded_cfg)
48
+
49
+ # defaults.py needs to import CfgNode
50
+ from .defaults import _C
51
+
52
+ latest_ver = _C.VERSION
53
+ assert (
54
+ latest_ver == self.VERSION
55
+ ), "CfgNode.merge_from_file is only allowed on a config object of latest version!"
56
+
57
+ logger = logging.getLogger(__name__)
58
+
59
+ loaded_ver = loaded_cfg.get("VERSION", None)
60
+ if loaded_ver is None:
61
+ from .compat import guess_version
62
+
63
+ loaded_ver = guess_version(loaded_cfg, cfg_filename)
64
+ assert loaded_ver <= self.VERSION, "Cannot merge a v{} config into a v{} config.".format(
65
+ loaded_ver, self.VERSION
66
+ )
67
+
68
+ if loaded_ver == self.VERSION:
69
+ self.merge_from_other_cfg(loaded_cfg)
70
+ else:
71
+ # compat.py needs to import CfgNode
72
+ from .compat import upgrade_config, downgrade_config
73
+
74
+ logger.warning(
75
+ "Loading an old v{} config file '{}' by automatically upgrading to v{}. "
76
+ "See docs/CHANGELOG.md for instructions to update your files.".format(
77
+ loaded_ver, cfg_filename, self.VERSION
78
+ )
79
+ )
80
+ # To convert, first obtain a full config at an old version
81
+ old_self = downgrade_config(self, to_version=loaded_ver)
82
+ old_self.merge_from_other_cfg(loaded_cfg)
83
+ new_config = upgrade_config(old_self)
84
+ self.clear()
85
+ self.update(new_config)
86
+
87
+ def dump(self, *args, **kwargs):
88
+ """
89
+ Returns:
90
+ str: a yaml string representation of the config
91
+ """
92
+ # to make it show up in docs
93
+ return super().dump(*args, **kwargs)
94
+
95
+
96
+ global_cfg = CfgNode()
97
+
98
+
99
+ def get_cfg() -> CfgNode:
100
+ """
101
+ Get a copy of the default config.
102
+
103
+ Returns:
104
+ a detectron2 CfgNode instance.
105
+ """
106
+ from .defaults import _C
107
+
108
+ return _C.clone()
109
+
110
+
111
+ def set_global_cfg(cfg: CfgNode) -> None:
112
+ """
113
+ Let the global config point to the given cfg.
114
+
115
+ Assume that the given "cfg" has the key "KEY", after calling
116
+ `set_global_cfg(cfg)`, the key can be accessed by:
117
+ ::
118
+ from annotator.oneformer.detectron2.config import global_cfg
119
+ print(global_cfg.KEY)
120
+
121
+ By using a hacky global config, you can access these configs anywhere,
122
+ without having to pass the config object or the values deep into the code.
123
+ This is a hacky feature introduced for quick prototyping / research exploration.
124
+ """
125
+ global global_cfg
126
+ global_cfg.clear()
127
+ global_cfg.update(cfg)
128
+
129
+
130
+ def configurable(init_func=None, *, from_config=None):
131
+ """
132
+ Decorate a function or a class's __init__ method so that it can be called
133
+ with a :class:`CfgNode` object using a :func:`from_config` function that translates
134
+ :class:`CfgNode` to arguments.
135
+
136
+ Examples:
137
+ ::
138
+ # Usage 1: Decorator on __init__:
139
+ class A:
140
+ @configurable
141
+ def __init__(self, a, b=2, c=3):
142
+ pass
143
+
144
+ @classmethod
145
+ def from_config(cls, cfg): # 'cfg' must be the first argument
146
+ # Returns kwargs to be passed to __init__
147
+ return {"a": cfg.A, "b": cfg.B}
148
+
149
+ a1 = A(a=1, b=2) # regular construction
150
+ a2 = A(cfg) # construct with a cfg
151
+ a3 = A(cfg, b=3, c=4) # construct with extra overwrite
152
+
153
+ # Usage 2: Decorator on any function. Needs an extra from_config argument:
154
+ @configurable(from_config=lambda cfg: {"a: cfg.A, "b": cfg.B})
155
+ def a_func(a, b=2, c=3):
156
+ pass
157
+
158
+ a1 = a_func(a=1, b=2) # regular call
159
+ a2 = a_func(cfg) # call with a cfg
160
+ a3 = a_func(cfg, b=3, c=4) # call with extra overwrite
161
+
162
+ Args:
163
+ init_func (callable): a class's ``__init__`` method in usage 1. The
164
+ class must have a ``from_config`` classmethod which takes `cfg` as
165
+ the first argument.
166
+ from_config (callable): the from_config function in usage 2. It must take `cfg`
167
+ as its first argument.
168
+ """
169
+
170
+ if init_func is not None:
171
+ assert (
172
+ inspect.isfunction(init_func)
173
+ and from_config is None
174
+ and init_func.__name__ == "__init__"
175
+ ), "Incorrect use of @configurable. Check API documentation for examples."
176
+
177
+ @functools.wraps(init_func)
178
+ def wrapped(self, *args, **kwargs):
179
+ try:
180
+ from_config_func = type(self).from_config
181
+ except AttributeError as e:
182
+ raise AttributeError(
183
+ "Class with @configurable must have a 'from_config' classmethod."
184
+ ) from e
185
+ if not inspect.ismethod(from_config_func):
186
+ raise TypeError("Class with @configurable must have a 'from_config' classmethod.")
187
+
188
+ if _called_with_cfg(*args, **kwargs):
189
+ explicit_args = _get_args_from_config(from_config_func, *args, **kwargs)
190
+ init_func(self, **explicit_args)
191
+ else:
192
+ init_func(self, *args, **kwargs)
193
+
194
+ return wrapped
195
+
196
+ else:
197
+ if from_config is None:
198
+ return configurable # @configurable() is made equivalent to @configurable
199
+ assert inspect.isfunction(
200
+ from_config
201
+ ), "from_config argument of configurable must be a function!"
202
+
203
+ def wrapper(orig_func):
204
+ @functools.wraps(orig_func)
205
+ def wrapped(*args, **kwargs):
206
+ if _called_with_cfg(*args, **kwargs):
207
+ explicit_args = _get_args_from_config(from_config, *args, **kwargs)
208
+ return orig_func(**explicit_args)
209
+ else:
210
+ return orig_func(*args, **kwargs)
211
+
212
+ wrapped.from_config = from_config
213
+ return wrapped
214
+
215
+ return wrapper
216
+
217
+
218
+ def _get_args_from_config(from_config_func, *args, **kwargs):
219
+ """
220
+ Use `from_config` to obtain explicit arguments.
221
+
222
+ Returns:
223
+ dict: arguments to be used for cls.__init__
224
+ """
225
+ signature = inspect.signature(from_config_func)
226
+ if list(signature.parameters.keys())[0] != "cfg":
227
+ if inspect.isfunction(from_config_func):
228
+ name = from_config_func.__name__
229
+ else:
230
+ name = f"{from_config_func.__self__}.from_config"
231
+ raise TypeError(f"{name} must take 'cfg' as the first argument!")
232
+ support_var_arg = any(
233
+ param.kind in [param.VAR_POSITIONAL, param.VAR_KEYWORD]
234
+ for param in signature.parameters.values()
235
+ )
236
+ if support_var_arg: # forward all arguments to from_config, if from_config accepts them
237
+ ret = from_config_func(*args, **kwargs)
238
+ else:
239
+ # forward supported arguments to from_config
240
+ supported_arg_names = set(signature.parameters.keys())
241
+ extra_kwargs = {}
242
+ for name in list(kwargs.keys()):
243
+ if name not in supported_arg_names:
244
+ extra_kwargs[name] = kwargs.pop(name)
245
+ ret = from_config_func(*args, **kwargs)
246
+ # forward the other arguments to __init__
247
+ ret.update(extra_kwargs)
248
+ return ret
249
+
250
+
251
+ def _called_with_cfg(*args, **kwargs):
252
+ """
253
+ Returns:
254
+ bool: whether the arguments contain CfgNode and should be considered
255
+ forwarded to from_config.
256
+ """
257
+ from omegaconf import DictConfig
258
+
259
+ if len(args) and isinstance(args[0], (_CfgNode, DictConfig)):
260
+ return True
261
+ if isinstance(kwargs.pop("cfg", None), (_CfgNode, DictConfig)):
262
+ return True
263
+ # `from_config`'s first argument is forced to be "cfg".
264
+ # So the above check covers all cases.
265
+ return False
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/defaults.py ADDED
@@ -0,0 +1,650 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from .config import CfgNode as CN
3
+
4
+ # NOTE: given the new config system
5
+ # (https://detectron2.readthedocs.io/en/latest/tutorials/lazyconfigs.html),
6
+ # we will stop adding new functionalities to default CfgNode.
7
+
8
+ # -----------------------------------------------------------------------------
9
+ # Convention about Training / Test specific parameters
10
+ # -----------------------------------------------------------------------------
11
+ # Whenever an argument can be either used for training or for testing, the
12
+ # corresponding name will be post-fixed by a _TRAIN for a training parameter,
13
+ # or _TEST for a test-specific parameter.
14
+ # For example, the number of images during training will be
15
+ # IMAGES_PER_BATCH_TRAIN, while the number of images for testing will be
16
+ # IMAGES_PER_BATCH_TEST
17
+
18
+ # -----------------------------------------------------------------------------
19
+ # Config definition
20
+ # -----------------------------------------------------------------------------
21
+
22
+ _C = CN()
23
+
24
+ # The version number, to upgrade from old configs to new ones if any
25
+ # changes happen. It's recommended to keep a VERSION in your config file.
26
+ _C.VERSION = 2
27
+
28
+ _C.MODEL = CN()
29
+ _C.MODEL.LOAD_PROPOSALS = False
30
+ _C.MODEL.MASK_ON = False
31
+ _C.MODEL.KEYPOINT_ON = False
32
+ _C.MODEL.DEVICE = "cuda"
33
+ _C.MODEL.META_ARCHITECTURE = "GeneralizedRCNN"
34
+
35
+ # Path (a file path, or URL like detectron2://.., https://..) to a checkpoint file
36
+ # to be loaded to the model. You can find available models in the model zoo.
37
+ _C.MODEL.WEIGHTS = ""
38
+
39
+ # Values to be used for image normalization (BGR order, since INPUT.FORMAT defaults to BGR).
40
+ # To train on images of different number of channels, just set different mean & std.
41
+ # Default values are the mean pixel value from ImageNet: [103.53, 116.28, 123.675]
42
+ _C.MODEL.PIXEL_MEAN = [103.530, 116.280, 123.675]
43
+ # When using pre-trained models in Detectron1 or any MSRA models,
44
+ # std has been absorbed into its conv1 weights, so the std needs to be set 1.
45
+ # Otherwise, you can use [57.375, 57.120, 58.395] (ImageNet std)
46
+ _C.MODEL.PIXEL_STD = [1.0, 1.0, 1.0]
47
+
48
+
49
+ # -----------------------------------------------------------------------------
50
+ # INPUT
51
+ # -----------------------------------------------------------------------------
52
+ _C.INPUT = CN()
53
+ # By default, {MIN,MAX}_SIZE options are used in transforms.ResizeShortestEdge.
54
+ # Please refer to ResizeShortestEdge for detailed definition.
55
+ # Size of the smallest side of the image during training
56
+ _C.INPUT.MIN_SIZE_TRAIN = (800,)
57
+ # Sample size of smallest side by choice or random selection from range give by
58
+ # INPUT.MIN_SIZE_TRAIN
59
+ _C.INPUT.MIN_SIZE_TRAIN_SAMPLING = "choice"
60
+ # Maximum size of the side of the image during training
61
+ _C.INPUT.MAX_SIZE_TRAIN = 1333
62
+ # Size of the smallest side of the image during testing. Set to zero to disable resize in testing.
63
+ _C.INPUT.MIN_SIZE_TEST = 800
64
+ # Maximum size of the side of the image during testing
65
+ _C.INPUT.MAX_SIZE_TEST = 1333
66
+ # Mode for flipping images used in data augmentation during training
67
+ # choose one of ["horizontal, "vertical", "none"]
68
+ _C.INPUT.RANDOM_FLIP = "horizontal"
69
+
70
+ # `True` if cropping is used for data augmentation during training
71
+ _C.INPUT.CROP = CN({"ENABLED": False})
72
+ # Cropping type. See documentation of `detectron2.data.transforms.RandomCrop` for explanation.
73
+ _C.INPUT.CROP.TYPE = "relative_range"
74
+ # Size of crop in range (0, 1] if CROP.TYPE is "relative" or "relative_range" and in number of
75
+ # pixels if CROP.TYPE is "absolute"
76
+ _C.INPUT.CROP.SIZE = [0.9, 0.9]
77
+
78
+
79
+ # Whether the model needs RGB, YUV, HSV etc.
80
+ # Should be one of the modes defined here, as we use PIL to read the image:
81
+ # https://pillow.readthedocs.io/en/stable/handbook/concepts.html#concept-modes
82
+ # with BGR being the one exception. One can set image format to BGR, we will
83
+ # internally use RGB for conversion and flip the channels over
84
+ _C.INPUT.FORMAT = "BGR"
85
+ # The ground truth mask format that the model will use.
86
+ # Mask R-CNN supports either "polygon" or "bitmask" as ground truth.
87
+ _C.INPUT.MASK_FORMAT = "polygon" # alternative: "bitmask"
88
+
89
+
90
+ # -----------------------------------------------------------------------------
91
+ # Dataset
92
+ # -----------------------------------------------------------------------------
93
+ _C.DATASETS = CN()
94
+ # List of the dataset names for training. Must be registered in DatasetCatalog
95
+ # Samples from these datasets will be merged and used as one dataset.
96
+ _C.DATASETS.TRAIN = ()
97
+ # List of the pre-computed proposal files for training, which must be consistent
98
+ # with datasets listed in DATASETS.TRAIN.
99
+ _C.DATASETS.PROPOSAL_FILES_TRAIN = ()
100
+ # Number of top scoring precomputed proposals to keep for training
101
+ _C.DATASETS.PRECOMPUTED_PROPOSAL_TOPK_TRAIN = 2000
102
+ # List of the dataset names for testing. Must be registered in DatasetCatalog
103
+ _C.DATASETS.TEST = ()
104
+ # List of the pre-computed proposal files for test, which must be consistent
105
+ # with datasets listed in DATASETS.TEST.
106
+ _C.DATASETS.PROPOSAL_FILES_TEST = ()
107
+ # Number of top scoring precomputed proposals to keep for test
108
+ _C.DATASETS.PRECOMPUTED_PROPOSAL_TOPK_TEST = 1000
109
+
110
+ # -----------------------------------------------------------------------------
111
+ # DataLoader
112
+ # -----------------------------------------------------------------------------
113
+ _C.DATALOADER = CN()
114
+ # Number of data loading threads
115
+ _C.DATALOADER.NUM_WORKERS = 4
116
+ # If True, each batch should contain only images for which the aspect ratio
117
+ # is compatible. This groups portrait images together, and landscape images
118
+ # are not batched with portrait images.
119
+ _C.DATALOADER.ASPECT_RATIO_GROUPING = True
120
+ # Options: TrainingSampler, RepeatFactorTrainingSampler
121
+ _C.DATALOADER.SAMPLER_TRAIN = "TrainingSampler"
122
+ # Repeat threshold for RepeatFactorTrainingSampler
123
+ _C.DATALOADER.REPEAT_THRESHOLD = 0.0
124
+ # Tf True, when working on datasets that have instance annotations, the
125
+ # training dataloader will filter out images without associated annotations
126
+ _C.DATALOADER.FILTER_EMPTY_ANNOTATIONS = True
127
+
128
+ # ---------------------------------------------------------------------------- #
129
+ # Backbone options
130
+ # ---------------------------------------------------------------------------- #
131
+ _C.MODEL.BACKBONE = CN()
132
+
133
+ _C.MODEL.BACKBONE.NAME = "build_resnet_backbone"
134
+ # Freeze the first several stages so they are not trained.
135
+ # There are 5 stages in ResNet. The first is a convolution, and the following
136
+ # stages are each group of residual blocks.
137
+ _C.MODEL.BACKBONE.FREEZE_AT = 2
138
+
139
+
140
+ # ---------------------------------------------------------------------------- #
141
+ # FPN options
142
+ # ---------------------------------------------------------------------------- #
143
+ _C.MODEL.FPN = CN()
144
+ # Names of the input feature maps to be used by FPN
145
+ # They must have contiguous power of 2 strides
146
+ # e.g., ["res2", "res3", "res4", "res5"]
147
+ _C.MODEL.FPN.IN_FEATURES = []
148
+ _C.MODEL.FPN.OUT_CHANNELS = 256
149
+
150
+ # Options: "" (no norm), "GN"
151
+ _C.MODEL.FPN.NORM = ""
152
+
153
+ # Types for fusing the FPN top-down and lateral features. Can be either "sum" or "avg"
154
+ _C.MODEL.FPN.FUSE_TYPE = "sum"
155
+
156
+
157
+ # ---------------------------------------------------------------------------- #
158
+ # Proposal generator options
159
+ # ---------------------------------------------------------------------------- #
160
+ _C.MODEL.PROPOSAL_GENERATOR = CN()
161
+ # Current proposal generators include "RPN", "RRPN" and "PrecomputedProposals"
162
+ _C.MODEL.PROPOSAL_GENERATOR.NAME = "RPN"
163
+ # Proposal height and width both need to be greater than MIN_SIZE
164
+ # (a the scale used during training or inference)
165
+ _C.MODEL.PROPOSAL_GENERATOR.MIN_SIZE = 0
166
+
167
+
168
+ # ---------------------------------------------------------------------------- #
169
+ # Anchor generator options
170
+ # ---------------------------------------------------------------------------- #
171
+ _C.MODEL.ANCHOR_GENERATOR = CN()
172
+ # The generator can be any name in the ANCHOR_GENERATOR registry
173
+ _C.MODEL.ANCHOR_GENERATOR.NAME = "DefaultAnchorGenerator"
174
+ # Anchor sizes (i.e. sqrt of area) in absolute pixels w.r.t. the network input.
175
+ # Format: list[list[float]]. SIZES[i] specifies the list of sizes to use for
176
+ # IN_FEATURES[i]; len(SIZES) must be equal to len(IN_FEATURES) or 1.
177
+ # When len(SIZES) == 1, SIZES[0] is used for all IN_FEATURES.
178
+ _C.MODEL.ANCHOR_GENERATOR.SIZES = [[32, 64, 128, 256, 512]]
179
+ # Anchor aspect ratios. For each area given in `SIZES`, anchors with different aspect
180
+ # ratios are generated by an anchor generator.
181
+ # Format: list[list[float]]. ASPECT_RATIOS[i] specifies the list of aspect ratios (H/W)
182
+ # to use for IN_FEATURES[i]; len(ASPECT_RATIOS) == len(IN_FEATURES) must be true,
183
+ # or len(ASPECT_RATIOS) == 1 is true and aspect ratio list ASPECT_RATIOS[0] is used
184
+ # for all IN_FEATURES.
185
+ _C.MODEL.ANCHOR_GENERATOR.ASPECT_RATIOS = [[0.5, 1.0, 2.0]]
186
+ # Anchor angles.
187
+ # list[list[float]], the angle in degrees, for each input feature map.
188
+ # ANGLES[i] specifies the list of angles for IN_FEATURES[i].
189
+ _C.MODEL.ANCHOR_GENERATOR.ANGLES = [[-90, 0, 90]]
190
+ # Relative offset between the center of the first anchor and the top-left corner of the image
191
+ # Value has to be in [0, 1). Recommend to use 0.5, which means half stride.
192
+ # The value is not expected to affect model accuracy.
193
+ _C.MODEL.ANCHOR_GENERATOR.OFFSET = 0.0
194
+
195
+ # ---------------------------------------------------------------------------- #
196
+ # RPN options
197
+ # ---------------------------------------------------------------------------- #
198
+ _C.MODEL.RPN = CN()
199
+ _C.MODEL.RPN.HEAD_NAME = "StandardRPNHead" # used by RPN_HEAD_REGISTRY
200
+
201
+ # Names of the input feature maps to be used by RPN
202
+ # e.g., ["p2", "p3", "p4", "p5", "p6"] for FPN
203
+ _C.MODEL.RPN.IN_FEATURES = ["res4"]
204
+ # Remove RPN anchors that go outside the image by BOUNDARY_THRESH pixels
205
+ # Set to -1 or a large value, e.g. 100000, to disable pruning anchors
206
+ _C.MODEL.RPN.BOUNDARY_THRESH = -1
207
+ # IOU overlap ratios [BG_IOU_THRESHOLD, FG_IOU_THRESHOLD]
208
+ # Minimum overlap required between an anchor and ground-truth box for the
209
+ # (anchor, gt box) pair to be a positive example (IoU >= FG_IOU_THRESHOLD
210
+ # ==> positive RPN example: 1)
211
+ # Maximum overlap allowed between an anchor and ground-truth box for the
212
+ # (anchor, gt box) pair to be a negative examples (IoU < BG_IOU_THRESHOLD
213
+ # ==> negative RPN example: 0)
214
+ # Anchors with overlap in between (BG_IOU_THRESHOLD <= IoU < FG_IOU_THRESHOLD)
215
+ # are ignored (-1)
216
+ _C.MODEL.RPN.IOU_THRESHOLDS = [0.3, 0.7]
217
+ _C.MODEL.RPN.IOU_LABELS = [0, -1, 1]
218
+ # Number of regions per image used to train RPN
219
+ _C.MODEL.RPN.BATCH_SIZE_PER_IMAGE = 256
220
+ # Target fraction of foreground (positive) examples per RPN minibatch
221
+ _C.MODEL.RPN.POSITIVE_FRACTION = 0.5
222
+ # Options are: "smooth_l1", "giou", "diou", "ciou"
223
+ _C.MODEL.RPN.BBOX_REG_LOSS_TYPE = "smooth_l1"
224
+ _C.MODEL.RPN.BBOX_REG_LOSS_WEIGHT = 1.0
225
+ # Weights on (dx, dy, dw, dh) for normalizing RPN anchor regression targets
226
+ _C.MODEL.RPN.BBOX_REG_WEIGHTS = (1.0, 1.0, 1.0, 1.0)
227
+ # The transition point from L1 to L2 loss. Set to 0.0 to make the loss simply L1.
228
+ _C.MODEL.RPN.SMOOTH_L1_BETA = 0.0
229
+ _C.MODEL.RPN.LOSS_WEIGHT = 1.0
230
+ # Number of top scoring RPN proposals to keep before applying NMS
231
+ # When FPN is used, this is *per FPN level* (not total)
232
+ _C.MODEL.RPN.PRE_NMS_TOPK_TRAIN = 12000
233
+ _C.MODEL.RPN.PRE_NMS_TOPK_TEST = 6000
234
+ # Number of top scoring RPN proposals to keep after applying NMS
235
+ # When FPN is used, this limit is applied per level and then again to the union
236
+ # of proposals from all levels
237
+ # NOTE: When FPN is used, the meaning of this config is different from Detectron1.
238
+ # It means per-batch topk in Detectron1, but per-image topk here.
239
+ # See the "find_top_rpn_proposals" function for details.
240
+ _C.MODEL.RPN.POST_NMS_TOPK_TRAIN = 2000
241
+ _C.MODEL.RPN.POST_NMS_TOPK_TEST = 1000
242
+ # NMS threshold used on RPN proposals
243
+ _C.MODEL.RPN.NMS_THRESH = 0.7
244
+ # Set this to -1 to use the same number of output channels as input channels.
245
+ _C.MODEL.RPN.CONV_DIMS = [-1]
246
+
247
+ # ---------------------------------------------------------------------------- #
248
+ # ROI HEADS options
249
+ # ---------------------------------------------------------------------------- #
250
+ _C.MODEL.ROI_HEADS = CN()
251
+ _C.MODEL.ROI_HEADS.NAME = "Res5ROIHeads"
252
+ # Number of foreground classes
253
+ _C.MODEL.ROI_HEADS.NUM_CLASSES = 80
254
+ # Names of the input feature maps to be used by ROI heads
255
+ # Currently all heads (box, mask, ...) use the same input feature map list
256
+ # e.g., ["p2", "p3", "p4", "p5"] is commonly used for FPN
257
+ _C.MODEL.ROI_HEADS.IN_FEATURES = ["res4"]
258
+ # IOU overlap ratios [IOU_THRESHOLD]
259
+ # Overlap threshold for an RoI to be considered background (if < IOU_THRESHOLD)
260
+ # Overlap threshold for an RoI to be considered foreground (if >= IOU_THRESHOLD)
261
+ _C.MODEL.ROI_HEADS.IOU_THRESHOLDS = [0.5]
262
+ _C.MODEL.ROI_HEADS.IOU_LABELS = [0, 1]
263
+ # RoI minibatch size *per image* (number of regions of interest [ROIs]) during training
264
+ # Total number of RoIs per training minibatch =
265
+ # ROI_HEADS.BATCH_SIZE_PER_IMAGE * SOLVER.IMS_PER_BATCH
266
+ # E.g., a common configuration is: 512 * 16 = 8192
267
+ _C.MODEL.ROI_HEADS.BATCH_SIZE_PER_IMAGE = 512
268
+ # Target fraction of RoI minibatch that is labeled foreground (i.e. class > 0)
269
+ _C.MODEL.ROI_HEADS.POSITIVE_FRACTION = 0.25
270
+
271
+ # Only used on test mode
272
+
273
+ # Minimum score threshold (assuming scores in a [0, 1] range); a value chosen to
274
+ # balance obtaining high recall with not having too many low precision
275
+ # detections that will slow down inference post processing steps (like NMS)
276
+ # A default threshold of 0.0 increases AP by ~0.2-0.3 but significantly slows down
277
+ # inference.
278
+ _C.MODEL.ROI_HEADS.SCORE_THRESH_TEST = 0.05
279
+ # Overlap threshold used for non-maximum suppression (suppress boxes with
280
+ # IoU >= this threshold)
281
+ _C.MODEL.ROI_HEADS.NMS_THRESH_TEST = 0.5
282
+ # If True, augment proposals with ground-truth boxes before sampling proposals to
283
+ # train ROI heads.
284
+ _C.MODEL.ROI_HEADS.PROPOSAL_APPEND_GT = True
285
+
286
+ # ---------------------------------------------------------------------------- #
287
+ # Box Head
288
+ # ---------------------------------------------------------------------------- #
289
+ _C.MODEL.ROI_BOX_HEAD = CN()
290
+ # C4 don't use head name option
291
+ # Options for non-C4 models: FastRCNNConvFCHead,
292
+ _C.MODEL.ROI_BOX_HEAD.NAME = ""
293
+ # Options are: "smooth_l1", "giou", "diou", "ciou"
294
+ _C.MODEL.ROI_BOX_HEAD.BBOX_REG_LOSS_TYPE = "smooth_l1"
295
+ # The final scaling coefficient on the box regression loss, used to balance the magnitude of its
296
+ # gradients with other losses in the model. See also `MODEL.ROI_KEYPOINT_HEAD.LOSS_WEIGHT`.
297
+ _C.MODEL.ROI_BOX_HEAD.BBOX_REG_LOSS_WEIGHT = 1.0
298
+ # Default weights on (dx, dy, dw, dh) for normalizing bbox regression targets
299
+ # These are empirically chosen to approximately lead to unit variance targets
300
+ _C.MODEL.ROI_BOX_HEAD.BBOX_REG_WEIGHTS = (10.0, 10.0, 5.0, 5.0)
301
+ # The transition point from L1 to L2 loss. Set to 0.0 to make the loss simply L1.
302
+ _C.MODEL.ROI_BOX_HEAD.SMOOTH_L1_BETA = 0.0
303
+ _C.MODEL.ROI_BOX_HEAD.POOLER_RESOLUTION = 14
304
+ _C.MODEL.ROI_BOX_HEAD.POOLER_SAMPLING_RATIO = 0
305
+ # Type of pooling operation applied to the incoming feature map for each RoI
306
+ _C.MODEL.ROI_BOX_HEAD.POOLER_TYPE = "ROIAlignV2"
307
+
308
+ _C.MODEL.ROI_BOX_HEAD.NUM_FC = 0
309
+ # Hidden layer dimension for FC layers in the RoI box head
310
+ _C.MODEL.ROI_BOX_HEAD.FC_DIM = 1024
311
+ _C.MODEL.ROI_BOX_HEAD.NUM_CONV = 0
312
+ # Channel dimension for Conv layers in the RoI box head
313
+ _C.MODEL.ROI_BOX_HEAD.CONV_DIM = 256
314
+ # Normalization method for the convolution layers.
315
+ # Options: "" (no norm), "GN", "SyncBN".
316
+ _C.MODEL.ROI_BOX_HEAD.NORM = ""
317
+ # Whether to use class agnostic for bbox regression
318
+ _C.MODEL.ROI_BOX_HEAD.CLS_AGNOSTIC_BBOX_REG = False
319
+ # If true, RoI heads use bounding boxes predicted by the box head rather than proposal boxes.
320
+ _C.MODEL.ROI_BOX_HEAD.TRAIN_ON_PRED_BOXES = False
321
+
322
+ # Federated loss can be used to improve the training of LVIS
323
+ _C.MODEL.ROI_BOX_HEAD.USE_FED_LOSS = False
324
+ # Sigmoid cross entrophy is used with federated loss
325
+ _C.MODEL.ROI_BOX_HEAD.USE_SIGMOID_CE = False
326
+ # The power value applied to image_count when calcualting frequency weight
327
+ _C.MODEL.ROI_BOX_HEAD.FED_LOSS_FREQ_WEIGHT_POWER = 0.5
328
+ # Number of classes to keep in total
329
+ _C.MODEL.ROI_BOX_HEAD.FED_LOSS_NUM_CLASSES = 50
330
+
331
+ # ---------------------------------------------------------------------------- #
332
+ # Cascaded Box Head
333
+ # ---------------------------------------------------------------------------- #
334
+ _C.MODEL.ROI_BOX_CASCADE_HEAD = CN()
335
+ # The number of cascade stages is implicitly defined by the length of the following two configs.
336
+ _C.MODEL.ROI_BOX_CASCADE_HEAD.BBOX_REG_WEIGHTS = (
337
+ (10.0, 10.0, 5.0, 5.0),
338
+ (20.0, 20.0, 10.0, 10.0),
339
+ (30.0, 30.0, 15.0, 15.0),
340
+ )
341
+ _C.MODEL.ROI_BOX_CASCADE_HEAD.IOUS = (0.5, 0.6, 0.7)
342
+
343
+
344
+ # ---------------------------------------------------------------------------- #
345
+ # Mask Head
346
+ # ---------------------------------------------------------------------------- #
347
+ _C.MODEL.ROI_MASK_HEAD = CN()
348
+ _C.MODEL.ROI_MASK_HEAD.NAME = "MaskRCNNConvUpsampleHead"
349
+ _C.MODEL.ROI_MASK_HEAD.POOLER_RESOLUTION = 14
350
+ _C.MODEL.ROI_MASK_HEAD.POOLER_SAMPLING_RATIO = 0
351
+ _C.MODEL.ROI_MASK_HEAD.NUM_CONV = 0 # The number of convs in the mask head
352
+ _C.MODEL.ROI_MASK_HEAD.CONV_DIM = 256
353
+ # Normalization method for the convolution layers.
354
+ # Options: "" (no norm), "GN", "SyncBN".
355
+ _C.MODEL.ROI_MASK_HEAD.NORM = ""
356
+ # Whether to use class agnostic for mask prediction
357
+ _C.MODEL.ROI_MASK_HEAD.CLS_AGNOSTIC_MASK = False
358
+ # Type of pooling operation applied to the incoming feature map for each RoI
359
+ _C.MODEL.ROI_MASK_HEAD.POOLER_TYPE = "ROIAlignV2"
360
+
361
+
362
+ # ---------------------------------------------------------------------------- #
363
+ # Keypoint Head
364
+ # ---------------------------------------------------------------------------- #
365
+ _C.MODEL.ROI_KEYPOINT_HEAD = CN()
366
+ _C.MODEL.ROI_KEYPOINT_HEAD.NAME = "KRCNNConvDeconvUpsampleHead"
367
+ _C.MODEL.ROI_KEYPOINT_HEAD.POOLER_RESOLUTION = 14
368
+ _C.MODEL.ROI_KEYPOINT_HEAD.POOLER_SAMPLING_RATIO = 0
369
+ _C.MODEL.ROI_KEYPOINT_HEAD.CONV_DIMS = tuple(512 for _ in range(8))
370
+ _C.MODEL.ROI_KEYPOINT_HEAD.NUM_KEYPOINTS = 17 # 17 is the number of keypoints in COCO.
371
+
372
+ # Images with too few (or no) keypoints are excluded from training.
373
+ _C.MODEL.ROI_KEYPOINT_HEAD.MIN_KEYPOINTS_PER_IMAGE = 1
374
+ # Normalize by the total number of visible keypoints in the minibatch if True.
375
+ # Otherwise, normalize by the total number of keypoints that could ever exist
376
+ # in the minibatch.
377
+ # The keypoint softmax loss is only calculated on visible keypoints.
378
+ # Since the number of visible keypoints can vary significantly between
379
+ # minibatches, this has the effect of up-weighting the importance of
380
+ # minibatches with few visible keypoints. (Imagine the extreme case of
381
+ # only one visible keypoint versus N: in the case of N, each one
382
+ # contributes 1/N to the gradient compared to the single keypoint
383
+ # determining the gradient direction). Instead, we can normalize the
384
+ # loss by the total number of keypoints, if it were the case that all
385
+ # keypoints were visible in a full minibatch. (Returning to the example,
386
+ # this means that the one visible keypoint contributes as much as each
387
+ # of the N keypoints.)
388
+ _C.MODEL.ROI_KEYPOINT_HEAD.NORMALIZE_LOSS_BY_VISIBLE_KEYPOINTS = True
389
+ # Multi-task loss weight to use for keypoints
390
+ # Recommended values:
391
+ # - use 1.0 if NORMALIZE_LOSS_BY_VISIBLE_KEYPOINTS is True
392
+ # - use 4.0 if NORMALIZE_LOSS_BY_VISIBLE_KEYPOINTS is False
393
+ _C.MODEL.ROI_KEYPOINT_HEAD.LOSS_WEIGHT = 1.0
394
+ # Type of pooling operation applied to the incoming feature map for each RoI
395
+ _C.MODEL.ROI_KEYPOINT_HEAD.POOLER_TYPE = "ROIAlignV2"
396
+
397
+ # ---------------------------------------------------------------------------- #
398
+ # Semantic Segmentation Head
399
+ # ---------------------------------------------------------------------------- #
400
+ _C.MODEL.SEM_SEG_HEAD = CN()
401
+ _C.MODEL.SEM_SEG_HEAD.NAME = "SemSegFPNHead"
402
+ _C.MODEL.SEM_SEG_HEAD.IN_FEATURES = ["p2", "p3", "p4", "p5"]
403
+ # Label in the semantic segmentation ground truth that is ignored, i.e., no loss is calculated for
404
+ # the correposnding pixel.
405
+ _C.MODEL.SEM_SEG_HEAD.IGNORE_VALUE = 255
406
+ # Number of classes in the semantic segmentation head
407
+ _C.MODEL.SEM_SEG_HEAD.NUM_CLASSES = 54
408
+ # Number of channels in the 3x3 convs inside semantic-FPN heads.
409
+ _C.MODEL.SEM_SEG_HEAD.CONVS_DIM = 128
410
+ # Outputs from semantic-FPN heads are up-scaled to the COMMON_STRIDE stride.
411
+ _C.MODEL.SEM_SEG_HEAD.COMMON_STRIDE = 4
412
+ # Normalization method for the convolution layers. Options: "" (no norm), "GN".
413
+ _C.MODEL.SEM_SEG_HEAD.NORM = "GN"
414
+ _C.MODEL.SEM_SEG_HEAD.LOSS_WEIGHT = 1.0
415
+
416
+ _C.MODEL.PANOPTIC_FPN = CN()
417
+ # Scaling of all losses from instance detection / segmentation head.
418
+ _C.MODEL.PANOPTIC_FPN.INSTANCE_LOSS_WEIGHT = 1.0
419
+
420
+ # options when combining instance & semantic segmentation outputs
421
+ _C.MODEL.PANOPTIC_FPN.COMBINE = CN({"ENABLED": True}) # "COMBINE.ENABLED" is deprecated & not used
422
+ _C.MODEL.PANOPTIC_FPN.COMBINE.OVERLAP_THRESH = 0.5
423
+ _C.MODEL.PANOPTIC_FPN.COMBINE.STUFF_AREA_LIMIT = 4096
424
+ _C.MODEL.PANOPTIC_FPN.COMBINE.INSTANCES_CONFIDENCE_THRESH = 0.5
425
+
426
+
427
+ # ---------------------------------------------------------------------------- #
428
+ # RetinaNet Head
429
+ # ---------------------------------------------------------------------------- #
430
+ _C.MODEL.RETINANET = CN()
431
+
432
+ # This is the number of foreground classes.
433
+ _C.MODEL.RETINANET.NUM_CLASSES = 80
434
+
435
+ _C.MODEL.RETINANET.IN_FEATURES = ["p3", "p4", "p5", "p6", "p7"]
436
+
437
+ # Convolutions to use in the cls and bbox tower
438
+ # NOTE: this doesn't include the last conv for logits
439
+ _C.MODEL.RETINANET.NUM_CONVS = 4
440
+
441
+ # IoU overlap ratio [bg, fg] for labeling anchors.
442
+ # Anchors with < bg are labeled negative (0)
443
+ # Anchors with >= bg and < fg are ignored (-1)
444
+ # Anchors with >= fg are labeled positive (1)
445
+ _C.MODEL.RETINANET.IOU_THRESHOLDS = [0.4, 0.5]
446
+ _C.MODEL.RETINANET.IOU_LABELS = [0, -1, 1]
447
+
448
+ # Prior prob for rare case (i.e. foreground) at the beginning of training.
449
+ # This is used to set the bias for the logits layer of the classifier subnet.
450
+ # This improves training stability in the case of heavy class imbalance.
451
+ _C.MODEL.RETINANET.PRIOR_PROB = 0.01
452
+
453
+ # Inference cls score threshold, only anchors with score > INFERENCE_TH are
454
+ # considered for inference (to improve speed)
455
+ _C.MODEL.RETINANET.SCORE_THRESH_TEST = 0.05
456
+ # Select topk candidates before NMS
457
+ _C.MODEL.RETINANET.TOPK_CANDIDATES_TEST = 1000
458
+ _C.MODEL.RETINANET.NMS_THRESH_TEST = 0.5
459
+
460
+ # Weights on (dx, dy, dw, dh) for normalizing Retinanet anchor regression targets
461
+ _C.MODEL.RETINANET.BBOX_REG_WEIGHTS = (1.0, 1.0, 1.0, 1.0)
462
+
463
+ # Loss parameters
464
+ _C.MODEL.RETINANET.FOCAL_LOSS_GAMMA = 2.0
465
+ _C.MODEL.RETINANET.FOCAL_LOSS_ALPHA = 0.25
466
+ _C.MODEL.RETINANET.SMOOTH_L1_LOSS_BETA = 0.1
467
+ # Options are: "smooth_l1", "giou", "diou", "ciou"
468
+ _C.MODEL.RETINANET.BBOX_REG_LOSS_TYPE = "smooth_l1"
469
+
470
+ # One of BN, SyncBN, FrozenBN, GN
471
+ # Only supports GN until unshared norm is implemented
472
+ _C.MODEL.RETINANET.NORM = ""
473
+
474
+
475
+ # ---------------------------------------------------------------------------- #
476
+ # ResNe[X]t options (ResNets = {ResNet, ResNeXt}
477
+ # Note that parts of a resnet may be used for both the backbone and the head
478
+ # These options apply to both
479
+ # ---------------------------------------------------------------------------- #
480
+ _C.MODEL.RESNETS = CN()
481
+
482
+ _C.MODEL.RESNETS.DEPTH = 50
483
+ _C.MODEL.RESNETS.OUT_FEATURES = ["res4"] # res4 for C4 backbone, res2..5 for FPN backbone
484
+
485
+ # Number of groups to use; 1 ==> ResNet; > 1 ==> ResNeXt
486
+ _C.MODEL.RESNETS.NUM_GROUPS = 1
487
+
488
+ # Options: FrozenBN, GN, "SyncBN", "BN"
489
+ _C.MODEL.RESNETS.NORM = "FrozenBN"
490
+
491
+ # Baseline width of each group.
492
+ # Scaling this parameters will scale the width of all bottleneck layers.
493
+ _C.MODEL.RESNETS.WIDTH_PER_GROUP = 64
494
+
495
+ # Place the stride 2 conv on the 1x1 filter
496
+ # Use True only for the original MSRA ResNet; use False for C2 and Torch models
497
+ _C.MODEL.RESNETS.STRIDE_IN_1X1 = True
498
+
499
+ # Apply dilation in stage "res5"
500
+ _C.MODEL.RESNETS.RES5_DILATION = 1
501
+
502
+ # Output width of res2. Scaling this parameters will scale the width of all 1x1 convs in ResNet
503
+ # For R18 and R34, this needs to be set to 64
504
+ _C.MODEL.RESNETS.RES2_OUT_CHANNELS = 256
505
+ _C.MODEL.RESNETS.STEM_OUT_CHANNELS = 64
506
+
507
+ # Apply Deformable Convolution in stages
508
+ # Specify if apply deform_conv on Res2, Res3, Res4, Res5
509
+ _C.MODEL.RESNETS.DEFORM_ON_PER_STAGE = [False, False, False, False]
510
+ # Use True to use modulated deform_conv (DeformableV2, https://arxiv.org/abs/1811.11168);
511
+ # Use False for DeformableV1.
512
+ _C.MODEL.RESNETS.DEFORM_MODULATED = False
513
+ # Number of groups in deformable conv.
514
+ _C.MODEL.RESNETS.DEFORM_NUM_GROUPS = 1
515
+
516
+
517
+ # ---------------------------------------------------------------------------- #
518
+ # Solver
519
+ # ---------------------------------------------------------------------------- #
520
+ _C.SOLVER = CN()
521
+
522
+ # Options: WarmupMultiStepLR, WarmupCosineLR.
523
+ # See detectron2/solver/build.py for definition.
524
+ _C.SOLVER.LR_SCHEDULER_NAME = "WarmupMultiStepLR"
525
+
526
+ _C.SOLVER.MAX_ITER = 40000
527
+
528
+ _C.SOLVER.BASE_LR = 0.001
529
+ # The end lr, only used by WarmupCosineLR
530
+ _C.SOLVER.BASE_LR_END = 0.0
531
+
532
+ _C.SOLVER.MOMENTUM = 0.9
533
+
534
+ _C.SOLVER.NESTEROV = False
535
+
536
+ _C.SOLVER.WEIGHT_DECAY = 0.0001
537
+ # The weight decay that's applied to parameters of normalization layers
538
+ # (typically the affine transformation)
539
+ _C.SOLVER.WEIGHT_DECAY_NORM = 0.0
540
+
541
+ _C.SOLVER.GAMMA = 0.1
542
+ # The iteration number to decrease learning rate by GAMMA.
543
+ _C.SOLVER.STEPS = (30000,)
544
+ # Number of decays in WarmupStepWithFixedGammaLR schedule
545
+ _C.SOLVER.NUM_DECAYS = 3
546
+
547
+ _C.SOLVER.WARMUP_FACTOR = 1.0 / 1000
548
+ _C.SOLVER.WARMUP_ITERS = 1000
549
+ _C.SOLVER.WARMUP_METHOD = "linear"
550
+ # Whether to rescale the interval for the learning schedule after warmup
551
+ _C.SOLVER.RESCALE_INTERVAL = False
552
+
553
+ # Save a checkpoint after every this number of iterations
554
+ _C.SOLVER.CHECKPOINT_PERIOD = 5000
555
+
556
+ # Number of images per batch across all machines. This is also the number
557
+ # of training images per step (i.e. per iteration). If we use 16 GPUs
558
+ # and IMS_PER_BATCH = 32, each GPU will see 2 images per batch.
559
+ # May be adjusted automatically if REFERENCE_WORLD_SIZE is set.
560
+ _C.SOLVER.IMS_PER_BATCH = 16
561
+
562
+ # The reference number of workers (GPUs) this config is meant to train with.
563
+ # It takes no effect when set to 0.
564
+ # With a non-zero value, it will be used by DefaultTrainer to compute a desired
565
+ # per-worker batch size, and then scale the other related configs (total batch size,
566
+ # learning rate, etc) to match the per-worker batch size.
567
+ # See documentation of `DefaultTrainer.auto_scale_workers` for details:
568
+ _C.SOLVER.REFERENCE_WORLD_SIZE = 0
569
+
570
+ # Detectron v1 (and previous detection code) used a 2x higher LR and 0 WD for
571
+ # biases. This is not useful (at least for recent models). You should avoid
572
+ # changing these and they exist only to reproduce Detectron v1 training if
573
+ # desired.
574
+ _C.SOLVER.BIAS_LR_FACTOR = 1.0
575
+ _C.SOLVER.WEIGHT_DECAY_BIAS = None # None means following WEIGHT_DECAY
576
+
577
+ # Gradient clipping
578
+ _C.SOLVER.CLIP_GRADIENTS = CN({"ENABLED": False})
579
+ # Type of gradient clipping, currently 2 values are supported:
580
+ # - "value": the absolute values of elements of each gradients are clipped
581
+ # - "norm": the norm of the gradient for each parameter is clipped thus
582
+ # affecting all elements in the parameter
583
+ _C.SOLVER.CLIP_GRADIENTS.CLIP_TYPE = "value"
584
+ # Maximum absolute value used for clipping gradients
585
+ _C.SOLVER.CLIP_GRADIENTS.CLIP_VALUE = 1.0
586
+ # Floating point number p for L-p norm to be used with the "norm"
587
+ # gradient clipping type; for L-inf, please specify .inf
588
+ _C.SOLVER.CLIP_GRADIENTS.NORM_TYPE = 2.0
589
+
590
+ # Enable automatic mixed precision for training
591
+ # Note that this does not change model's inference behavior.
592
+ # To use AMP in inference, run inference under autocast()
593
+ _C.SOLVER.AMP = CN({"ENABLED": False})
594
+
595
+ # ---------------------------------------------------------------------------- #
596
+ # Specific test options
597
+ # ---------------------------------------------------------------------------- #
598
+ _C.TEST = CN()
599
+ # For end-to-end tests to verify the expected accuracy.
600
+ # Each item is [task, metric, value, tolerance]
601
+ # e.g.: [['bbox', 'AP', 38.5, 0.2]]
602
+ _C.TEST.EXPECTED_RESULTS = []
603
+ # The period (in terms of steps) to evaluate the model during training.
604
+ # Set to 0 to disable.
605
+ _C.TEST.EVAL_PERIOD = 0
606
+ # The sigmas used to calculate keypoint OKS. See http://cocodataset.org/#keypoints-eval
607
+ # When empty, it will use the defaults in COCO.
608
+ # Otherwise it should be a list[float] with the same length as ROI_KEYPOINT_HEAD.NUM_KEYPOINTS.
609
+ _C.TEST.KEYPOINT_OKS_SIGMAS = []
610
+ # Maximum number of detections to return per image during inference (100 is
611
+ # based on the limit established for the COCO dataset).
612
+ _C.TEST.DETECTIONS_PER_IMAGE = 100
613
+
614
+ _C.TEST.AUG = CN({"ENABLED": False})
615
+ _C.TEST.AUG.MIN_SIZES = (400, 500, 600, 700, 800, 900, 1000, 1100, 1200)
616
+ _C.TEST.AUG.MAX_SIZE = 4000
617
+ _C.TEST.AUG.FLIP = True
618
+
619
+ _C.TEST.PRECISE_BN = CN({"ENABLED": False})
620
+ _C.TEST.PRECISE_BN.NUM_ITER = 200
621
+
622
+ # ---------------------------------------------------------------------------- #
623
+ # Misc options
624
+ # ---------------------------------------------------------------------------- #
625
+ # Directory where output files are written
626
+ _C.OUTPUT_DIR = "./output"
627
+ # Set seed to negative to fully randomize everything.
628
+ # Set seed to positive to use a fixed seed. Note that a fixed seed increases
629
+ # reproducibility but does not guarantee fully deterministic behavior.
630
+ # Disabling all parallelism further increases reproducibility.
631
+ _C.SEED = -1
632
+ # Benchmark different cudnn algorithms.
633
+ # If input images have very different sizes, this option will have large overhead
634
+ # for about 10k iterations. It usually hurts total time, but can benefit for certain models.
635
+ # If input images have the same or similar sizes, benchmark is often helpful.
636
+ _C.CUDNN_BENCHMARK = False
637
+ # The period (in terms of steps) for minibatch visualization at train time.
638
+ # Set to 0 to disable.
639
+ _C.VIS_PERIOD = 0
640
+
641
+ # global config is for quick hack purposes.
642
+ # You can set them in command line or config files,
643
+ # and access it with:
644
+ #
645
+ # from annotator.oneformer.detectron2.config import global_cfg
646
+ # print(global_cfg.HACK)
647
+ #
648
+ # Do not commit any configs into it.
649
+ _C.GLOBAL = CN()
650
+ _C.GLOBAL.HACK = 1.0
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/instantiate.py ADDED
@@ -0,0 +1,88 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+
3
+ import collections.abc as abc
4
+ import dataclasses
5
+ import logging
6
+ from typing import Any
7
+
8
+ from annotator.oneformer.detectron2.utils.registry import _convert_target_to_string, locate
9
+
10
+ __all__ = ["dump_dataclass", "instantiate"]
11
+
12
+
13
+ def dump_dataclass(obj: Any):
14
+ """
15
+ Dump a dataclass recursively into a dict that can be later instantiated.
16
+
17
+ Args:
18
+ obj: a dataclass object
19
+
20
+ Returns:
21
+ dict
22
+ """
23
+ assert dataclasses.is_dataclass(obj) and not isinstance(
24
+ obj, type
25
+ ), "dump_dataclass() requires an instance of a dataclass."
26
+ ret = {"_target_": _convert_target_to_string(type(obj))}
27
+ for f in dataclasses.fields(obj):
28
+ v = getattr(obj, f.name)
29
+ if dataclasses.is_dataclass(v):
30
+ v = dump_dataclass(v)
31
+ if isinstance(v, (list, tuple)):
32
+ v = [dump_dataclass(x) if dataclasses.is_dataclass(x) else x for x in v]
33
+ ret[f.name] = v
34
+ return ret
35
+
36
+
37
+ def instantiate(cfg):
38
+ """
39
+ Recursively instantiate objects defined in dictionaries by
40
+ "_target_" and arguments.
41
+
42
+ Args:
43
+ cfg: a dict-like object with "_target_" that defines the caller, and
44
+ other keys that define the arguments
45
+
46
+ Returns:
47
+ object instantiated by cfg
48
+ """
49
+ from omegaconf import ListConfig, DictConfig, OmegaConf
50
+
51
+ if isinstance(cfg, ListConfig):
52
+ lst = [instantiate(x) for x in cfg]
53
+ return ListConfig(lst, flags={"allow_objects": True})
54
+ if isinstance(cfg, list):
55
+ # Specialize for list, because many classes take
56
+ # list[objects] as arguments, such as ResNet, DatasetMapper
57
+ return [instantiate(x) for x in cfg]
58
+
59
+ # If input is a DictConfig backed by dataclasses (i.e. omegaconf's structured config),
60
+ # instantiate it to the actual dataclass.
61
+ if isinstance(cfg, DictConfig) and dataclasses.is_dataclass(cfg._metadata.object_type):
62
+ return OmegaConf.to_object(cfg)
63
+
64
+ if isinstance(cfg, abc.Mapping) and "_target_" in cfg:
65
+ # conceptually equivalent to hydra.utils.instantiate(cfg) with _convert_=all,
66
+ # but faster: https://github.com/facebookresearch/hydra/issues/1200
67
+ cfg = {k: instantiate(v) for k, v in cfg.items()}
68
+ cls = cfg.pop("_target_")
69
+ cls = instantiate(cls)
70
+
71
+ if isinstance(cls, str):
72
+ cls_name = cls
73
+ cls = locate(cls_name)
74
+ assert cls is not None, cls_name
75
+ else:
76
+ try:
77
+ cls_name = cls.__module__ + "." + cls.__qualname__
78
+ except Exception:
79
+ # target could be anything, so the above could fail
80
+ cls_name = str(cls)
81
+ assert callable(cls), f"_target_ {cls} does not define a callable object"
82
+ try:
83
+ return cls(**cfg)
84
+ except TypeError:
85
+ logger = logging.getLogger(__name__)
86
+ logger.error(f"Error when instantiating {cls_name}!")
87
+ raise
88
+ return cfg # return as-is if don't know what to do
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/config/lazy.py ADDED
@@ -0,0 +1,435 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+
3
+ import ast
4
+ import builtins
5
+ import collections.abc as abc
6
+ import importlib
7
+ import inspect
8
+ import logging
9
+ import os
10
+ import uuid
11
+ from contextlib import contextmanager
12
+ from copy import deepcopy
13
+ from dataclasses import is_dataclass
14
+ from typing import List, Tuple, Union
15
+ import yaml
16
+ from omegaconf import DictConfig, ListConfig, OmegaConf, SCMode
17
+
18
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
19
+ from annotator.oneformer.detectron2.utils.registry import _convert_target_to_string
20
+
21
+ __all__ = ["LazyCall", "LazyConfig"]
22
+
23
+
24
+ class LazyCall:
25
+ """
26
+ Wrap a callable so that when it's called, the call will not be executed,
27
+ but returns a dict that describes the call.
28
+
29
+ LazyCall object has to be called with only keyword arguments. Positional
30
+ arguments are not yet supported.
31
+
32
+ Examples:
33
+ ::
34
+ from annotator.oneformer.detectron2.config import instantiate, LazyCall
35
+
36
+ layer_cfg = LazyCall(nn.Conv2d)(in_channels=32, out_channels=32)
37
+ layer_cfg.out_channels = 64 # can edit it afterwards
38
+ layer = instantiate(layer_cfg)
39
+ """
40
+
41
+ def __init__(self, target):
42
+ if not (callable(target) or isinstance(target, (str, abc.Mapping))):
43
+ raise TypeError(
44
+ f"target of LazyCall must be a callable or defines a callable! Got {target}"
45
+ )
46
+ self._target = target
47
+
48
+ def __call__(self, **kwargs):
49
+ if is_dataclass(self._target):
50
+ # omegaconf object cannot hold dataclass type
51
+ # https://github.com/omry/omegaconf/issues/784
52
+ target = _convert_target_to_string(self._target)
53
+ else:
54
+ target = self._target
55
+ kwargs["_target_"] = target
56
+
57
+ return DictConfig(content=kwargs, flags={"allow_objects": True})
58
+
59
+
60
+ def _visit_dict_config(cfg, func):
61
+ """
62
+ Apply func recursively to all DictConfig in cfg.
63
+ """
64
+ if isinstance(cfg, DictConfig):
65
+ func(cfg)
66
+ for v in cfg.values():
67
+ _visit_dict_config(v, func)
68
+ elif isinstance(cfg, ListConfig):
69
+ for v in cfg:
70
+ _visit_dict_config(v, func)
71
+
72
+
73
+ def _validate_py_syntax(filename):
74
+ # see also https://github.com/open-mmlab/mmcv/blob/master/mmcv/utils/config.py
75
+ with PathManager.open(filename, "r") as f:
76
+ content = f.read()
77
+ try:
78
+ ast.parse(content)
79
+ except SyntaxError as e:
80
+ raise SyntaxError(f"Config file {filename} has syntax error!") from e
81
+
82
+
83
+ def _cast_to_config(obj):
84
+ # if given a dict, return DictConfig instead
85
+ if isinstance(obj, dict):
86
+ return DictConfig(obj, flags={"allow_objects": True})
87
+ return obj
88
+
89
+
90
+ _CFG_PACKAGE_NAME = "detectron2._cfg_loader"
91
+ """
92
+ A namespace to put all imported config into.
93
+ """
94
+
95
+
96
+ def _random_package_name(filename):
97
+ # generate a random package name when loading config files
98
+ return _CFG_PACKAGE_NAME + str(uuid.uuid4())[:4] + "." + os.path.basename(filename)
99
+
100
+
101
+ @contextmanager
102
+ def _patch_import():
103
+ """
104
+ Enhance relative import statements in config files, so that they:
105
+ 1. locate files purely based on relative location, regardless of packages.
106
+ e.g. you can import file without having __init__
107
+ 2. do not cache modules globally; modifications of module states has no side effect
108
+ 3. support other storage system through PathManager, so config files can be in the cloud
109
+ 4. imported dict are turned into omegaconf.DictConfig automatically
110
+ """
111
+ old_import = builtins.__import__
112
+
113
+ def find_relative_file(original_file, relative_import_path, level):
114
+ # NOTE: "from . import x" is not handled. Because then it's unclear
115
+ # if such import should produce `x` as a python module or DictConfig.
116
+ # This can be discussed further if needed.
117
+ relative_import_err = """
118
+ Relative import of directories is not allowed within config files.
119
+ Within a config file, relative import can only import other config files.
120
+ """.replace(
121
+ "\n", " "
122
+ )
123
+ if not len(relative_import_path):
124
+ raise ImportError(relative_import_err)
125
+
126
+ cur_file = os.path.dirname(original_file)
127
+ for _ in range(level - 1):
128
+ cur_file = os.path.dirname(cur_file)
129
+ cur_name = relative_import_path.lstrip(".")
130
+ for part in cur_name.split("."):
131
+ cur_file = os.path.join(cur_file, part)
132
+ if not cur_file.endswith(".py"):
133
+ cur_file += ".py"
134
+ if not PathManager.isfile(cur_file):
135
+ cur_file_no_suffix = cur_file[: -len(".py")]
136
+ if PathManager.isdir(cur_file_no_suffix):
137
+ raise ImportError(f"Cannot import from {cur_file_no_suffix}." + relative_import_err)
138
+ else:
139
+ raise ImportError(
140
+ f"Cannot import name {relative_import_path} from "
141
+ f"{original_file}: {cur_file} does not exist."
142
+ )
143
+ return cur_file
144
+
145
+ def new_import(name, globals=None, locals=None, fromlist=(), level=0):
146
+ if (
147
+ # Only deal with relative imports inside config files
148
+ level != 0
149
+ and globals is not None
150
+ and (globals.get("__package__", "") or "").startswith(_CFG_PACKAGE_NAME)
151
+ ):
152
+ cur_file = find_relative_file(globals["__file__"], name, level)
153
+ _validate_py_syntax(cur_file)
154
+ spec = importlib.machinery.ModuleSpec(
155
+ _random_package_name(cur_file), None, origin=cur_file
156
+ )
157
+ module = importlib.util.module_from_spec(spec)
158
+ module.__file__ = cur_file
159
+ with PathManager.open(cur_file) as f:
160
+ content = f.read()
161
+ exec(compile(content, cur_file, "exec"), module.__dict__)
162
+ for name in fromlist: # turn imported dict into DictConfig automatically
163
+ val = _cast_to_config(module.__dict__[name])
164
+ module.__dict__[name] = val
165
+ return module
166
+ return old_import(name, globals, locals, fromlist=fromlist, level=level)
167
+
168
+ builtins.__import__ = new_import
169
+ yield new_import
170
+ builtins.__import__ = old_import
171
+
172
+
173
+ class LazyConfig:
174
+ """
175
+ Provide methods to save, load, and overrides an omegaconf config object
176
+ which may contain definition of lazily-constructed objects.
177
+ """
178
+
179
+ @staticmethod
180
+ def load_rel(filename: str, keys: Union[None, str, Tuple[str, ...]] = None):
181
+ """
182
+ Similar to :meth:`load()`, but load path relative to the caller's
183
+ source file.
184
+
185
+ This has the same functionality as a relative import, except that this method
186
+ accepts filename as a string, so more characters are allowed in the filename.
187
+ """
188
+ caller_frame = inspect.stack()[1]
189
+ caller_fname = caller_frame[0].f_code.co_filename
190
+ assert caller_fname != "<string>", "load_rel Unable to find caller"
191
+ caller_dir = os.path.dirname(caller_fname)
192
+ filename = os.path.join(caller_dir, filename)
193
+ return LazyConfig.load(filename, keys)
194
+
195
+ @staticmethod
196
+ def load(filename: str, keys: Union[None, str, Tuple[str, ...]] = None):
197
+ """
198
+ Load a config file.
199
+
200
+ Args:
201
+ filename: absolute path or relative path w.r.t. the current working directory
202
+ keys: keys to load and return. If not given, return all keys
203
+ (whose values are config objects) in a dict.
204
+ """
205
+ has_keys = keys is not None
206
+ filename = filename.replace("/./", "/") # redundant
207
+ if os.path.splitext(filename)[1] not in [".py", ".yaml", ".yml"]:
208
+ raise ValueError(f"Config file {filename} has to be a python or yaml file.")
209
+ if filename.endswith(".py"):
210
+ _validate_py_syntax(filename)
211
+
212
+ with _patch_import():
213
+ # Record the filename
214
+ module_namespace = {
215
+ "__file__": filename,
216
+ "__package__": _random_package_name(filename),
217
+ }
218
+ with PathManager.open(filename) as f:
219
+ content = f.read()
220
+ # Compile first with filename to:
221
+ # 1. make filename appears in stacktrace
222
+ # 2. make load_rel able to find its parent's (possibly remote) location
223
+ exec(compile(content, filename, "exec"), module_namespace)
224
+
225
+ ret = module_namespace
226
+ else:
227
+ with PathManager.open(filename) as f:
228
+ obj = yaml.unsafe_load(f)
229
+ ret = OmegaConf.create(obj, flags={"allow_objects": True})
230
+
231
+ if has_keys:
232
+ if isinstance(keys, str):
233
+ return _cast_to_config(ret[keys])
234
+ else:
235
+ return tuple(_cast_to_config(ret[a]) for a in keys)
236
+ else:
237
+ if filename.endswith(".py"):
238
+ # when not specified, only load those that are config objects
239
+ ret = DictConfig(
240
+ {
241
+ name: _cast_to_config(value)
242
+ for name, value in ret.items()
243
+ if isinstance(value, (DictConfig, ListConfig, dict))
244
+ and not name.startswith("_")
245
+ },
246
+ flags={"allow_objects": True},
247
+ )
248
+ return ret
249
+
250
+ @staticmethod
251
+ def save(cfg, filename: str):
252
+ """
253
+ Save a config object to a yaml file.
254
+ Note that when the config dictionary contains complex objects (e.g. lambda),
255
+ it can't be saved to yaml. In that case we will print an error and
256
+ attempt to save to a pkl file instead.
257
+
258
+ Args:
259
+ cfg: an omegaconf config object
260
+ filename: yaml file name to save the config file
261
+ """
262
+ logger = logging.getLogger(__name__)
263
+ try:
264
+ cfg = deepcopy(cfg)
265
+ except Exception:
266
+ pass
267
+ else:
268
+ # if it's deep-copyable, then...
269
+ def _replace_type_by_name(x):
270
+ if "_target_" in x and callable(x._target_):
271
+ try:
272
+ x._target_ = _convert_target_to_string(x._target_)
273
+ except AttributeError:
274
+ pass
275
+
276
+ # not necessary, but makes yaml looks nicer
277
+ _visit_dict_config(cfg, _replace_type_by_name)
278
+
279
+ save_pkl = False
280
+ try:
281
+ dict = OmegaConf.to_container(
282
+ cfg,
283
+ # Do not resolve interpolation when saving, i.e. do not turn ${a} into
284
+ # actual values when saving.
285
+ resolve=False,
286
+ # Save structures (dataclasses) in a format that can be instantiated later.
287
+ # Without this option, the type information of the dataclass will be erased.
288
+ structured_config_mode=SCMode.INSTANTIATE,
289
+ )
290
+ dumped = yaml.dump(dict, default_flow_style=None, allow_unicode=True, width=9999)
291
+ with PathManager.open(filename, "w") as f:
292
+ f.write(dumped)
293
+
294
+ try:
295
+ _ = yaml.unsafe_load(dumped) # test that it is loadable
296
+ except Exception:
297
+ logger.warning(
298
+ "The config contains objects that cannot serialize to a valid yaml. "
299
+ f"{filename} is human-readable but cannot be loaded."
300
+ )
301
+ save_pkl = True
302
+ except Exception:
303
+ logger.exception("Unable to serialize the config to yaml. Error:")
304
+ save_pkl = True
305
+
306
+ if save_pkl:
307
+ new_filename = filename + ".pkl"
308
+ # try:
309
+ # # retry by pickle
310
+ # with PathManager.open(new_filename, "wb") as f:
311
+ # cloudpickle.dump(cfg, f)
312
+ # logger.warning(f"Config is saved using cloudpickle at {new_filename}.")
313
+ # except Exception:
314
+ # pass
315
+
316
+ @staticmethod
317
+ def apply_overrides(cfg, overrides: List[str]):
318
+ """
319
+ In-place override contents of cfg.
320
+
321
+ Args:
322
+ cfg: an omegaconf config object
323
+ overrides: list of strings in the format of "a=b" to override configs.
324
+ See https://hydra.cc/docs/next/advanced/override_grammar/basic/
325
+ for syntax.
326
+
327
+ Returns:
328
+ the cfg object
329
+ """
330
+
331
+ def safe_update(cfg, key, value):
332
+ parts = key.split(".")
333
+ for idx in range(1, len(parts)):
334
+ prefix = ".".join(parts[:idx])
335
+ v = OmegaConf.select(cfg, prefix, default=None)
336
+ if v is None:
337
+ break
338
+ if not OmegaConf.is_config(v):
339
+ raise KeyError(
340
+ f"Trying to update key {key}, but {prefix} "
341
+ f"is not a config, but has type {type(v)}."
342
+ )
343
+ OmegaConf.update(cfg, key, value, merge=True)
344
+
345
+ try:
346
+ from hydra.core.override_parser.overrides_parser import OverridesParser
347
+
348
+ has_hydra = True
349
+ except ImportError:
350
+ has_hydra = False
351
+
352
+ if has_hydra:
353
+ parser = OverridesParser.create()
354
+ overrides = parser.parse_overrides(overrides)
355
+ for o in overrides:
356
+ key = o.key_or_group
357
+ value = o.value()
358
+ if o.is_delete():
359
+ # TODO support this
360
+ raise NotImplementedError("deletion is not yet a supported override")
361
+ safe_update(cfg, key, value)
362
+ else:
363
+ # Fallback. Does not support all the features and error checking like hydra.
364
+ for o in overrides:
365
+ key, value = o.split("=")
366
+ try:
367
+ value = eval(value, {})
368
+ except NameError:
369
+ pass
370
+ safe_update(cfg, key, value)
371
+ return cfg
372
+
373
+ # @staticmethod
374
+ # def to_py(cfg, prefix: str = "cfg."):
375
+ # """
376
+ # Try to convert a config object into Python-like psuedo code.
377
+ #
378
+ # Note that perfect conversion is not always possible. So the returned
379
+ # results are mainly meant to be human-readable, and not meant to be executed.
380
+ #
381
+ # Args:
382
+ # cfg: an omegaconf config object
383
+ # prefix: root name for the resulting code (default: "cfg.")
384
+ #
385
+ #
386
+ # Returns:
387
+ # str of formatted Python code
388
+ # """
389
+ # import black
390
+ #
391
+ # cfg = OmegaConf.to_container(cfg, resolve=True)
392
+ #
393
+ # def _to_str(obj, prefix=None, inside_call=False):
394
+ # if prefix is None:
395
+ # prefix = []
396
+ # if isinstance(obj, abc.Mapping) and "_target_" in obj:
397
+ # # Dict representing a function call
398
+ # target = _convert_target_to_string(obj.pop("_target_"))
399
+ # args = []
400
+ # for k, v in sorted(obj.items()):
401
+ # args.append(f"{k}={_to_str(v, inside_call=True)}")
402
+ # args = ", ".join(args)
403
+ # call = f"{target}({args})"
404
+ # return "".join(prefix) + call
405
+ # elif isinstance(obj, abc.Mapping) and not inside_call:
406
+ # # Dict that is not inside a call is a list of top-level config objects that we
407
+ # # render as one object per line with dot separated prefixes
408
+ # key_list = []
409
+ # for k, v in sorted(obj.items()):
410
+ # if isinstance(v, abc.Mapping) and "_target_" not in v:
411
+ # key_list.append(_to_str(v, prefix=prefix + [k + "."]))
412
+ # else:
413
+ # key = "".join(prefix) + k
414
+ # key_list.append(f"{key}={_to_str(v)}")
415
+ # return "\n".join(key_list)
416
+ # elif isinstance(obj, abc.Mapping):
417
+ # # Dict that is inside a call is rendered as a regular dict
418
+ # return (
419
+ # "{"
420
+ # + ",".join(
421
+ # f"{repr(k)}: {_to_str(v, inside_call=inside_call)}"
422
+ # for k, v in sorted(obj.items())
423
+ # )
424
+ # + "}"
425
+ # )
426
+ # elif isinstance(obj, list):
427
+ # return "[" + ",".join(_to_str(x, inside_call=inside_call) for x in obj) + "]"
428
+ # else:
429
+ # return repr(obj)
430
+ #
431
+ # py_str = _to_str(cfg, prefix=[prefix])
432
+ # try:
433
+ # return black.format_str(py_str, mode=black.Mode())
434
+ # except black.InvalidInput:
435
+ # return py_str
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/__init__.py ADDED
@@ -0,0 +1,19 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from . import transforms # isort:skip
3
+
4
+ from .build import (
5
+ build_batch_data_loader,
6
+ build_detection_test_loader,
7
+ build_detection_train_loader,
8
+ get_detection_dataset_dicts,
9
+ load_proposals_into_dataset,
10
+ print_instances_class_histogram,
11
+ )
12
+ from .catalog import DatasetCatalog, MetadataCatalog, Metadata
13
+ from .common import DatasetFromList, MapDataset, ToIterableDataset
14
+ from .dataset_mapper import DatasetMapper
15
+
16
+ # ensure the builtin datasets are registered
17
+ from . import datasets, samplers # isort:skip
18
+
19
+ __all__ = [k for k in globals().keys() if not k.startswith("_")]
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/benchmark.py ADDED
@@ -0,0 +1,225 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import logging
3
+ import numpy as np
4
+ from itertools import count
5
+ from typing import List, Tuple
6
+ import torch
7
+ import tqdm
8
+ from fvcore.common.timer import Timer
9
+
10
+ from annotator.oneformer.detectron2.utils import comm
11
+
12
+ from .build import build_batch_data_loader
13
+ from .common import DatasetFromList, MapDataset
14
+ from .samplers import TrainingSampler
15
+
16
+ logger = logging.getLogger(__name__)
17
+
18
+
19
+ class _EmptyMapDataset(torch.utils.data.Dataset):
20
+ """
21
+ Map anything to emptiness.
22
+ """
23
+
24
+ def __init__(self, dataset):
25
+ self.ds = dataset
26
+
27
+ def __len__(self):
28
+ return len(self.ds)
29
+
30
+ def __getitem__(self, idx):
31
+ _ = self.ds[idx]
32
+ return [0]
33
+
34
+
35
+ def iter_benchmark(
36
+ iterator, num_iter: int, warmup: int = 5, max_time_seconds: float = 60
37
+ ) -> Tuple[float, List[float]]:
38
+ """
39
+ Benchmark an iterator/iterable for `num_iter` iterations with an extra
40
+ `warmup` iterations of warmup.
41
+ End early if `max_time_seconds` time is spent on iterations.
42
+
43
+ Returns:
44
+ float: average time (seconds) per iteration
45
+ list[float]: time spent on each iteration. Sometimes useful for further analysis.
46
+ """
47
+ num_iter, warmup = int(num_iter), int(warmup)
48
+
49
+ iterator = iter(iterator)
50
+ for _ in range(warmup):
51
+ next(iterator)
52
+ timer = Timer()
53
+ all_times = []
54
+ for curr_iter in tqdm.trange(num_iter):
55
+ start = timer.seconds()
56
+ if start > max_time_seconds:
57
+ num_iter = curr_iter
58
+ break
59
+ next(iterator)
60
+ all_times.append(timer.seconds() - start)
61
+ avg = timer.seconds() / num_iter
62
+ return avg, all_times
63
+
64
+
65
+ class DataLoaderBenchmark:
66
+ """
67
+ Some common benchmarks that help understand perf bottleneck of a standard dataloader
68
+ made of dataset, mapper and sampler.
69
+ """
70
+
71
+ def __init__(
72
+ self,
73
+ dataset,
74
+ *,
75
+ mapper,
76
+ sampler=None,
77
+ total_batch_size,
78
+ num_workers=0,
79
+ max_time_seconds: int = 90,
80
+ ):
81
+ """
82
+ Args:
83
+ max_time_seconds (int): maximum time to spent for each benchmark
84
+ other args: same as in `build.py:build_detection_train_loader`
85
+ """
86
+ if isinstance(dataset, list):
87
+ dataset = DatasetFromList(dataset, copy=False, serialize=True)
88
+ if sampler is None:
89
+ sampler = TrainingSampler(len(dataset))
90
+
91
+ self.dataset = dataset
92
+ self.mapper = mapper
93
+ self.sampler = sampler
94
+ self.total_batch_size = total_batch_size
95
+ self.num_workers = num_workers
96
+ self.per_gpu_batch_size = self.total_batch_size // comm.get_world_size()
97
+
98
+ self.max_time_seconds = max_time_seconds
99
+
100
+ def _benchmark(self, iterator, num_iter, warmup, msg=None):
101
+ avg, all_times = iter_benchmark(iterator, num_iter, warmup, self.max_time_seconds)
102
+ if msg is not None:
103
+ self._log_time(msg, avg, all_times)
104
+ return avg, all_times
105
+
106
+ def _log_time(self, msg, avg, all_times, distributed=False):
107
+ percentiles = [np.percentile(all_times, k, interpolation="nearest") for k in [1, 5, 95, 99]]
108
+ if not distributed:
109
+ logger.info(
110
+ f"{msg}: avg={1.0/avg:.1f} it/s, "
111
+ f"p1={percentiles[0]:.2g}s, p5={percentiles[1]:.2g}s, "
112
+ f"p95={percentiles[2]:.2g}s, p99={percentiles[3]:.2g}s."
113
+ )
114
+ return
115
+ avg_per_gpu = comm.all_gather(avg)
116
+ percentiles_per_gpu = comm.all_gather(percentiles)
117
+ if comm.get_rank() > 0:
118
+ return
119
+ for idx, avg, percentiles in zip(count(), avg_per_gpu, percentiles_per_gpu):
120
+ logger.info(
121
+ f"GPU{idx} {msg}: avg={1.0/avg:.1f} it/s, "
122
+ f"p1={percentiles[0]:.2g}s, p5={percentiles[1]:.2g}s, "
123
+ f"p95={percentiles[2]:.2g}s, p99={percentiles[3]:.2g}s."
124
+ )
125
+
126
+ def benchmark_dataset(self, num_iter, warmup=5):
127
+ """
128
+ Benchmark the speed of taking raw samples from the dataset.
129
+ """
130
+
131
+ def loader():
132
+ while True:
133
+ for k in self.sampler:
134
+ yield self.dataset[k]
135
+
136
+ self._benchmark(loader(), num_iter, warmup, "Dataset Alone")
137
+
138
+ def benchmark_mapper(self, num_iter, warmup=5):
139
+ """
140
+ Benchmark the speed of taking raw samples from the dataset and map
141
+ them in a single process.
142
+ """
143
+
144
+ def loader():
145
+ while True:
146
+ for k in self.sampler:
147
+ yield self.mapper(self.dataset[k])
148
+
149
+ self._benchmark(loader(), num_iter, warmup, "Single Process Mapper (sec/sample)")
150
+
151
+ def benchmark_workers(self, num_iter, warmup=10):
152
+ """
153
+ Benchmark the dataloader by tuning num_workers to [0, 1, self.num_workers].
154
+ """
155
+ candidates = [0, 1]
156
+ if self.num_workers not in candidates:
157
+ candidates.append(self.num_workers)
158
+
159
+ dataset = MapDataset(self.dataset, self.mapper)
160
+ for n in candidates:
161
+ loader = build_batch_data_loader(
162
+ dataset,
163
+ self.sampler,
164
+ self.total_batch_size,
165
+ num_workers=n,
166
+ )
167
+ self._benchmark(
168
+ iter(loader),
169
+ num_iter * max(n, 1),
170
+ warmup * max(n, 1),
171
+ f"DataLoader ({n} workers, bs={self.per_gpu_batch_size})",
172
+ )
173
+ del loader
174
+
175
+ def benchmark_IPC(self, num_iter, warmup=10):
176
+ """
177
+ Benchmark the dataloader where each worker outputs nothing. This
178
+ eliminates the IPC overhead compared to the regular dataloader.
179
+
180
+ PyTorch multiprocessing's IPC only optimizes for torch tensors.
181
+ Large numpy arrays or other data structure may incur large IPC overhead.
182
+ """
183
+ n = self.num_workers
184
+ dataset = _EmptyMapDataset(MapDataset(self.dataset, self.mapper))
185
+ loader = build_batch_data_loader(
186
+ dataset, self.sampler, self.total_batch_size, num_workers=n
187
+ )
188
+ self._benchmark(
189
+ iter(loader),
190
+ num_iter * max(n, 1),
191
+ warmup * max(n, 1),
192
+ f"DataLoader ({n} workers, bs={self.per_gpu_batch_size}) w/o comm",
193
+ )
194
+
195
+ def benchmark_distributed(self, num_iter, warmup=10):
196
+ """
197
+ Benchmark the dataloader in each distributed worker, and log results of
198
+ all workers. This helps understand the final performance as well as
199
+ the variances among workers.
200
+
201
+ It also prints startup time (first iter) of the dataloader.
202
+ """
203
+ gpu = comm.get_world_size()
204
+ dataset = MapDataset(self.dataset, self.mapper)
205
+ n = self.num_workers
206
+ loader = build_batch_data_loader(
207
+ dataset, self.sampler, self.total_batch_size, num_workers=n
208
+ )
209
+
210
+ timer = Timer()
211
+ loader = iter(loader)
212
+ next(loader)
213
+ startup_time = timer.seconds()
214
+ logger.info("Dataloader startup time: {:.2f} seconds".format(startup_time))
215
+
216
+ comm.synchronize()
217
+
218
+ avg, all_times = self._benchmark(loader, num_iter * max(n, 1), warmup * max(n, 1))
219
+ del loader
220
+ self._log_time(
221
+ f"DataLoader ({gpu} GPUs x {n} workers, total bs={self.total_batch_size})",
222
+ avg,
223
+ all_times,
224
+ True,
225
+ )
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/build.py ADDED
@@ -0,0 +1,556 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import itertools
3
+ import logging
4
+ import numpy as np
5
+ import operator
6
+ import pickle
7
+ from typing import Any, Callable, Dict, List, Optional, Union
8
+ import torch
9
+ import torch.utils.data as torchdata
10
+ from tabulate import tabulate
11
+ from termcolor import colored
12
+
13
+ from annotator.oneformer.detectron2.config import configurable
14
+ from annotator.oneformer.detectron2.structures import BoxMode
15
+ from annotator.oneformer.detectron2.utils.comm import get_world_size
16
+ from annotator.oneformer.detectron2.utils.env import seed_all_rng
17
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
18
+ from annotator.oneformer.detectron2.utils.logger import _log_api_usage, log_first_n
19
+
20
+ from .catalog import DatasetCatalog, MetadataCatalog
21
+ from .common import AspectRatioGroupedDataset, DatasetFromList, MapDataset, ToIterableDataset
22
+ from .dataset_mapper import DatasetMapper
23
+ from .detection_utils import check_metadata_consistency
24
+ from .samplers import (
25
+ InferenceSampler,
26
+ RandomSubsetTrainingSampler,
27
+ RepeatFactorTrainingSampler,
28
+ TrainingSampler,
29
+ )
30
+
31
+ """
32
+ This file contains the default logic to build a dataloader for training or testing.
33
+ """
34
+
35
+ __all__ = [
36
+ "build_batch_data_loader",
37
+ "build_detection_train_loader",
38
+ "build_detection_test_loader",
39
+ "get_detection_dataset_dicts",
40
+ "load_proposals_into_dataset",
41
+ "print_instances_class_histogram",
42
+ ]
43
+
44
+
45
+ def filter_images_with_only_crowd_annotations(dataset_dicts):
46
+ """
47
+ Filter out images with none annotations or only crowd annotations
48
+ (i.e., images without non-crowd annotations).
49
+ A common training-time preprocessing on COCO dataset.
50
+
51
+ Args:
52
+ dataset_dicts (list[dict]): annotations in Detectron2 Dataset format.
53
+
54
+ Returns:
55
+ list[dict]: the same format, but filtered.
56
+ """
57
+ num_before = len(dataset_dicts)
58
+
59
+ def valid(anns):
60
+ for ann in anns:
61
+ if ann.get("iscrowd", 0) == 0:
62
+ return True
63
+ return False
64
+
65
+ dataset_dicts = [x for x in dataset_dicts if valid(x["annotations"])]
66
+ num_after = len(dataset_dicts)
67
+ logger = logging.getLogger(__name__)
68
+ logger.info(
69
+ "Removed {} images with no usable annotations. {} images left.".format(
70
+ num_before - num_after, num_after
71
+ )
72
+ )
73
+ return dataset_dicts
74
+
75
+
76
+ def filter_images_with_few_keypoints(dataset_dicts, min_keypoints_per_image):
77
+ """
78
+ Filter out images with too few number of keypoints.
79
+
80
+ Args:
81
+ dataset_dicts (list[dict]): annotations in Detectron2 Dataset format.
82
+
83
+ Returns:
84
+ list[dict]: the same format as dataset_dicts, but filtered.
85
+ """
86
+ num_before = len(dataset_dicts)
87
+
88
+ def visible_keypoints_in_image(dic):
89
+ # Each keypoints field has the format [x1, y1, v1, ...], where v is visibility
90
+ annotations = dic["annotations"]
91
+ return sum(
92
+ (np.array(ann["keypoints"][2::3]) > 0).sum()
93
+ for ann in annotations
94
+ if "keypoints" in ann
95
+ )
96
+
97
+ dataset_dicts = [
98
+ x for x in dataset_dicts if visible_keypoints_in_image(x) >= min_keypoints_per_image
99
+ ]
100
+ num_after = len(dataset_dicts)
101
+ logger = logging.getLogger(__name__)
102
+ logger.info(
103
+ "Removed {} images with fewer than {} keypoints.".format(
104
+ num_before - num_after, min_keypoints_per_image
105
+ )
106
+ )
107
+ return dataset_dicts
108
+
109
+
110
+ def load_proposals_into_dataset(dataset_dicts, proposal_file):
111
+ """
112
+ Load precomputed object proposals into the dataset.
113
+
114
+ The proposal file should be a pickled dict with the following keys:
115
+
116
+ - "ids": list[int] or list[str], the image ids
117
+ - "boxes": list[np.ndarray], each is an Nx4 array of boxes corresponding to the image id
118
+ - "objectness_logits": list[np.ndarray], each is an N sized array of objectness scores
119
+ corresponding to the boxes.
120
+ - "bbox_mode": the BoxMode of the boxes array. Defaults to ``BoxMode.XYXY_ABS``.
121
+
122
+ Args:
123
+ dataset_dicts (list[dict]): annotations in Detectron2 Dataset format.
124
+ proposal_file (str): file path of pre-computed proposals, in pkl format.
125
+
126
+ Returns:
127
+ list[dict]: the same format as dataset_dicts, but added proposal field.
128
+ """
129
+ logger = logging.getLogger(__name__)
130
+ logger.info("Loading proposals from: {}".format(proposal_file))
131
+
132
+ with PathManager.open(proposal_file, "rb") as f:
133
+ proposals = pickle.load(f, encoding="latin1")
134
+
135
+ # Rename the key names in D1 proposal files
136
+ rename_keys = {"indexes": "ids", "scores": "objectness_logits"}
137
+ for key in rename_keys:
138
+ if key in proposals:
139
+ proposals[rename_keys[key]] = proposals.pop(key)
140
+
141
+ # Fetch the indexes of all proposals that are in the dataset
142
+ # Convert image_id to str since they could be int.
143
+ img_ids = set({str(record["image_id"]) for record in dataset_dicts})
144
+ id_to_index = {str(id): i for i, id in enumerate(proposals["ids"]) if str(id) in img_ids}
145
+
146
+ # Assuming default bbox_mode of precomputed proposals are 'XYXY_ABS'
147
+ bbox_mode = BoxMode(proposals["bbox_mode"]) if "bbox_mode" in proposals else BoxMode.XYXY_ABS
148
+
149
+ for record in dataset_dicts:
150
+ # Get the index of the proposal
151
+ i = id_to_index[str(record["image_id"])]
152
+
153
+ boxes = proposals["boxes"][i]
154
+ objectness_logits = proposals["objectness_logits"][i]
155
+ # Sort the proposals in descending order of the scores
156
+ inds = objectness_logits.argsort()[::-1]
157
+ record["proposal_boxes"] = boxes[inds]
158
+ record["proposal_objectness_logits"] = objectness_logits[inds]
159
+ record["proposal_bbox_mode"] = bbox_mode
160
+
161
+ return dataset_dicts
162
+
163
+
164
+ def print_instances_class_histogram(dataset_dicts, class_names):
165
+ """
166
+ Args:
167
+ dataset_dicts (list[dict]): list of dataset dicts.
168
+ class_names (list[str]): list of class names (zero-indexed).
169
+ """
170
+ num_classes = len(class_names)
171
+ hist_bins = np.arange(num_classes + 1)
172
+ histogram = np.zeros((num_classes,), dtype=np.int)
173
+ for entry in dataset_dicts:
174
+ annos = entry["annotations"]
175
+ classes = np.asarray(
176
+ [x["category_id"] for x in annos if not x.get("iscrowd", 0)], dtype=np.int
177
+ )
178
+ if len(classes):
179
+ assert classes.min() >= 0, f"Got an invalid category_id={classes.min()}"
180
+ assert (
181
+ classes.max() < num_classes
182
+ ), f"Got an invalid category_id={classes.max()} for a dataset of {num_classes} classes"
183
+ histogram += np.histogram(classes, bins=hist_bins)[0]
184
+
185
+ N_COLS = min(6, len(class_names) * 2)
186
+
187
+ def short_name(x):
188
+ # make long class names shorter. useful for lvis
189
+ if len(x) > 13:
190
+ return x[:11] + ".."
191
+ return x
192
+
193
+ data = list(
194
+ itertools.chain(*[[short_name(class_names[i]), int(v)] for i, v in enumerate(histogram)])
195
+ )
196
+ total_num_instances = sum(data[1::2])
197
+ data.extend([None] * (N_COLS - (len(data) % N_COLS)))
198
+ if num_classes > 1:
199
+ data.extend(["total", total_num_instances])
200
+ data = itertools.zip_longest(*[data[i::N_COLS] for i in range(N_COLS)])
201
+ table = tabulate(
202
+ data,
203
+ headers=["category", "#instances"] * (N_COLS // 2),
204
+ tablefmt="pipe",
205
+ numalign="left",
206
+ stralign="center",
207
+ )
208
+ log_first_n(
209
+ logging.INFO,
210
+ "Distribution of instances among all {} categories:\n".format(num_classes)
211
+ + colored(table, "cyan"),
212
+ key="message",
213
+ )
214
+
215
+
216
+ def get_detection_dataset_dicts(
217
+ names,
218
+ filter_empty=True,
219
+ min_keypoints=0,
220
+ proposal_files=None,
221
+ check_consistency=True,
222
+ ):
223
+ """
224
+ Load and prepare dataset dicts for instance detection/segmentation and semantic segmentation.
225
+
226
+ Args:
227
+ names (str or list[str]): a dataset name or a list of dataset names
228
+ filter_empty (bool): whether to filter out images without instance annotations
229
+ min_keypoints (int): filter out images with fewer keypoints than
230
+ `min_keypoints`. Set to 0 to do nothing.
231
+ proposal_files (list[str]): if given, a list of object proposal files
232
+ that match each dataset in `names`.
233
+ check_consistency (bool): whether to check if datasets have consistent metadata.
234
+
235
+ Returns:
236
+ list[dict]: a list of dicts following the standard dataset dict format.
237
+ """
238
+ if isinstance(names, str):
239
+ names = [names]
240
+ assert len(names), names
241
+ dataset_dicts = [DatasetCatalog.get(dataset_name) for dataset_name in names]
242
+
243
+ if isinstance(dataset_dicts[0], torchdata.Dataset):
244
+ if len(dataset_dicts) > 1:
245
+ # ConcatDataset does not work for iterable style dataset.
246
+ # We could support concat for iterable as well, but it's often
247
+ # not a good idea to concat iterables anyway.
248
+ return torchdata.ConcatDataset(dataset_dicts)
249
+ return dataset_dicts[0]
250
+
251
+ for dataset_name, dicts in zip(names, dataset_dicts):
252
+ assert len(dicts), "Dataset '{}' is empty!".format(dataset_name)
253
+
254
+ if proposal_files is not None:
255
+ assert len(names) == len(proposal_files)
256
+ # load precomputed proposals from proposal files
257
+ dataset_dicts = [
258
+ load_proposals_into_dataset(dataset_i_dicts, proposal_file)
259
+ for dataset_i_dicts, proposal_file in zip(dataset_dicts, proposal_files)
260
+ ]
261
+
262
+ dataset_dicts = list(itertools.chain.from_iterable(dataset_dicts))
263
+
264
+ has_instances = "annotations" in dataset_dicts[0]
265
+ if filter_empty and has_instances:
266
+ dataset_dicts = filter_images_with_only_crowd_annotations(dataset_dicts)
267
+ if min_keypoints > 0 and has_instances:
268
+ dataset_dicts = filter_images_with_few_keypoints(dataset_dicts, min_keypoints)
269
+
270
+ if check_consistency and has_instances:
271
+ try:
272
+ class_names = MetadataCatalog.get(names[0]).thing_classes
273
+ check_metadata_consistency("thing_classes", names)
274
+ print_instances_class_histogram(dataset_dicts, class_names)
275
+ except AttributeError: # class names are not available for this dataset
276
+ pass
277
+
278
+ assert len(dataset_dicts), "No valid data found in {}.".format(",".join(names))
279
+ return dataset_dicts
280
+
281
+
282
+ def build_batch_data_loader(
283
+ dataset,
284
+ sampler,
285
+ total_batch_size,
286
+ *,
287
+ aspect_ratio_grouping=False,
288
+ num_workers=0,
289
+ collate_fn=None,
290
+ ):
291
+ """
292
+ Build a batched dataloader. The main differences from `torch.utils.data.DataLoader` are:
293
+ 1. support aspect ratio grouping options
294
+ 2. use no "batch collation", because this is common for detection training
295
+
296
+ Args:
297
+ dataset (torch.utils.data.Dataset): a pytorch map-style or iterable dataset.
298
+ sampler (torch.utils.data.sampler.Sampler or None): a sampler that produces indices.
299
+ Must be provided iff. ``dataset`` is a map-style dataset.
300
+ total_batch_size, aspect_ratio_grouping, num_workers, collate_fn: see
301
+ :func:`build_detection_train_loader`.
302
+
303
+ Returns:
304
+ iterable[list]. Length of each list is the batch size of the current
305
+ GPU. Each element in the list comes from the dataset.
306
+ """
307
+ world_size = get_world_size()
308
+ assert (
309
+ total_batch_size > 0 and total_batch_size % world_size == 0
310
+ ), "Total batch size ({}) must be divisible by the number of gpus ({}).".format(
311
+ total_batch_size, world_size
312
+ )
313
+ batch_size = total_batch_size // world_size
314
+
315
+ if isinstance(dataset, torchdata.IterableDataset):
316
+ assert sampler is None, "sampler must be None if dataset is IterableDataset"
317
+ else:
318
+ dataset = ToIterableDataset(dataset, sampler)
319
+
320
+ if aspect_ratio_grouping:
321
+ data_loader = torchdata.DataLoader(
322
+ dataset,
323
+ num_workers=num_workers,
324
+ collate_fn=operator.itemgetter(0), # don't batch, but yield individual elements
325
+ worker_init_fn=worker_init_reset_seed,
326
+ ) # yield individual mapped dict
327
+ data_loader = AspectRatioGroupedDataset(data_loader, batch_size)
328
+ if collate_fn is None:
329
+ return data_loader
330
+ return MapDataset(data_loader, collate_fn)
331
+ else:
332
+ return torchdata.DataLoader(
333
+ dataset,
334
+ batch_size=batch_size,
335
+ drop_last=True,
336
+ num_workers=num_workers,
337
+ collate_fn=trivial_batch_collator if collate_fn is None else collate_fn,
338
+ worker_init_fn=worker_init_reset_seed,
339
+ )
340
+
341
+
342
+ def _train_loader_from_config(cfg, mapper=None, *, dataset=None, sampler=None):
343
+ if dataset is None:
344
+ dataset = get_detection_dataset_dicts(
345
+ cfg.DATASETS.TRAIN,
346
+ filter_empty=cfg.DATALOADER.FILTER_EMPTY_ANNOTATIONS,
347
+ min_keypoints=cfg.MODEL.ROI_KEYPOINT_HEAD.MIN_KEYPOINTS_PER_IMAGE
348
+ if cfg.MODEL.KEYPOINT_ON
349
+ else 0,
350
+ proposal_files=cfg.DATASETS.PROPOSAL_FILES_TRAIN if cfg.MODEL.LOAD_PROPOSALS else None,
351
+ )
352
+ _log_api_usage("dataset." + cfg.DATASETS.TRAIN[0])
353
+
354
+ if mapper is None:
355
+ mapper = DatasetMapper(cfg, True)
356
+
357
+ if sampler is None:
358
+ sampler_name = cfg.DATALOADER.SAMPLER_TRAIN
359
+ logger = logging.getLogger(__name__)
360
+ if isinstance(dataset, torchdata.IterableDataset):
361
+ logger.info("Not using any sampler since the dataset is IterableDataset.")
362
+ sampler = None
363
+ else:
364
+ logger.info("Using training sampler {}".format(sampler_name))
365
+ if sampler_name == "TrainingSampler":
366
+ sampler = TrainingSampler(len(dataset))
367
+ elif sampler_name == "RepeatFactorTrainingSampler":
368
+ repeat_factors = RepeatFactorTrainingSampler.repeat_factors_from_category_frequency(
369
+ dataset, cfg.DATALOADER.REPEAT_THRESHOLD
370
+ )
371
+ sampler = RepeatFactorTrainingSampler(repeat_factors)
372
+ elif sampler_name == "RandomSubsetTrainingSampler":
373
+ sampler = RandomSubsetTrainingSampler(
374
+ len(dataset), cfg.DATALOADER.RANDOM_SUBSET_RATIO
375
+ )
376
+ else:
377
+ raise ValueError("Unknown training sampler: {}".format(sampler_name))
378
+
379
+ return {
380
+ "dataset": dataset,
381
+ "sampler": sampler,
382
+ "mapper": mapper,
383
+ "total_batch_size": cfg.SOLVER.IMS_PER_BATCH,
384
+ "aspect_ratio_grouping": cfg.DATALOADER.ASPECT_RATIO_GROUPING,
385
+ "num_workers": cfg.DATALOADER.NUM_WORKERS,
386
+ }
387
+
388
+
389
+ @configurable(from_config=_train_loader_from_config)
390
+ def build_detection_train_loader(
391
+ dataset,
392
+ *,
393
+ mapper,
394
+ sampler=None,
395
+ total_batch_size,
396
+ aspect_ratio_grouping=True,
397
+ num_workers=0,
398
+ collate_fn=None,
399
+ ):
400
+ """
401
+ Build a dataloader for object detection with some default features.
402
+
403
+ Args:
404
+ dataset (list or torch.utils.data.Dataset): a list of dataset dicts,
405
+ or a pytorch dataset (either map-style or iterable). It can be obtained
406
+ by using :func:`DatasetCatalog.get` or :func:`get_detection_dataset_dicts`.
407
+ mapper (callable): a callable which takes a sample (dict) from dataset and
408
+ returns the format to be consumed by the model.
409
+ When using cfg, the default choice is ``DatasetMapper(cfg, is_train=True)``.
410
+ sampler (torch.utils.data.sampler.Sampler or None): a sampler that produces
411
+ indices to be applied on ``dataset``.
412
+ If ``dataset`` is map-style, the default sampler is a :class:`TrainingSampler`,
413
+ which coordinates an infinite random shuffle sequence across all workers.
414
+ Sampler must be None if ``dataset`` is iterable.
415
+ total_batch_size (int): total batch size across all workers.
416
+ aspect_ratio_grouping (bool): whether to group images with similar
417
+ aspect ratio for efficiency. When enabled, it requires each
418
+ element in dataset be a dict with keys "width" and "height".
419
+ num_workers (int): number of parallel data loading workers
420
+ collate_fn: a function that determines how to do batching, same as the argument of
421
+ `torch.utils.data.DataLoader`. Defaults to do no collation and return a list of
422
+ data. No collation is OK for small batch size and simple data structures.
423
+ If your batch size is large and each sample contains too many small tensors,
424
+ it's more efficient to collate them in data loader.
425
+
426
+ Returns:
427
+ torch.utils.data.DataLoader:
428
+ a dataloader. Each output from it is a ``list[mapped_element]`` of length
429
+ ``total_batch_size / num_workers``, where ``mapped_element`` is produced
430
+ by the ``mapper``.
431
+ """
432
+ if isinstance(dataset, list):
433
+ dataset = DatasetFromList(dataset, copy=False)
434
+ if mapper is not None:
435
+ dataset = MapDataset(dataset, mapper)
436
+
437
+ if isinstance(dataset, torchdata.IterableDataset):
438
+ assert sampler is None, "sampler must be None if dataset is IterableDataset"
439
+ else:
440
+ if sampler is None:
441
+ sampler = TrainingSampler(len(dataset))
442
+ assert isinstance(sampler, torchdata.Sampler), f"Expect a Sampler but got {type(sampler)}"
443
+ return build_batch_data_loader(
444
+ dataset,
445
+ sampler,
446
+ total_batch_size,
447
+ aspect_ratio_grouping=aspect_ratio_grouping,
448
+ num_workers=num_workers,
449
+ collate_fn=collate_fn,
450
+ )
451
+
452
+
453
+ def _test_loader_from_config(cfg, dataset_name, mapper=None):
454
+ """
455
+ Uses the given `dataset_name` argument (instead of the names in cfg), because the
456
+ standard practice is to evaluate each test set individually (not combining them).
457
+ """
458
+ if isinstance(dataset_name, str):
459
+ dataset_name = [dataset_name]
460
+
461
+ dataset = get_detection_dataset_dicts(
462
+ dataset_name,
463
+ filter_empty=False,
464
+ proposal_files=[
465
+ cfg.DATASETS.PROPOSAL_FILES_TEST[list(cfg.DATASETS.TEST).index(x)] for x in dataset_name
466
+ ]
467
+ if cfg.MODEL.LOAD_PROPOSALS
468
+ else None,
469
+ )
470
+ if mapper is None:
471
+ mapper = DatasetMapper(cfg, False)
472
+ return {
473
+ "dataset": dataset,
474
+ "mapper": mapper,
475
+ "num_workers": cfg.DATALOADER.NUM_WORKERS,
476
+ "sampler": InferenceSampler(len(dataset))
477
+ if not isinstance(dataset, torchdata.IterableDataset)
478
+ else None,
479
+ }
480
+
481
+
482
+ @configurable(from_config=_test_loader_from_config)
483
+ def build_detection_test_loader(
484
+ dataset: Union[List[Any], torchdata.Dataset],
485
+ *,
486
+ mapper: Callable[[Dict[str, Any]], Any],
487
+ sampler: Optional[torchdata.Sampler] = None,
488
+ batch_size: int = 1,
489
+ num_workers: int = 0,
490
+ collate_fn: Optional[Callable[[List[Any]], Any]] = None,
491
+ ) -> torchdata.DataLoader:
492
+ """
493
+ Similar to `build_detection_train_loader`, with default batch size = 1,
494
+ and sampler = :class:`InferenceSampler`. This sampler coordinates all workers
495
+ to produce the exact set of all samples.
496
+
497
+ Args:
498
+ dataset: a list of dataset dicts,
499
+ or a pytorch dataset (either map-style or iterable). They can be obtained
500
+ by using :func:`DatasetCatalog.get` or :func:`get_detection_dataset_dicts`.
501
+ mapper: a callable which takes a sample (dict) from dataset
502
+ and returns the format to be consumed by the model.
503
+ When using cfg, the default choice is ``DatasetMapper(cfg, is_train=False)``.
504
+ sampler: a sampler that produces
505
+ indices to be applied on ``dataset``. Default to :class:`InferenceSampler`,
506
+ which splits the dataset across all workers. Sampler must be None
507
+ if `dataset` is iterable.
508
+ batch_size: the batch size of the data loader to be created.
509
+ Default to 1 image per worker since this is the standard when reporting
510
+ inference time in papers.
511
+ num_workers: number of parallel data loading workers
512
+ collate_fn: same as the argument of `torch.utils.data.DataLoader`.
513
+ Defaults to do no collation and return a list of data.
514
+
515
+ Returns:
516
+ DataLoader: a torch DataLoader, that loads the given detection
517
+ dataset, with test-time transformation and batching.
518
+
519
+ Examples:
520
+ ::
521
+ data_loader = build_detection_test_loader(
522
+ DatasetRegistry.get("my_test"),
523
+ mapper=DatasetMapper(...))
524
+
525
+ # or, instantiate with a CfgNode:
526
+ data_loader = build_detection_test_loader(cfg, "my_test")
527
+ """
528
+ if isinstance(dataset, list):
529
+ dataset = DatasetFromList(dataset, copy=False)
530
+ if mapper is not None:
531
+ dataset = MapDataset(dataset, mapper)
532
+ if isinstance(dataset, torchdata.IterableDataset):
533
+ assert sampler is None, "sampler must be None if dataset is IterableDataset"
534
+ else:
535
+ if sampler is None:
536
+ sampler = InferenceSampler(len(dataset))
537
+ return torchdata.DataLoader(
538
+ dataset,
539
+ batch_size=batch_size,
540
+ sampler=sampler,
541
+ drop_last=False,
542
+ num_workers=num_workers,
543
+ collate_fn=trivial_batch_collator if collate_fn is None else collate_fn,
544
+ )
545
+
546
+
547
+ def trivial_batch_collator(batch):
548
+ """
549
+ A batch collator that does nothing.
550
+ """
551
+ return batch
552
+
553
+
554
+ def worker_init_reset_seed(worker_id):
555
+ initial_seed = torch.initial_seed() % 2**31
556
+ seed_all_rng(initial_seed + worker_id)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/catalog.py ADDED
@@ -0,0 +1,236 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import copy
3
+ import logging
4
+ import types
5
+ from collections import UserDict
6
+ from typing import List
7
+
8
+ from annotator.oneformer.detectron2.utils.logger import log_first_n
9
+
10
+ __all__ = ["DatasetCatalog", "MetadataCatalog", "Metadata"]
11
+
12
+
13
+ class _DatasetCatalog(UserDict):
14
+ """
15
+ A global dictionary that stores information about the datasets and how to obtain them.
16
+
17
+ It contains a mapping from strings
18
+ (which are names that identify a dataset, e.g. "coco_2014_train")
19
+ to a function which parses the dataset and returns the samples in the
20
+ format of `list[dict]`.
21
+
22
+ The returned dicts should be in Detectron2 Dataset format (See DATASETS.md for details)
23
+ if used with the data loader functionalities in `data/build.py,data/detection_transform.py`.
24
+
25
+ The purpose of having this catalog is to make it easy to choose
26
+ different datasets, by just using the strings in the config.
27
+ """
28
+
29
+ def register(self, name, func):
30
+ """
31
+ Args:
32
+ name (str): the name that identifies a dataset, e.g. "coco_2014_train".
33
+ func (callable): a callable which takes no arguments and returns a list of dicts.
34
+ It must return the same results if called multiple times.
35
+ """
36
+ assert callable(func), "You must register a function with `DatasetCatalog.register`!"
37
+ assert name not in self, "Dataset '{}' is already registered!".format(name)
38
+ self[name] = func
39
+
40
+ def get(self, name):
41
+ """
42
+ Call the registered function and return its results.
43
+
44
+ Args:
45
+ name (str): the name that identifies a dataset, e.g. "coco_2014_train".
46
+
47
+ Returns:
48
+ list[dict]: dataset annotations.
49
+ """
50
+ try:
51
+ f = self[name]
52
+ except KeyError as e:
53
+ raise KeyError(
54
+ "Dataset '{}' is not registered! Available datasets are: {}".format(
55
+ name, ", ".join(list(self.keys()))
56
+ )
57
+ ) from e
58
+ return f()
59
+
60
+ def list(self) -> List[str]:
61
+ """
62
+ List all registered datasets.
63
+
64
+ Returns:
65
+ list[str]
66
+ """
67
+ return list(self.keys())
68
+
69
+ def remove(self, name):
70
+ """
71
+ Alias of ``pop``.
72
+ """
73
+ self.pop(name)
74
+
75
+ def __str__(self):
76
+ return "DatasetCatalog(registered datasets: {})".format(", ".join(self.keys()))
77
+
78
+ __repr__ = __str__
79
+
80
+
81
+ DatasetCatalog = _DatasetCatalog()
82
+ DatasetCatalog.__doc__ = (
83
+ _DatasetCatalog.__doc__
84
+ + """
85
+ .. automethod:: detectron2.data.catalog.DatasetCatalog.register
86
+ .. automethod:: detectron2.data.catalog.DatasetCatalog.get
87
+ """
88
+ )
89
+
90
+
91
+ class Metadata(types.SimpleNamespace):
92
+ """
93
+ A class that supports simple attribute setter/getter.
94
+ It is intended for storing metadata of a dataset and make it accessible globally.
95
+
96
+ Examples:
97
+ ::
98
+ # somewhere when you load the data:
99
+ MetadataCatalog.get("mydataset").thing_classes = ["person", "dog"]
100
+
101
+ # somewhere when you print statistics or visualize:
102
+ classes = MetadataCatalog.get("mydataset").thing_classes
103
+ """
104
+
105
+ # the name of the dataset
106
+ # set default to N/A so that `self.name` in the errors will not trigger getattr again
107
+ name: str = "N/A"
108
+
109
+ _RENAMED = {
110
+ "class_names": "thing_classes",
111
+ "dataset_id_to_contiguous_id": "thing_dataset_id_to_contiguous_id",
112
+ "stuff_class_names": "stuff_classes",
113
+ }
114
+
115
+ def __getattr__(self, key):
116
+ if key in self._RENAMED:
117
+ log_first_n(
118
+ logging.WARNING,
119
+ "Metadata '{}' was renamed to '{}'!".format(key, self._RENAMED[key]),
120
+ n=10,
121
+ )
122
+ return getattr(self, self._RENAMED[key])
123
+
124
+ # "name" exists in every metadata
125
+ if len(self.__dict__) > 1:
126
+ raise AttributeError(
127
+ "Attribute '{}' does not exist in the metadata of dataset '{}'. Available "
128
+ "keys are {}.".format(key, self.name, str(self.__dict__.keys()))
129
+ )
130
+ else:
131
+ raise AttributeError(
132
+ f"Attribute '{key}' does not exist in the metadata of dataset '{self.name}': "
133
+ "metadata is empty."
134
+ )
135
+
136
+ def __setattr__(self, key, val):
137
+ if key in self._RENAMED:
138
+ log_first_n(
139
+ logging.WARNING,
140
+ "Metadata '{}' was renamed to '{}'!".format(key, self._RENAMED[key]),
141
+ n=10,
142
+ )
143
+ setattr(self, self._RENAMED[key], val)
144
+
145
+ # Ensure that metadata of the same name stays consistent
146
+ try:
147
+ oldval = getattr(self, key)
148
+ assert oldval == val, (
149
+ "Attribute '{}' in the metadata of '{}' cannot be set "
150
+ "to a different value!\n{} != {}".format(key, self.name, oldval, val)
151
+ )
152
+ except AttributeError:
153
+ super().__setattr__(key, val)
154
+
155
+ def as_dict(self):
156
+ """
157
+ Returns all the metadata as a dict.
158
+ Note that modifications to the returned dict will not reflect on the Metadata object.
159
+ """
160
+ return copy.copy(self.__dict__)
161
+
162
+ def set(self, **kwargs):
163
+ """
164
+ Set multiple metadata with kwargs.
165
+ """
166
+ for k, v in kwargs.items():
167
+ setattr(self, k, v)
168
+ return self
169
+
170
+ def get(self, key, default=None):
171
+ """
172
+ Access an attribute and return its value if exists.
173
+ Otherwise return default.
174
+ """
175
+ try:
176
+ return getattr(self, key)
177
+ except AttributeError:
178
+ return default
179
+
180
+
181
+ class _MetadataCatalog(UserDict):
182
+ """
183
+ MetadataCatalog is a global dictionary that provides access to
184
+ :class:`Metadata` of a given dataset.
185
+
186
+ The metadata associated with a certain name is a singleton: once created, the
187
+ metadata will stay alive and will be returned by future calls to ``get(name)``.
188
+
189
+ It's like global variables, so don't abuse it.
190
+ It's meant for storing knowledge that's constant and shared across the execution
191
+ of the program, e.g.: the class names in COCO.
192
+ """
193
+
194
+ def get(self, name):
195
+ """
196
+ Args:
197
+ name (str): name of a dataset (e.g. coco_2014_train).
198
+
199
+ Returns:
200
+ Metadata: The :class:`Metadata` instance associated with this name,
201
+ or create an empty one if none is available.
202
+ """
203
+ assert len(name)
204
+ r = super().get(name, None)
205
+ if r is None:
206
+ r = self[name] = Metadata(name=name)
207
+ return r
208
+
209
+ def list(self):
210
+ """
211
+ List all registered metadata.
212
+
213
+ Returns:
214
+ list[str]: keys (names of datasets) of all registered metadata
215
+ """
216
+ return list(self.keys())
217
+
218
+ def remove(self, name):
219
+ """
220
+ Alias of ``pop``.
221
+ """
222
+ self.pop(name)
223
+
224
+ def __str__(self):
225
+ return "MetadataCatalog(registered metadata: {})".format(", ".join(self.keys()))
226
+
227
+ __repr__ = __str__
228
+
229
+
230
+ MetadataCatalog = _MetadataCatalog()
231
+ MetadataCatalog.__doc__ = (
232
+ _MetadataCatalog.__doc__
233
+ + """
234
+ .. automethod:: detectron2.data.catalog.MetadataCatalog.get
235
+ """
236
+ )
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/common.py ADDED
@@ -0,0 +1,301 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import contextlib
3
+ import copy
4
+ import itertools
5
+ import logging
6
+ import numpy as np
7
+ import pickle
8
+ import random
9
+ from typing import Callable, Union
10
+ import torch
11
+ import torch.utils.data as data
12
+ from torch.utils.data.sampler import Sampler
13
+
14
+ from annotator.oneformer.detectron2.utils.serialize import PicklableWrapper
15
+
16
+ __all__ = ["MapDataset", "DatasetFromList", "AspectRatioGroupedDataset", "ToIterableDataset"]
17
+
18
+ logger = logging.getLogger(__name__)
19
+
20
+
21
+ def _shard_iterator_dataloader_worker(iterable):
22
+ # Shard the iterable if we're currently inside pytorch dataloader worker.
23
+ worker_info = data.get_worker_info()
24
+ if worker_info is None or worker_info.num_workers == 1:
25
+ # do nothing
26
+ yield from iterable
27
+ else:
28
+ yield from itertools.islice(iterable, worker_info.id, None, worker_info.num_workers)
29
+
30
+
31
+ class _MapIterableDataset(data.IterableDataset):
32
+ """
33
+ Map a function over elements in an IterableDataset.
34
+
35
+ Similar to pytorch's MapIterDataPipe, but support filtering when map_func
36
+ returns None.
37
+
38
+ This class is not public-facing. Will be called by `MapDataset`.
39
+ """
40
+
41
+ def __init__(self, dataset, map_func):
42
+ self._dataset = dataset
43
+ self._map_func = PicklableWrapper(map_func) # wrap so that a lambda will work
44
+
45
+ def __len__(self):
46
+ return len(self._dataset)
47
+
48
+ def __iter__(self):
49
+ for x in map(self._map_func, self._dataset):
50
+ if x is not None:
51
+ yield x
52
+
53
+
54
+ class MapDataset(data.Dataset):
55
+ """
56
+ Map a function over the elements in a dataset.
57
+ """
58
+
59
+ def __init__(self, dataset, map_func):
60
+ """
61
+ Args:
62
+ dataset: a dataset where map function is applied. Can be either
63
+ map-style or iterable dataset. When given an iterable dataset,
64
+ the returned object will also be an iterable dataset.
65
+ map_func: a callable which maps the element in dataset. map_func can
66
+ return None to skip the data (e.g. in case of errors).
67
+ How None is handled depends on the style of `dataset`.
68
+ If `dataset` is map-style, it randomly tries other elements.
69
+ If `dataset` is iterable, it skips the data and tries the next.
70
+ """
71
+ self._dataset = dataset
72
+ self._map_func = PicklableWrapper(map_func) # wrap so that a lambda will work
73
+
74
+ self._rng = random.Random(42)
75
+ self._fallback_candidates = set(range(len(dataset)))
76
+
77
+ def __new__(cls, dataset, map_func):
78
+ is_iterable = isinstance(dataset, data.IterableDataset)
79
+ if is_iterable:
80
+ return _MapIterableDataset(dataset, map_func)
81
+ else:
82
+ return super().__new__(cls)
83
+
84
+ def __getnewargs__(self):
85
+ return self._dataset, self._map_func
86
+
87
+ def __len__(self):
88
+ return len(self._dataset)
89
+
90
+ def __getitem__(self, idx):
91
+ retry_count = 0
92
+ cur_idx = int(idx)
93
+
94
+ while True:
95
+ data = self._map_func(self._dataset[cur_idx])
96
+ if data is not None:
97
+ self._fallback_candidates.add(cur_idx)
98
+ return data
99
+
100
+ # _map_func fails for this idx, use a random new index from the pool
101
+ retry_count += 1
102
+ self._fallback_candidates.discard(cur_idx)
103
+ cur_idx = self._rng.sample(self._fallback_candidates, k=1)[0]
104
+
105
+ if retry_count >= 3:
106
+ logger = logging.getLogger(__name__)
107
+ logger.warning(
108
+ "Failed to apply `_map_func` for idx: {}, retry count: {}".format(
109
+ idx, retry_count
110
+ )
111
+ )
112
+
113
+
114
+ class _TorchSerializedList(object):
115
+ """
116
+ A list-like object whose items are serialized and stored in a torch tensor. When
117
+ launching a process that uses TorchSerializedList with "fork" start method,
118
+ the subprocess can read the same buffer without triggering copy-on-access. When
119
+ launching a process that uses TorchSerializedList with "spawn/forkserver" start
120
+ method, the list will be pickled by a special ForkingPickler registered by PyTorch
121
+ that moves data to shared memory. In both cases, this allows parent and child
122
+ processes to share RAM for the list data, hence avoids the issue in
123
+ https://github.com/pytorch/pytorch/issues/13246.
124
+
125
+ See also https://ppwwyyxx.com/blog/2022/Demystify-RAM-Usage-in-Multiprocess-DataLoader/
126
+ on how it works.
127
+ """
128
+
129
+ def __init__(self, lst: list):
130
+ self._lst = lst
131
+
132
+ def _serialize(data):
133
+ buffer = pickle.dumps(data, protocol=-1)
134
+ return np.frombuffer(buffer, dtype=np.uint8)
135
+
136
+ logger.info(
137
+ "Serializing {} elements to byte tensors and concatenating them all ...".format(
138
+ len(self._lst)
139
+ )
140
+ )
141
+ self._lst = [_serialize(x) for x in self._lst]
142
+ self._addr = np.asarray([len(x) for x in self._lst], dtype=np.int64)
143
+ self._addr = torch.from_numpy(np.cumsum(self._addr))
144
+ self._lst = torch.from_numpy(np.concatenate(self._lst))
145
+ logger.info("Serialized dataset takes {:.2f} MiB".format(len(self._lst) / 1024**2))
146
+
147
+ def __len__(self):
148
+ return len(self._addr)
149
+
150
+ def __getitem__(self, idx):
151
+ start_addr = 0 if idx == 0 else self._addr[idx - 1].item()
152
+ end_addr = self._addr[idx].item()
153
+ bytes = memoryview(self._lst[start_addr:end_addr].numpy())
154
+
155
+ # @lint-ignore PYTHONPICKLEISBAD
156
+ return pickle.loads(bytes)
157
+
158
+
159
+ _DEFAULT_DATASET_FROM_LIST_SERIALIZE_METHOD = _TorchSerializedList
160
+
161
+
162
+ @contextlib.contextmanager
163
+ def set_default_dataset_from_list_serialize_method(new):
164
+ """
165
+ Context manager for using custom serialize function when creating DatasetFromList
166
+ """
167
+
168
+ global _DEFAULT_DATASET_FROM_LIST_SERIALIZE_METHOD
169
+ orig = _DEFAULT_DATASET_FROM_LIST_SERIALIZE_METHOD
170
+ _DEFAULT_DATASET_FROM_LIST_SERIALIZE_METHOD = new
171
+ yield
172
+ _DEFAULT_DATASET_FROM_LIST_SERIALIZE_METHOD = orig
173
+
174
+
175
+ class DatasetFromList(data.Dataset):
176
+ """
177
+ Wrap a list to a torch Dataset. It produces elements of the list as data.
178
+ """
179
+
180
+ def __init__(
181
+ self,
182
+ lst: list,
183
+ copy: bool = True,
184
+ serialize: Union[bool, Callable] = True,
185
+ ):
186
+ """
187
+ Args:
188
+ lst (list): a list which contains elements to produce.
189
+ copy (bool): whether to deepcopy the element when producing it,
190
+ so that the result can be modified in place without affecting the
191
+ source in the list.
192
+ serialize (bool or callable): whether to serialize the stroage to other
193
+ backend. If `True`, the default serialize method will be used, if given
194
+ a callable, the callable will be used as serialize method.
195
+ """
196
+ self._lst = lst
197
+ self._copy = copy
198
+ if not isinstance(serialize, (bool, Callable)):
199
+ raise TypeError(f"Unsupported type for argument `serailzie`: {serialize}")
200
+ self._serialize = serialize is not False
201
+
202
+ if self._serialize:
203
+ serialize_method = (
204
+ serialize
205
+ if isinstance(serialize, Callable)
206
+ else _DEFAULT_DATASET_FROM_LIST_SERIALIZE_METHOD
207
+ )
208
+ logger.info(f"Serializing the dataset using: {serialize_method}")
209
+ self._lst = serialize_method(self._lst)
210
+
211
+ def __len__(self):
212
+ return len(self._lst)
213
+
214
+ def __getitem__(self, idx):
215
+ if self._copy and not self._serialize:
216
+ return copy.deepcopy(self._lst[idx])
217
+ else:
218
+ return self._lst[idx]
219
+
220
+
221
+ class ToIterableDataset(data.IterableDataset):
222
+ """
223
+ Convert an old indices-based (also called map-style) dataset
224
+ to an iterable-style dataset.
225
+ """
226
+
227
+ def __init__(self, dataset: data.Dataset, sampler: Sampler, shard_sampler: bool = True):
228
+ """
229
+ Args:
230
+ dataset: an old-style dataset with ``__getitem__``
231
+ sampler: a cheap iterable that produces indices to be applied on ``dataset``.
232
+ shard_sampler: whether to shard the sampler based on the current pytorch data loader
233
+ worker id. When an IterableDataset is forked by pytorch's DataLoader into multiple
234
+ workers, it is responsible for sharding its data based on worker id so that workers
235
+ don't produce identical data.
236
+
237
+ Most samplers (like our TrainingSampler) do not shard based on dataloader worker id
238
+ and this argument should be set to True. But certain samplers may be already
239
+ sharded, in that case this argument should be set to False.
240
+ """
241
+ assert not isinstance(dataset, data.IterableDataset), dataset
242
+ assert isinstance(sampler, Sampler), sampler
243
+ self.dataset = dataset
244
+ self.sampler = sampler
245
+ self.shard_sampler = shard_sampler
246
+
247
+ def __iter__(self):
248
+ if not self.shard_sampler:
249
+ sampler = self.sampler
250
+ else:
251
+ # With map-style dataset, `DataLoader(dataset, sampler)` runs the
252
+ # sampler in main process only. But `DataLoader(ToIterableDataset(dataset, sampler))`
253
+ # will run sampler in every of the N worker. So we should only keep 1/N of the ids on
254
+ # each worker. The assumption is that sampler is cheap to iterate so it's fine to
255
+ # discard ids in workers.
256
+ sampler = _shard_iterator_dataloader_worker(self.sampler)
257
+ for idx in sampler:
258
+ yield self.dataset[idx]
259
+
260
+ def __len__(self):
261
+ return len(self.sampler)
262
+
263
+
264
+ class AspectRatioGroupedDataset(data.IterableDataset):
265
+ """
266
+ Batch data that have similar aspect ratio together.
267
+ In this implementation, images whose aspect ratio < (or >) 1 will
268
+ be batched together.
269
+ This improves training speed because the images then need less padding
270
+ to form a batch.
271
+
272
+ It assumes the underlying dataset produces dicts with "width" and "height" keys.
273
+ It will then produce a list of original dicts with length = batch_size,
274
+ all with similar aspect ratios.
275
+ """
276
+
277
+ def __init__(self, dataset, batch_size):
278
+ """
279
+ Args:
280
+ dataset: an iterable. Each element must be a dict with keys
281
+ "width" and "height", which will be used to batch data.
282
+ batch_size (int):
283
+ """
284
+ self.dataset = dataset
285
+ self.batch_size = batch_size
286
+ self._buckets = [[] for _ in range(2)]
287
+ # Hard-coded two aspect ratio groups: w > h and w < h.
288
+ # Can add support for more aspect ratio groups, but doesn't seem useful
289
+
290
+ def __iter__(self):
291
+ for d in self.dataset:
292
+ w, h = d["width"], d["height"]
293
+ bucket_id = 0 if w > h else 1
294
+ bucket = self._buckets[bucket_id]
295
+ bucket.append(d)
296
+ if len(bucket) == self.batch_size:
297
+ data = bucket[:]
298
+ # Clear bucket first, because code after yield is not
299
+ # guaranteed to execute
300
+ del bucket[:]
301
+ yield data
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/dataset_mapper.py ADDED
@@ -0,0 +1,191 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import copy
3
+ import logging
4
+ import numpy as np
5
+ from typing import List, Optional, Union
6
+ import torch
7
+
8
+ from annotator.oneformer.detectron2.config import configurable
9
+
10
+ from . import detection_utils as utils
11
+ from . import transforms as T
12
+
13
+ """
14
+ This file contains the default mapping that's applied to "dataset dicts".
15
+ """
16
+
17
+ __all__ = ["DatasetMapper"]
18
+
19
+
20
+ class DatasetMapper:
21
+ """
22
+ A callable which takes a dataset dict in Detectron2 Dataset format,
23
+ and map it into a format used by the model.
24
+
25
+ This is the default callable to be used to map your dataset dict into training data.
26
+ You may need to follow it to implement your own one for customized logic,
27
+ such as a different way to read or transform images.
28
+ See :doc:`/tutorials/data_loading` for details.
29
+
30
+ The callable currently does the following:
31
+
32
+ 1. Read the image from "file_name"
33
+ 2. Applies cropping/geometric transforms to the image and annotations
34
+ 3. Prepare data and annotations to Tensor and :class:`Instances`
35
+ """
36
+
37
+ @configurable
38
+ def __init__(
39
+ self,
40
+ is_train: bool,
41
+ *,
42
+ augmentations: List[Union[T.Augmentation, T.Transform]],
43
+ image_format: str,
44
+ use_instance_mask: bool = False,
45
+ use_keypoint: bool = False,
46
+ instance_mask_format: str = "polygon",
47
+ keypoint_hflip_indices: Optional[np.ndarray] = None,
48
+ precomputed_proposal_topk: Optional[int] = None,
49
+ recompute_boxes: bool = False,
50
+ ):
51
+ """
52
+ NOTE: this interface is experimental.
53
+
54
+ Args:
55
+ is_train: whether it's used in training or inference
56
+ augmentations: a list of augmentations or deterministic transforms to apply
57
+ image_format: an image format supported by :func:`detection_utils.read_image`.
58
+ use_instance_mask: whether to process instance segmentation annotations, if available
59
+ use_keypoint: whether to process keypoint annotations if available
60
+ instance_mask_format: one of "polygon" or "bitmask". Process instance segmentation
61
+ masks into this format.
62
+ keypoint_hflip_indices: see :func:`detection_utils.create_keypoint_hflip_indices`
63
+ precomputed_proposal_topk: if given, will load pre-computed
64
+ proposals from dataset_dict and keep the top k proposals for each image.
65
+ recompute_boxes: whether to overwrite bounding box annotations
66
+ by computing tight bounding boxes from instance mask annotations.
67
+ """
68
+ if recompute_boxes:
69
+ assert use_instance_mask, "recompute_boxes requires instance masks"
70
+ # fmt: off
71
+ self.is_train = is_train
72
+ self.augmentations = T.AugmentationList(augmentations)
73
+ self.image_format = image_format
74
+ self.use_instance_mask = use_instance_mask
75
+ self.instance_mask_format = instance_mask_format
76
+ self.use_keypoint = use_keypoint
77
+ self.keypoint_hflip_indices = keypoint_hflip_indices
78
+ self.proposal_topk = precomputed_proposal_topk
79
+ self.recompute_boxes = recompute_boxes
80
+ # fmt: on
81
+ logger = logging.getLogger(__name__)
82
+ mode = "training" if is_train else "inference"
83
+ logger.info(f"[DatasetMapper] Augmentations used in {mode}: {augmentations}")
84
+
85
+ @classmethod
86
+ def from_config(cls, cfg, is_train: bool = True):
87
+ augs = utils.build_augmentation(cfg, is_train)
88
+ if cfg.INPUT.CROP.ENABLED and is_train:
89
+ augs.insert(0, T.RandomCrop(cfg.INPUT.CROP.TYPE, cfg.INPUT.CROP.SIZE))
90
+ recompute_boxes = cfg.MODEL.MASK_ON
91
+ else:
92
+ recompute_boxes = False
93
+
94
+ ret = {
95
+ "is_train": is_train,
96
+ "augmentations": augs,
97
+ "image_format": cfg.INPUT.FORMAT,
98
+ "use_instance_mask": cfg.MODEL.MASK_ON,
99
+ "instance_mask_format": cfg.INPUT.MASK_FORMAT,
100
+ "use_keypoint": cfg.MODEL.KEYPOINT_ON,
101
+ "recompute_boxes": recompute_boxes,
102
+ }
103
+
104
+ if cfg.MODEL.KEYPOINT_ON:
105
+ ret["keypoint_hflip_indices"] = utils.create_keypoint_hflip_indices(cfg.DATASETS.TRAIN)
106
+
107
+ if cfg.MODEL.LOAD_PROPOSALS:
108
+ ret["precomputed_proposal_topk"] = (
109
+ cfg.DATASETS.PRECOMPUTED_PROPOSAL_TOPK_TRAIN
110
+ if is_train
111
+ else cfg.DATASETS.PRECOMPUTED_PROPOSAL_TOPK_TEST
112
+ )
113
+ return ret
114
+
115
+ def _transform_annotations(self, dataset_dict, transforms, image_shape):
116
+ # USER: Modify this if you want to keep them for some reason.
117
+ for anno in dataset_dict["annotations"]:
118
+ if not self.use_instance_mask:
119
+ anno.pop("segmentation", None)
120
+ if not self.use_keypoint:
121
+ anno.pop("keypoints", None)
122
+
123
+ # USER: Implement additional transformations if you have other types of data
124
+ annos = [
125
+ utils.transform_instance_annotations(
126
+ obj, transforms, image_shape, keypoint_hflip_indices=self.keypoint_hflip_indices
127
+ )
128
+ for obj in dataset_dict.pop("annotations")
129
+ if obj.get("iscrowd", 0) == 0
130
+ ]
131
+ instances = utils.annotations_to_instances(
132
+ annos, image_shape, mask_format=self.instance_mask_format
133
+ )
134
+
135
+ # After transforms such as cropping are applied, the bounding box may no longer
136
+ # tightly bound the object. As an example, imagine a triangle object
137
+ # [(0,0), (2,0), (0,2)] cropped by a box [(1,0),(2,2)] (XYXY format). The tight
138
+ # bounding box of the cropped triangle should be [(1,0),(2,1)], which is not equal to
139
+ # the intersection of original bounding box and the cropping box.
140
+ if self.recompute_boxes:
141
+ instances.gt_boxes = instances.gt_masks.get_bounding_boxes()
142
+ dataset_dict["instances"] = utils.filter_empty_instances(instances)
143
+
144
+ def __call__(self, dataset_dict):
145
+ """
146
+ Args:
147
+ dataset_dict (dict): Metadata of one image, in Detectron2 Dataset format.
148
+
149
+ Returns:
150
+ dict: a format that builtin models in detectron2 accept
151
+ """
152
+ dataset_dict = copy.deepcopy(dataset_dict) # it will be modified by code below
153
+ # USER: Write your own image loading if it's not from a file
154
+ image = utils.read_image(dataset_dict["file_name"], format=self.image_format)
155
+ utils.check_image_size(dataset_dict, image)
156
+
157
+ # USER: Remove if you don't do semantic/panoptic segmentation.
158
+ if "sem_seg_file_name" in dataset_dict:
159
+ sem_seg_gt = utils.read_image(dataset_dict.pop("sem_seg_file_name"), "L").squeeze(2)
160
+ else:
161
+ sem_seg_gt = None
162
+
163
+ aug_input = T.AugInput(image, sem_seg=sem_seg_gt)
164
+ transforms = self.augmentations(aug_input)
165
+ image, sem_seg_gt = aug_input.image, aug_input.sem_seg
166
+
167
+ image_shape = image.shape[:2] # h, w
168
+ # Pytorch's dataloader is efficient on torch.Tensor due to shared-memory,
169
+ # but not efficient on large generic data structures due to the use of pickle & mp.Queue.
170
+ # Therefore it's important to use torch.Tensor.
171
+ dataset_dict["image"] = torch.as_tensor(np.ascontiguousarray(image.transpose(2, 0, 1)))
172
+ if sem_seg_gt is not None:
173
+ dataset_dict["sem_seg"] = torch.as_tensor(sem_seg_gt.astype("long"))
174
+
175
+ # USER: Remove if you don't use pre-computed proposals.
176
+ # Most users would not need this feature.
177
+ if self.proposal_topk is not None:
178
+ utils.transform_proposals(
179
+ dataset_dict, image_shape, transforms, proposal_topk=self.proposal_topk
180
+ )
181
+
182
+ if not self.is_train:
183
+ # USER: Modify this if you want to keep them for some reason.
184
+ dataset_dict.pop("annotations", None)
185
+ dataset_dict.pop("sem_seg_file_name", None)
186
+ return dataset_dict
187
+
188
+ if "annotations" in dataset_dict:
189
+ self._transform_annotations(dataset_dict, transforms, image_shape)
190
+
191
+ return dataset_dict
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+ ### Common Datasets
4
+
5
+ The dataset implemented here do not need to load the data into the final format.
6
+ It should provide the minimal data structure needed to use the dataset, so it can be very efficient.
7
+
8
+ For example, for an image dataset, just provide the file names and labels, but don't read the images.
9
+ Let the downstream decide how to read.
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/__init__.py ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from .coco import load_coco_json, load_sem_seg, register_coco_instances, convert_to_coco_json
3
+ from .coco_panoptic import register_coco_panoptic, register_coco_panoptic_separated
4
+ from .lvis import load_lvis_json, register_lvis_instances, get_lvis_instances_meta
5
+ from .pascal_voc import load_voc_instances, register_pascal_voc
6
+ from . import builtin as _builtin # ensure the builtin datasets are registered
7
+
8
+
9
+ __all__ = [k for k in globals().keys() if not k.startswith("_")]
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/builtin.py ADDED
@@ -0,0 +1,259 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+
5
+ """
6
+ This file registers pre-defined datasets at hard-coded paths, and their metadata.
7
+
8
+ We hard-code metadata for common datasets. This will enable:
9
+ 1. Consistency check when loading the datasets
10
+ 2. Use models on these standard datasets directly and run demos,
11
+ without having to download the dataset annotations
12
+
13
+ We hard-code some paths to the dataset that's assumed to
14
+ exist in "./datasets/".
15
+
16
+ Users SHOULD NOT use this file to create new dataset / metadata for new dataset.
17
+ To add new dataset, refer to the tutorial "docs/DATASETS.md".
18
+ """
19
+
20
+ import os
21
+
22
+ from annotator.oneformer.detectron2.data import DatasetCatalog, MetadataCatalog
23
+
24
+ from .builtin_meta import ADE20K_SEM_SEG_CATEGORIES, _get_builtin_metadata
25
+ from .cityscapes import load_cityscapes_instances, load_cityscapes_semantic
26
+ from .cityscapes_panoptic import register_all_cityscapes_panoptic
27
+ from .coco import load_sem_seg, register_coco_instances
28
+ from .coco_panoptic import register_coco_panoptic, register_coco_panoptic_separated
29
+ from .lvis import get_lvis_instances_meta, register_lvis_instances
30
+ from .pascal_voc import register_pascal_voc
31
+
32
+ # ==== Predefined datasets and splits for COCO ==========
33
+
34
+ _PREDEFINED_SPLITS_COCO = {}
35
+ _PREDEFINED_SPLITS_COCO["coco"] = {
36
+ "coco_2014_train": ("coco/train2014", "coco/annotations/instances_train2014.json"),
37
+ "coco_2014_val": ("coco/val2014", "coco/annotations/instances_val2014.json"),
38
+ "coco_2014_minival": ("coco/val2014", "coco/annotations/instances_minival2014.json"),
39
+ "coco_2014_valminusminival": (
40
+ "coco/val2014",
41
+ "coco/annotations/instances_valminusminival2014.json",
42
+ ),
43
+ "coco_2017_train": ("coco/train2017", "coco/annotations/instances_train2017.json"),
44
+ "coco_2017_val": ("coco/val2017", "coco/annotations/instances_val2017.json"),
45
+ "coco_2017_test": ("coco/test2017", "coco/annotations/image_info_test2017.json"),
46
+ "coco_2017_test-dev": ("coco/test2017", "coco/annotations/image_info_test-dev2017.json"),
47
+ "coco_2017_val_100": ("coco/val2017", "coco/annotations/instances_val2017_100.json"),
48
+ }
49
+
50
+ _PREDEFINED_SPLITS_COCO["coco_person"] = {
51
+ "keypoints_coco_2014_train": (
52
+ "coco/train2014",
53
+ "coco/annotations/person_keypoints_train2014.json",
54
+ ),
55
+ "keypoints_coco_2014_val": ("coco/val2014", "coco/annotations/person_keypoints_val2014.json"),
56
+ "keypoints_coco_2014_minival": (
57
+ "coco/val2014",
58
+ "coco/annotations/person_keypoints_minival2014.json",
59
+ ),
60
+ "keypoints_coco_2014_valminusminival": (
61
+ "coco/val2014",
62
+ "coco/annotations/person_keypoints_valminusminival2014.json",
63
+ ),
64
+ "keypoints_coco_2017_train": (
65
+ "coco/train2017",
66
+ "coco/annotations/person_keypoints_train2017.json",
67
+ ),
68
+ "keypoints_coco_2017_val": ("coco/val2017", "coco/annotations/person_keypoints_val2017.json"),
69
+ "keypoints_coco_2017_val_100": (
70
+ "coco/val2017",
71
+ "coco/annotations/person_keypoints_val2017_100.json",
72
+ ),
73
+ }
74
+
75
+
76
+ _PREDEFINED_SPLITS_COCO_PANOPTIC = {
77
+ "coco_2017_train_panoptic": (
78
+ # This is the original panoptic annotation directory
79
+ "coco/panoptic_train2017",
80
+ "coco/annotations/panoptic_train2017.json",
81
+ # This directory contains semantic annotations that are
82
+ # converted from panoptic annotations.
83
+ # It is used by PanopticFPN.
84
+ # You can use the script at detectron2/datasets/prepare_panoptic_fpn.py
85
+ # to create these directories.
86
+ "coco/panoptic_stuff_train2017",
87
+ ),
88
+ "coco_2017_val_panoptic": (
89
+ "coco/panoptic_val2017",
90
+ "coco/annotations/panoptic_val2017.json",
91
+ "coco/panoptic_stuff_val2017",
92
+ ),
93
+ "coco_2017_val_100_panoptic": (
94
+ "coco/panoptic_val2017_100",
95
+ "coco/annotations/panoptic_val2017_100.json",
96
+ "coco/panoptic_stuff_val2017_100",
97
+ ),
98
+ }
99
+
100
+
101
+ def register_all_coco(root):
102
+ for dataset_name, splits_per_dataset in _PREDEFINED_SPLITS_COCO.items():
103
+ for key, (image_root, json_file) in splits_per_dataset.items():
104
+ # Assume pre-defined datasets live in `./datasets`.
105
+ register_coco_instances(
106
+ key,
107
+ _get_builtin_metadata(dataset_name),
108
+ os.path.join(root, json_file) if "://" not in json_file else json_file,
109
+ os.path.join(root, image_root),
110
+ )
111
+
112
+ for (
113
+ prefix,
114
+ (panoptic_root, panoptic_json, semantic_root),
115
+ ) in _PREDEFINED_SPLITS_COCO_PANOPTIC.items():
116
+ prefix_instances = prefix[: -len("_panoptic")]
117
+ instances_meta = MetadataCatalog.get(prefix_instances)
118
+ image_root, instances_json = instances_meta.image_root, instances_meta.json_file
119
+ # The "separated" version of COCO panoptic segmentation dataset,
120
+ # e.g. used by Panoptic FPN
121
+ register_coco_panoptic_separated(
122
+ prefix,
123
+ _get_builtin_metadata("coco_panoptic_separated"),
124
+ image_root,
125
+ os.path.join(root, panoptic_root),
126
+ os.path.join(root, panoptic_json),
127
+ os.path.join(root, semantic_root),
128
+ instances_json,
129
+ )
130
+ # The "standard" version of COCO panoptic segmentation dataset,
131
+ # e.g. used by Panoptic-DeepLab
132
+ register_coco_panoptic(
133
+ prefix,
134
+ _get_builtin_metadata("coco_panoptic_standard"),
135
+ image_root,
136
+ os.path.join(root, panoptic_root),
137
+ os.path.join(root, panoptic_json),
138
+ instances_json,
139
+ )
140
+
141
+
142
+ # ==== Predefined datasets and splits for LVIS ==========
143
+
144
+
145
+ _PREDEFINED_SPLITS_LVIS = {
146
+ "lvis_v1": {
147
+ "lvis_v1_train": ("coco/", "lvis/lvis_v1_train.json"),
148
+ "lvis_v1_val": ("coco/", "lvis/lvis_v1_val.json"),
149
+ "lvis_v1_test_dev": ("coco/", "lvis/lvis_v1_image_info_test_dev.json"),
150
+ "lvis_v1_test_challenge": ("coco/", "lvis/lvis_v1_image_info_test_challenge.json"),
151
+ },
152
+ "lvis_v0.5": {
153
+ "lvis_v0.5_train": ("coco/", "lvis/lvis_v0.5_train.json"),
154
+ "lvis_v0.5_val": ("coco/", "lvis/lvis_v0.5_val.json"),
155
+ "lvis_v0.5_val_rand_100": ("coco/", "lvis/lvis_v0.5_val_rand_100.json"),
156
+ "lvis_v0.5_test": ("coco/", "lvis/lvis_v0.5_image_info_test.json"),
157
+ },
158
+ "lvis_v0.5_cocofied": {
159
+ "lvis_v0.5_train_cocofied": ("coco/", "lvis/lvis_v0.5_train_cocofied.json"),
160
+ "lvis_v0.5_val_cocofied": ("coco/", "lvis/lvis_v0.5_val_cocofied.json"),
161
+ },
162
+ }
163
+
164
+
165
+ def register_all_lvis(root):
166
+ for dataset_name, splits_per_dataset in _PREDEFINED_SPLITS_LVIS.items():
167
+ for key, (image_root, json_file) in splits_per_dataset.items():
168
+ register_lvis_instances(
169
+ key,
170
+ get_lvis_instances_meta(dataset_name),
171
+ os.path.join(root, json_file) if "://" not in json_file else json_file,
172
+ os.path.join(root, image_root),
173
+ )
174
+
175
+
176
+ # ==== Predefined splits for raw cityscapes images ===========
177
+ _RAW_CITYSCAPES_SPLITS = {
178
+ "cityscapes_fine_{task}_train": ("cityscapes/leftImg8bit/train/", "cityscapes/gtFine/train/"),
179
+ "cityscapes_fine_{task}_val": ("cityscapes/leftImg8bit/val/", "cityscapes/gtFine/val/"),
180
+ "cityscapes_fine_{task}_test": ("cityscapes/leftImg8bit/test/", "cityscapes/gtFine/test/"),
181
+ }
182
+
183
+
184
+ def register_all_cityscapes(root):
185
+ for key, (image_dir, gt_dir) in _RAW_CITYSCAPES_SPLITS.items():
186
+ meta = _get_builtin_metadata("cityscapes")
187
+ image_dir = os.path.join(root, image_dir)
188
+ gt_dir = os.path.join(root, gt_dir)
189
+
190
+ inst_key = key.format(task="instance_seg")
191
+ DatasetCatalog.register(
192
+ inst_key,
193
+ lambda x=image_dir, y=gt_dir: load_cityscapes_instances(
194
+ x, y, from_json=True, to_polygons=True
195
+ ),
196
+ )
197
+ MetadataCatalog.get(inst_key).set(
198
+ image_dir=image_dir, gt_dir=gt_dir, evaluator_type="cityscapes_instance", **meta
199
+ )
200
+
201
+ sem_key = key.format(task="sem_seg")
202
+ DatasetCatalog.register(
203
+ sem_key, lambda x=image_dir, y=gt_dir: load_cityscapes_semantic(x, y)
204
+ )
205
+ MetadataCatalog.get(sem_key).set(
206
+ image_dir=image_dir,
207
+ gt_dir=gt_dir,
208
+ evaluator_type="cityscapes_sem_seg",
209
+ ignore_label=255,
210
+ **meta,
211
+ )
212
+
213
+
214
+ # ==== Predefined splits for PASCAL VOC ===========
215
+ def register_all_pascal_voc(root):
216
+ SPLITS = [
217
+ ("voc_2007_trainval", "VOC2007", "trainval"),
218
+ ("voc_2007_train", "VOC2007", "train"),
219
+ ("voc_2007_val", "VOC2007", "val"),
220
+ ("voc_2007_test", "VOC2007", "test"),
221
+ ("voc_2012_trainval", "VOC2012", "trainval"),
222
+ ("voc_2012_train", "VOC2012", "train"),
223
+ ("voc_2012_val", "VOC2012", "val"),
224
+ ]
225
+ for name, dirname, split in SPLITS:
226
+ year = 2007 if "2007" in name else 2012
227
+ register_pascal_voc(name, os.path.join(root, dirname), split, year)
228
+ MetadataCatalog.get(name).evaluator_type = "pascal_voc"
229
+
230
+
231
+ def register_all_ade20k(root):
232
+ root = os.path.join(root, "ADEChallengeData2016")
233
+ for name, dirname in [("train", "training"), ("val", "validation")]:
234
+ image_dir = os.path.join(root, "images", dirname)
235
+ gt_dir = os.path.join(root, "annotations_detectron2", dirname)
236
+ name = f"ade20k_sem_seg_{name}"
237
+ DatasetCatalog.register(
238
+ name, lambda x=image_dir, y=gt_dir: load_sem_seg(y, x, gt_ext="png", image_ext="jpg")
239
+ )
240
+ MetadataCatalog.get(name).set(
241
+ stuff_classes=ADE20K_SEM_SEG_CATEGORIES[:],
242
+ image_root=image_dir,
243
+ sem_seg_root=gt_dir,
244
+ evaluator_type="sem_seg",
245
+ ignore_label=255,
246
+ )
247
+
248
+
249
+ # True for open source;
250
+ # Internally at fb, we register them elsewhere
251
+ if __name__.endswith(".builtin"):
252
+ # Assume pre-defined datasets live in `./datasets`.
253
+ _root = os.path.expanduser(os.getenv("DETECTRON2_DATASETS", "datasets"))
254
+ register_all_coco(_root)
255
+ register_all_lvis(_root)
256
+ register_all_cityscapes(_root)
257
+ register_all_cityscapes_panoptic(_root)
258
+ register_all_pascal_voc(_root)
259
+ register_all_ade20k(_root)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/builtin_meta.py ADDED
@@ -0,0 +1,350 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ """
5
+ Note:
6
+ For your custom dataset, there is no need to hard-code metadata anywhere in the code.
7
+ For example, for COCO-format dataset, metadata will be obtained automatically
8
+ when calling `load_coco_json`. For other dataset, metadata may also be obtained in other ways
9
+ during loading.
10
+
11
+ However, we hard-coded metadata for a few common dataset here.
12
+ The only goal is to allow users who don't have these dataset to use pre-trained models.
13
+ Users don't have to download a COCO json (which contains metadata), in order to visualize a
14
+ COCO model (with correct class names and colors).
15
+ """
16
+
17
+
18
+ # All coco categories, together with their nice-looking visualization colors
19
+ # It's from https://github.com/cocodataset/panopticapi/blob/master/panoptic_coco_categories.json
20
+ COCO_CATEGORIES = [
21
+ {"color": [220, 20, 60], "isthing": 1, "id": 1, "name": "person"},
22
+ {"color": [119, 11, 32], "isthing": 1, "id": 2, "name": "bicycle"},
23
+ {"color": [0, 0, 142], "isthing": 1, "id": 3, "name": "car"},
24
+ {"color": [0, 0, 230], "isthing": 1, "id": 4, "name": "motorcycle"},
25
+ {"color": [106, 0, 228], "isthing": 1, "id": 5, "name": "airplane"},
26
+ {"color": [0, 60, 100], "isthing": 1, "id": 6, "name": "bus"},
27
+ {"color": [0, 80, 100], "isthing": 1, "id": 7, "name": "train"},
28
+ {"color": [0, 0, 70], "isthing": 1, "id": 8, "name": "truck"},
29
+ {"color": [0, 0, 192], "isthing": 1, "id": 9, "name": "boat"},
30
+ {"color": [250, 170, 30], "isthing": 1, "id": 10, "name": "traffic light"},
31
+ {"color": [100, 170, 30], "isthing": 1, "id": 11, "name": "fire hydrant"},
32
+ {"color": [220, 220, 0], "isthing": 1, "id": 13, "name": "stop sign"},
33
+ {"color": [175, 116, 175], "isthing": 1, "id": 14, "name": "parking meter"},
34
+ {"color": [250, 0, 30], "isthing": 1, "id": 15, "name": "bench"},
35
+ {"color": [165, 42, 42], "isthing": 1, "id": 16, "name": "bird"},
36
+ {"color": [255, 77, 255], "isthing": 1, "id": 17, "name": "cat"},
37
+ {"color": [0, 226, 252], "isthing": 1, "id": 18, "name": "dog"},
38
+ {"color": [182, 182, 255], "isthing": 1, "id": 19, "name": "horse"},
39
+ {"color": [0, 82, 0], "isthing": 1, "id": 20, "name": "sheep"},
40
+ {"color": [120, 166, 157], "isthing": 1, "id": 21, "name": "cow"},
41
+ {"color": [110, 76, 0], "isthing": 1, "id": 22, "name": "elephant"},
42
+ {"color": [174, 57, 255], "isthing": 1, "id": 23, "name": "bear"},
43
+ {"color": [199, 100, 0], "isthing": 1, "id": 24, "name": "zebra"},
44
+ {"color": [72, 0, 118], "isthing": 1, "id": 25, "name": "giraffe"},
45
+ {"color": [255, 179, 240], "isthing": 1, "id": 27, "name": "backpack"},
46
+ {"color": [0, 125, 92], "isthing": 1, "id": 28, "name": "umbrella"},
47
+ {"color": [209, 0, 151], "isthing": 1, "id": 31, "name": "handbag"},
48
+ {"color": [188, 208, 182], "isthing": 1, "id": 32, "name": "tie"},
49
+ {"color": [0, 220, 176], "isthing": 1, "id": 33, "name": "suitcase"},
50
+ {"color": [255, 99, 164], "isthing": 1, "id": 34, "name": "frisbee"},
51
+ {"color": [92, 0, 73], "isthing": 1, "id": 35, "name": "skis"},
52
+ {"color": [133, 129, 255], "isthing": 1, "id": 36, "name": "snowboard"},
53
+ {"color": [78, 180, 255], "isthing": 1, "id": 37, "name": "sports ball"},
54
+ {"color": [0, 228, 0], "isthing": 1, "id": 38, "name": "kite"},
55
+ {"color": [174, 255, 243], "isthing": 1, "id": 39, "name": "baseball bat"},
56
+ {"color": [45, 89, 255], "isthing": 1, "id": 40, "name": "baseball glove"},
57
+ {"color": [134, 134, 103], "isthing": 1, "id": 41, "name": "skateboard"},
58
+ {"color": [145, 148, 174], "isthing": 1, "id": 42, "name": "surfboard"},
59
+ {"color": [255, 208, 186], "isthing": 1, "id": 43, "name": "tennis racket"},
60
+ {"color": [197, 226, 255], "isthing": 1, "id": 44, "name": "bottle"},
61
+ {"color": [171, 134, 1], "isthing": 1, "id": 46, "name": "wine glass"},
62
+ {"color": [109, 63, 54], "isthing": 1, "id": 47, "name": "cup"},
63
+ {"color": [207, 138, 255], "isthing": 1, "id": 48, "name": "fork"},
64
+ {"color": [151, 0, 95], "isthing": 1, "id": 49, "name": "knife"},
65
+ {"color": [9, 80, 61], "isthing": 1, "id": 50, "name": "spoon"},
66
+ {"color": [84, 105, 51], "isthing": 1, "id": 51, "name": "bowl"},
67
+ {"color": [74, 65, 105], "isthing": 1, "id": 52, "name": "banana"},
68
+ {"color": [166, 196, 102], "isthing": 1, "id": 53, "name": "apple"},
69
+ {"color": [208, 195, 210], "isthing": 1, "id": 54, "name": "sandwich"},
70
+ {"color": [255, 109, 65], "isthing": 1, "id": 55, "name": "orange"},
71
+ {"color": [0, 143, 149], "isthing": 1, "id": 56, "name": "broccoli"},
72
+ {"color": [179, 0, 194], "isthing": 1, "id": 57, "name": "carrot"},
73
+ {"color": [209, 99, 106], "isthing": 1, "id": 58, "name": "hot dog"},
74
+ {"color": [5, 121, 0], "isthing": 1, "id": 59, "name": "pizza"},
75
+ {"color": [227, 255, 205], "isthing": 1, "id": 60, "name": "donut"},
76
+ {"color": [147, 186, 208], "isthing": 1, "id": 61, "name": "cake"},
77
+ {"color": [153, 69, 1], "isthing": 1, "id": 62, "name": "chair"},
78
+ {"color": [3, 95, 161], "isthing": 1, "id": 63, "name": "couch"},
79
+ {"color": [163, 255, 0], "isthing": 1, "id": 64, "name": "potted plant"},
80
+ {"color": [119, 0, 170], "isthing": 1, "id": 65, "name": "bed"},
81
+ {"color": [0, 182, 199], "isthing": 1, "id": 67, "name": "dining table"},
82
+ {"color": [0, 165, 120], "isthing": 1, "id": 70, "name": "toilet"},
83
+ {"color": [183, 130, 88], "isthing": 1, "id": 72, "name": "tv"},
84
+ {"color": [95, 32, 0], "isthing": 1, "id": 73, "name": "laptop"},
85
+ {"color": [130, 114, 135], "isthing": 1, "id": 74, "name": "mouse"},
86
+ {"color": [110, 129, 133], "isthing": 1, "id": 75, "name": "remote"},
87
+ {"color": [166, 74, 118], "isthing": 1, "id": 76, "name": "keyboard"},
88
+ {"color": [219, 142, 185], "isthing": 1, "id": 77, "name": "cell phone"},
89
+ {"color": [79, 210, 114], "isthing": 1, "id": 78, "name": "microwave"},
90
+ {"color": [178, 90, 62], "isthing": 1, "id": 79, "name": "oven"},
91
+ {"color": [65, 70, 15], "isthing": 1, "id": 80, "name": "toaster"},
92
+ {"color": [127, 167, 115], "isthing": 1, "id": 81, "name": "sink"},
93
+ {"color": [59, 105, 106], "isthing": 1, "id": 82, "name": "refrigerator"},
94
+ {"color": [142, 108, 45], "isthing": 1, "id": 84, "name": "book"},
95
+ {"color": [196, 172, 0], "isthing": 1, "id": 85, "name": "clock"},
96
+ {"color": [95, 54, 80], "isthing": 1, "id": 86, "name": "vase"},
97
+ {"color": [128, 76, 255], "isthing": 1, "id": 87, "name": "scissors"},
98
+ {"color": [201, 57, 1], "isthing": 1, "id": 88, "name": "teddy bear"},
99
+ {"color": [246, 0, 122], "isthing": 1, "id": 89, "name": "hair drier"},
100
+ {"color": [191, 162, 208], "isthing": 1, "id": 90, "name": "toothbrush"},
101
+ {"color": [255, 255, 128], "isthing": 0, "id": 92, "name": "banner"},
102
+ {"color": [147, 211, 203], "isthing": 0, "id": 93, "name": "blanket"},
103
+ {"color": [150, 100, 100], "isthing": 0, "id": 95, "name": "bridge"},
104
+ {"color": [168, 171, 172], "isthing": 0, "id": 100, "name": "cardboard"},
105
+ {"color": [146, 112, 198], "isthing": 0, "id": 107, "name": "counter"},
106
+ {"color": [210, 170, 100], "isthing": 0, "id": 109, "name": "curtain"},
107
+ {"color": [92, 136, 89], "isthing": 0, "id": 112, "name": "door-stuff"},
108
+ {"color": [218, 88, 184], "isthing": 0, "id": 118, "name": "floor-wood"},
109
+ {"color": [241, 129, 0], "isthing": 0, "id": 119, "name": "flower"},
110
+ {"color": [217, 17, 255], "isthing": 0, "id": 122, "name": "fruit"},
111
+ {"color": [124, 74, 181], "isthing": 0, "id": 125, "name": "gravel"},
112
+ {"color": [70, 70, 70], "isthing": 0, "id": 128, "name": "house"},
113
+ {"color": [255, 228, 255], "isthing": 0, "id": 130, "name": "light"},
114
+ {"color": [154, 208, 0], "isthing": 0, "id": 133, "name": "mirror-stuff"},
115
+ {"color": [193, 0, 92], "isthing": 0, "id": 138, "name": "net"},
116
+ {"color": [76, 91, 113], "isthing": 0, "id": 141, "name": "pillow"},
117
+ {"color": [255, 180, 195], "isthing": 0, "id": 144, "name": "platform"},
118
+ {"color": [106, 154, 176], "isthing": 0, "id": 145, "name": "playingfield"},
119
+ {"color": [230, 150, 140], "isthing": 0, "id": 147, "name": "railroad"},
120
+ {"color": [60, 143, 255], "isthing": 0, "id": 148, "name": "river"},
121
+ {"color": [128, 64, 128], "isthing": 0, "id": 149, "name": "road"},
122
+ {"color": [92, 82, 55], "isthing": 0, "id": 151, "name": "roof"},
123
+ {"color": [254, 212, 124], "isthing": 0, "id": 154, "name": "sand"},
124
+ {"color": [73, 77, 174], "isthing": 0, "id": 155, "name": "sea"},
125
+ {"color": [255, 160, 98], "isthing": 0, "id": 156, "name": "shelf"},
126
+ {"color": [255, 255, 255], "isthing": 0, "id": 159, "name": "snow"},
127
+ {"color": [104, 84, 109], "isthing": 0, "id": 161, "name": "stairs"},
128
+ {"color": [169, 164, 131], "isthing": 0, "id": 166, "name": "tent"},
129
+ {"color": [225, 199, 255], "isthing": 0, "id": 168, "name": "towel"},
130
+ {"color": [137, 54, 74], "isthing": 0, "id": 171, "name": "wall-brick"},
131
+ {"color": [135, 158, 223], "isthing": 0, "id": 175, "name": "wall-stone"},
132
+ {"color": [7, 246, 231], "isthing": 0, "id": 176, "name": "wall-tile"},
133
+ {"color": [107, 255, 200], "isthing": 0, "id": 177, "name": "wall-wood"},
134
+ {"color": [58, 41, 149], "isthing": 0, "id": 178, "name": "water-other"},
135
+ {"color": [183, 121, 142], "isthing": 0, "id": 180, "name": "window-blind"},
136
+ {"color": [255, 73, 97], "isthing": 0, "id": 181, "name": "window-other"},
137
+ {"color": [107, 142, 35], "isthing": 0, "id": 184, "name": "tree-merged"},
138
+ {"color": [190, 153, 153], "isthing": 0, "id": 185, "name": "fence-merged"},
139
+ {"color": [146, 139, 141], "isthing": 0, "id": 186, "name": "ceiling-merged"},
140
+ {"color": [70, 130, 180], "isthing": 0, "id": 187, "name": "sky-other-merged"},
141
+ {"color": [134, 199, 156], "isthing": 0, "id": 188, "name": "cabinet-merged"},
142
+ {"color": [209, 226, 140], "isthing": 0, "id": 189, "name": "table-merged"},
143
+ {"color": [96, 36, 108], "isthing": 0, "id": 190, "name": "floor-other-merged"},
144
+ {"color": [96, 96, 96], "isthing": 0, "id": 191, "name": "pavement-merged"},
145
+ {"color": [64, 170, 64], "isthing": 0, "id": 192, "name": "mountain-merged"},
146
+ {"color": [152, 251, 152], "isthing": 0, "id": 193, "name": "grass-merged"},
147
+ {"color": [208, 229, 228], "isthing": 0, "id": 194, "name": "dirt-merged"},
148
+ {"color": [206, 186, 171], "isthing": 0, "id": 195, "name": "paper-merged"},
149
+ {"color": [152, 161, 64], "isthing": 0, "id": 196, "name": "food-other-merged"},
150
+ {"color": [116, 112, 0], "isthing": 0, "id": 197, "name": "building-other-merged"},
151
+ {"color": [0, 114, 143], "isthing": 0, "id": 198, "name": "rock-merged"},
152
+ {"color": [102, 102, 156], "isthing": 0, "id": 199, "name": "wall-other-merged"},
153
+ {"color": [250, 141, 255], "isthing": 0, "id": 200, "name": "rug-merged"},
154
+ ]
155
+
156
+ # fmt: off
157
+ COCO_PERSON_KEYPOINT_NAMES = (
158
+ "nose",
159
+ "left_eye", "right_eye",
160
+ "left_ear", "right_ear",
161
+ "left_shoulder", "right_shoulder",
162
+ "left_elbow", "right_elbow",
163
+ "left_wrist", "right_wrist",
164
+ "left_hip", "right_hip",
165
+ "left_knee", "right_knee",
166
+ "left_ankle", "right_ankle",
167
+ )
168
+ # fmt: on
169
+
170
+ # Pairs of keypoints that should be exchanged under horizontal flipping
171
+ COCO_PERSON_KEYPOINT_FLIP_MAP = (
172
+ ("left_eye", "right_eye"),
173
+ ("left_ear", "right_ear"),
174
+ ("left_shoulder", "right_shoulder"),
175
+ ("left_elbow", "right_elbow"),
176
+ ("left_wrist", "right_wrist"),
177
+ ("left_hip", "right_hip"),
178
+ ("left_knee", "right_knee"),
179
+ ("left_ankle", "right_ankle"),
180
+ )
181
+
182
+ # rules for pairs of keypoints to draw a line between, and the line color to use.
183
+ KEYPOINT_CONNECTION_RULES = [
184
+ # face
185
+ ("left_ear", "left_eye", (102, 204, 255)),
186
+ ("right_ear", "right_eye", (51, 153, 255)),
187
+ ("left_eye", "nose", (102, 0, 204)),
188
+ ("nose", "right_eye", (51, 102, 255)),
189
+ # upper-body
190
+ ("left_shoulder", "right_shoulder", (255, 128, 0)),
191
+ ("left_shoulder", "left_elbow", (153, 255, 204)),
192
+ ("right_shoulder", "right_elbow", (128, 229, 255)),
193
+ ("left_elbow", "left_wrist", (153, 255, 153)),
194
+ ("right_elbow", "right_wrist", (102, 255, 224)),
195
+ # lower-body
196
+ ("left_hip", "right_hip", (255, 102, 0)),
197
+ ("left_hip", "left_knee", (255, 255, 77)),
198
+ ("right_hip", "right_knee", (153, 255, 204)),
199
+ ("left_knee", "left_ankle", (191, 255, 128)),
200
+ ("right_knee", "right_ankle", (255, 195, 77)),
201
+ ]
202
+
203
+ # All Cityscapes categories, together with their nice-looking visualization colors
204
+ # It's from https://github.com/mcordts/cityscapesScripts/blob/master/cityscapesscripts/helpers/labels.py # noqa
205
+ CITYSCAPES_CATEGORIES = [
206
+ {"color": (128, 64, 128), "isthing": 0, "id": 7, "trainId": 0, "name": "road"},
207
+ {"color": (244, 35, 232), "isthing": 0, "id": 8, "trainId": 1, "name": "sidewalk"},
208
+ {"color": (70, 70, 70), "isthing": 0, "id": 11, "trainId": 2, "name": "building"},
209
+ {"color": (102, 102, 156), "isthing": 0, "id": 12, "trainId": 3, "name": "wall"},
210
+ {"color": (190, 153, 153), "isthing": 0, "id": 13, "trainId": 4, "name": "fence"},
211
+ {"color": (153, 153, 153), "isthing": 0, "id": 17, "trainId": 5, "name": "pole"},
212
+ {"color": (250, 170, 30), "isthing": 0, "id": 19, "trainId": 6, "name": "traffic light"},
213
+ {"color": (220, 220, 0), "isthing": 0, "id": 20, "trainId": 7, "name": "traffic sign"},
214
+ {"color": (107, 142, 35), "isthing": 0, "id": 21, "trainId": 8, "name": "vegetation"},
215
+ {"color": (152, 251, 152), "isthing": 0, "id": 22, "trainId": 9, "name": "terrain"},
216
+ {"color": (70, 130, 180), "isthing": 0, "id": 23, "trainId": 10, "name": "sky"},
217
+ {"color": (220, 20, 60), "isthing": 1, "id": 24, "trainId": 11, "name": "person"},
218
+ {"color": (255, 0, 0), "isthing": 1, "id": 25, "trainId": 12, "name": "rider"},
219
+ {"color": (0, 0, 142), "isthing": 1, "id": 26, "trainId": 13, "name": "car"},
220
+ {"color": (0, 0, 70), "isthing": 1, "id": 27, "trainId": 14, "name": "truck"},
221
+ {"color": (0, 60, 100), "isthing": 1, "id": 28, "trainId": 15, "name": "bus"},
222
+ {"color": (0, 80, 100), "isthing": 1, "id": 31, "trainId": 16, "name": "train"},
223
+ {"color": (0, 0, 230), "isthing": 1, "id": 32, "trainId": 17, "name": "motorcycle"},
224
+ {"color": (119, 11, 32), "isthing": 1, "id": 33, "trainId": 18, "name": "bicycle"},
225
+ ]
226
+
227
+ # fmt: off
228
+ ADE20K_SEM_SEG_CATEGORIES = [
229
+ "wall", "building", "sky", "floor", "tree", "ceiling", "road, route", "bed", "window ", "grass", "cabinet", "sidewalk, pavement", "person", "earth, ground", "door", "table", "mountain, mount", "plant", "curtain", "chair", "car", "water", "painting, picture", "sofa", "shelf", "house", "sea", "mirror", "rug", "field", "armchair", "seat", "fence", "desk", "rock, stone", "wardrobe, closet, press", "lamp", "tub", "rail", "cushion", "base, pedestal, stand", "box", "column, pillar", "signboard, sign", "chest of drawers, chest, bureau, dresser", "counter", "sand", "sink", "skyscraper", "fireplace", "refrigerator, icebox", "grandstand, covered stand", "path", "stairs", "runway", "case, display case, showcase, vitrine", "pool table, billiard table, snooker table", "pillow", "screen door, screen", "stairway, staircase", "river", "bridge, span", "bookcase", "blind, screen", "coffee table", "toilet, can, commode, crapper, pot, potty, stool, throne", "flower", "book", "hill", "bench", "countertop", "stove", "palm, palm tree", "kitchen island", "computer", "swivel chair", "boat", "bar", "arcade machine", "hovel, hut, hutch, shack, shanty", "bus", "towel", "light", "truck", "tower", "chandelier", "awning, sunshade, sunblind", "street lamp", "booth", "tv", "plane", "dirt track", "clothes", "pole", "land, ground, soil", "bannister, banister, balustrade, balusters, handrail", "escalator, moving staircase, moving stairway", "ottoman, pouf, pouffe, puff, hassock", "bottle", "buffet, counter, sideboard", "poster, posting, placard, notice, bill, card", "stage", "van", "ship", "fountain", "conveyer belt, conveyor belt, conveyer, conveyor, transporter", "canopy", "washer, automatic washer, washing machine", "plaything, toy", "pool", "stool", "barrel, cask", "basket, handbasket", "falls", "tent", "bag", "minibike, motorbike", "cradle", "oven", "ball", "food, solid food", "step, stair", "tank, storage tank", "trade name", "microwave", "pot", "animal", "bicycle", "lake", "dishwasher", "screen", "blanket, cover", "sculpture", "hood, exhaust hood", "sconce", "vase", "traffic light", "tray", "trash can", "fan", "pier", "crt screen", "plate", "monitor", "bulletin board", "shower", "radiator", "glass, drinking glass", "clock", "flag", # noqa
230
+ ]
231
+ # After processed by `prepare_ade20k_sem_seg.py`, id 255 means ignore
232
+ # fmt: on
233
+
234
+
235
+ def _get_coco_instances_meta():
236
+ thing_ids = [k["id"] for k in COCO_CATEGORIES if k["isthing"] == 1]
237
+ thing_colors = [k["color"] for k in COCO_CATEGORIES if k["isthing"] == 1]
238
+ assert len(thing_ids) == 80, len(thing_ids)
239
+ # Mapping from the incontiguous COCO category id to an id in [0, 79]
240
+ thing_dataset_id_to_contiguous_id = {k: i for i, k in enumerate(thing_ids)}
241
+ thing_classes = [k["name"] for k in COCO_CATEGORIES if k["isthing"] == 1]
242
+ ret = {
243
+ "thing_dataset_id_to_contiguous_id": thing_dataset_id_to_contiguous_id,
244
+ "thing_classes": thing_classes,
245
+ "thing_colors": thing_colors,
246
+ }
247
+ return ret
248
+
249
+
250
+ def _get_coco_panoptic_separated_meta():
251
+ """
252
+ Returns metadata for "separated" version of the panoptic segmentation dataset.
253
+ """
254
+ stuff_ids = [k["id"] for k in COCO_CATEGORIES if k["isthing"] == 0]
255
+ assert len(stuff_ids) == 53, len(stuff_ids)
256
+
257
+ # For semantic segmentation, this mapping maps from contiguous stuff id
258
+ # (in [0, 53], used in models) to ids in the dataset (used for processing results)
259
+ # The id 0 is mapped to an extra category "thing".
260
+ stuff_dataset_id_to_contiguous_id = {k: i + 1 for i, k in enumerate(stuff_ids)}
261
+ # When converting COCO panoptic annotations to semantic annotations
262
+ # We label the "thing" category to 0
263
+ stuff_dataset_id_to_contiguous_id[0] = 0
264
+
265
+ # 54 names for COCO stuff categories (including "things")
266
+ stuff_classes = ["things"] + [
267
+ k["name"].replace("-other", "").replace("-merged", "")
268
+ for k in COCO_CATEGORIES
269
+ if k["isthing"] == 0
270
+ ]
271
+
272
+ # NOTE: I randomly picked a color for things
273
+ stuff_colors = [[82, 18, 128]] + [k["color"] for k in COCO_CATEGORIES if k["isthing"] == 0]
274
+ ret = {
275
+ "stuff_dataset_id_to_contiguous_id": stuff_dataset_id_to_contiguous_id,
276
+ "stuff_classes": stuff_classes,
277
+ "stuff_colors": stuff_colors,
278
+ }
279
+ ret.update(_get_coco_instances_meta())
280
+ return ret
281
+
282
+
283
+ def _get_builtin_metadata(dataset_name):
284
+ if dataset_name == "coco":
285
+ return _get_coco_instances_meta()
286
+ if dataset_name == "coco_panoptic_separated":
287
+ return _get_coco_panoptic_separated_meta()
288
+ elif dataset_name == "coco_panoptic_standard":
289
+ meta = {}
290
+ # The following metadata maps contiguous id from [0, #thing categories +
291
+ # #stuff categories) to their names and colors. We have to replica of the
292
+ # same name and color under "thing_*" and "stuff_*" because the current
293
+ # visualization function in D2 handles thing and class classes differently
294
+ # due to some heuristic used in Panoptic FPN. We keep the same naming to
295
+ # enable reusing existing visualization functions.
296
+ thing_classes = [k["name"] for k in COCO_CATEGORIES]
297
+ thing_colors = [k["color"] for k in COCO_CATEGORIES]
298
+ stuff_classes = [k["name"] for k in COCO_CATEGORIES]
299
+ stuff_colors = [k["color"] for k in COCO_CATEGORIES]
300
+
301
+ meta["thing_classes"] = thing_classes
302
+ meta["thing_colors"] = thing_colors
303
+ meta["stuff_classes"] = stuff_classes
304
+ meta["stuff_colors"] = stuff_colors
305
+
306
+ # Convert category id for training:
307
+ # category id: like semantic segmentation, it is the class id for each
308
+ # pixel. Since there are some classes not used in evaluation, the category
309
+ # id is not always contiguous and thus we have two set of category ids:
310
+ # - original category id: category id in the original dataset, mainly
311
+ # used for evaluation.
312
+ # - contiguous category id: [0, #classes), in order to train the linear
313
+ # softmax classifier.
314
+ thing_dataset_id_to_contiguous_id = {}
315
+ stuff_dataset_id_to_contiguous_id = {}
316
+
317
+ for i, cat in enumerate(COCO_CATEGORIES):
318
+ if cat["isthing"]:
319
+ thing_dataset_id_to_contiguous_id[cat["id"]] = i
320
+ else:
321
+ stuff_dataset_id_to_contiguous_id[cat["id"]] = i
322
+
323
+ meta["thing_dataset_id_to_contiguous_id"] = thing_dataset_id_to_contiguous_id
324
+ meta["stuff_dataset_id_to_contiguous_id"] = stuff_dataset_id_to_contiguous_id
325
+
326
+ return meta
327
+ elif dataset_name == "coco_person":
328
+ return {
329
+ "thing_classes": ["person"],
330
+ "keypoint_names": COCO_PERSON_KEYPOINT_NAMES,
331
+ "keypoint_flip_map": COCO_PERSON_KEYPOINT_FLIP_MAP,
332
+ "keypoint_connection_rules": KEYPOINT_CONNECTION_RULES,
333
+ }
334
+ elif dataset_name == "cityscapes":
335
+ # fmt: off
336
+ CITYSCAPES_THING_CLASSES = [
337
+ "person", "rider", "car", "truck",
338
+ "bus", "train", "motorcycle", "bicycle",
339
+ ]
340
+ CITYSCAPES_STUFF_CLASSES = [
341
+ "road", "sidewalk", "building", "wall", "fence", "pole", "traffic light",
342
+ "traffic sign", "vegetation", "terrain", "sky", "person", "rider", "car",
343
+ "truck", "bus", "train", "motorcycle", "bicycle",
344
+ ]
345
+ # fmt: on
346
+ return {
347
+ "thing_classes": CITYSCAPES_THING_CLASSES,
348
+ "stuff_classes": CITYSCAPES_STUFF_CLASSES,
349
+ }
350
+ raise KeyError("No built-in metadata for dataset {}".format(dataset_name))
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/cityscapes.py ADDED
@@ -0,0 +1,329 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import functools
3
+ import json
4
+ import logging
5
+ import multiprocessing as mp
6
+ import numpy as np
7
+ import os
8
+ from itertools import chain
9
+ import pycocotools.mask as mask_util
10
+ from PIL import Image
11
+
12
+ from annotator.oneformer.detectron2.structures import BoxMode
13
+ from annotator.oneformer.detectron2.utils.comm import get_world_size
14
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
15
+ from annotator.oneformer.detectron2.utils.logger import setup_logger
16
+
17
+ try:
18
+ import cv2 # noqa
19
+ except ImportError:
20
+ # OpenCV is an optional dependency at the moment
21
+ pass
22
+
23
+
24
+ logger = logging.getLogger(__name__)
25
+
26
+
27
+ def _get_cityscapes_files(image_dir, gt_dir):
28
+ files = []
29
+ # scan through the directory
30
+ cities = PathManager.ls(image_dir)
31
+ logger.info(f"{len(cities)} cities found in '{image_dir}'.")
32
+ for city in cities:
33
+ city_img_dir = os.path.join(image_dir, city)
34
+ city_gt_dir = os.path.join(gt_dir, city)
35
+ for basename in PathManager.ls(city_img_dir):
36
+ image_file = os.path.join(city_img_dir, basename)
37
+
38
+ suffix = "leftImg8bit.png"
39
+ assert basename.endswith(suffix), basename
40
+ basename = basename[: -len(suffix)]
41
+
42
+ instance_file = os.path.join(city_gt_dir, basename + "gtFine_instanceIds.png")
43
+ label_file = os.path.join(city_gt_dir, basename + "gtFine_labelIds.png")
44
+ json_file = os.path.join(city_gt_dir, basename + "gtFine_polygons.json")
45
+
46
+ files.append((image_file, instance_file, label_file, json_file))
47
+ assert len(files), "No images found in {}".format(image_dir)
48
+ for f in files[0]:
49
+ assert PathManager.isfile(f), f
50
+ return files
51
+
52
+
53
+ def load_cityscapes_instances(image_dir, gt_dir, from_json=True, to_polygons=True):
54
+ """
55
+ Args:
56
+ image_dir (str): path to the raw dataset. e.g., "~/cityscapes/leftImg8bit/train".
57
+ gt_dir (str): path to the raw annotations. e.g., "~/cityscapes/gtFine/train".
58
+ from_json (bool): whether to read annotations from the raw json file or the png files.
59
+ to_polygons (bool): whether to represent the segmentation as polygons
60
+ (COCO's format) instead of masks (cityscapes's format).
61
+
62
+ Returns:
63
+ list[dict]: a list of dicts in Detectron2 standard format. (See
64
+ `Using Custom Datasets </tutorials/datasets.html>`_ )
65
+ """
66
+ if from_json:
67
+ assert to_polygons, (
68
+ "Cityscapes's json annotations are in polygon format. "
69
+ "Converting to mask format is not supported now."
70
+ )
71
+ files = _get_cityscapes_files(image_dir, gt_dir)
72
+
73
+ logger.info("Preprocessing cityscapes annotations ...")
74
+ # This is still not fast: all workers will execute duplicate works and will
75
+ # take up to 10m on a 8GPU server.
76
+ pool = mp.Pool(processes=max(mp.cpu_count() // get_world_size() // 2, 4))
77
+
78
+ ret = pool.map(
79
+ functools.partial(_cityscapes_files_to_dict, from_json=from_json, to_polygons=to_polygons),
80
+ files,
81
+ )
82
+ logger.info("Loaded {} images from {}".format(len(ret), image_dir))
83
+
84
+ # Map cityscape ids to contiguous ids
85
+ from cityscapesscripts.helpers.labels import labels
86
+
87
+ labels = [l for l in labels if l.hasInstances and not l.ignoreInEval]
88
+ dataset_id_to_contiguous_id = {l.id: idx for idx, l in enumerate(labels)}
89
+ for dict_per_image in ret:
90
+ for anno in dict_per_image["annotations"]:
91
+ anno["category_id"] = dataset_id_to_contiguous_id[anno["category_id"]]
92
+ return ret
93
+
94
+
95
+ def load_cityscapes_semantic(image_dir, gt_dir):
96
+ """
97
+ Args:
98
+ image_dir (str): path to the raw dataset. e.g., "~/cityscapes/leftImg8bit/train".
99
+ gt_dir (str): path to the raw annotations. e.g., "~/cityscapes/gtFine/train".
100
+
101
+ Returns:
102
+ list[dict]: a list of dict, each has "file_name" and
103
+ "sem_seg_file_name".
104
+ """
105
+ ret = []
106
+ # gt_dir is small and contain many small files. make sense to fetch to local first
107
+ gt_dir = PathManager.get_local_path(gt_dir)
108
+ for image_file, _, label_file, json_file in _get_cityscapes_files(image_dir, gt_dir):
109
+ label_file = label_file.replace("labelIds", "labelTrainIds")
110
+
111
+ with PathManager.open(json_file, "r") as f:
112
+ jsonobj = json.load(f)
113
+ ret.append(
114
+ {
115
+ "file_name": image_file,
116
+ "sem_seg_file_name": label_file,
117
+ "height": jsonobj["imgHeight"],
118
+ "width": jsonobj["imgWidth"],
119
+ }
120
+ )
121
+ assert len(ret), f"No images found in {image_dir}!"
122
+ assert PathManager.isfile(
123
+ ret[0]["sem_seg_file_name"]
124
+ ), "Please generate labelTrainIds.png with cityscapesscripts/preparation/createTrainIdLabelImgs.py" # noqa
125
+ return ret
126
+
127
+
128
+ def _cityscapes_files_to_dict(files, from_json, to_polygons):
129
+ """
130
+ Parse cityscapes annotation files to a instance segmentation dataset dict.
131
+
132
+ Args:
133
+ files (tuple): consists of (image_file, instance_id_file, label_id_file, json_file)
134
+ from_json (bool): whether to read annotations from the raw json file or the png files.
135
+ to_polygons (bool): whether to represent the segmentation as polygons
136
+ (COCO's format) instead of masks (cityscapes's format).
137
+
138
+ Returns:
139
+ A dict in Detectron2 Dataset format.
140
+ """
141
+ from cityscapesscripts.helpers.labels import id2label, name2label
142
+
143
+ image_file, instance_id_file, _, json_file = files
144
+
145
+ annos = []
146
+
147
+ if from_json:
148
+ from shapely.geometry import MultiPolygon, Polygon
149
+
150
+ with PathManager.open(json_file, "r") as f:
151
+ jsonobj = json.load(f)
152
+ ret = {
153
+ "file_name": image_file,
154
+ "image_id": os.path.basename(image_file),
155
+ "height": jsonobj["imgHeight"],
156
+ "width": jsonobj["imgWidth"],
157
+ }
158
+
159
+ # `polygons_union` contains the union of all valid polygons.
160
+ polygons_union = Polygon()
161
+
162
+ # CityscapesScripts draw the polygons in sequential order
163
+ # and each polygon *overwrites* existing ones. See
164
+ # (https://github.com/mcordts/cityscapesScripts/blob/master/cityscapesscripts/preparation/json2instanceImg.py) # noqa
165
+ # We use reverse order, and each polygon *avoids* early ones.
166
+ # This will resolve the ploygon overlaps in the same way as CityscapesScripts.
167
+ for obj in jsonobj["objects"][::-1]:
168
+ if "deleted" in obj: # cityscapes data format specific
169
+ continue
170
+ label_name = obj["label"]
171
+
172
+ try:
173
+ label = name2label[label_name]
174
+ except KeyError:
175
+ if label_name.endswith("group"): # crowd area
176
+ label = name2label[label_name[: -len("group")]]
177
+ else:
178
+ raise
179
+ if label.id < 0: # cityscapes data format
180
+ continue
181
+
182
+ # Cityscapes's raw annotations uses integer coordinates
183
+ # Therefore +0.5 here
184
+ poly_coord = np.asarray(obj["polygon"], dtype="f4") + 0.5
185
+ # CityscapesScript uses PIL.ImageDraw.polygon to rasterize
186
+ # polygons for evaluation. This function operates in integer space
187
+ # and draws each pixel whose center falls into the polygon.
188
+ # Therefore it draws a polygon which is 0.5 "fatter" in expectation.
189
+ # We therefore dilate the input polygon by 0.5 as our input.
190
+ poly = Polygon(poly_coord).buffer(0.5, resolution=4)
191
+
192
+ if not label.hasInstances or label.ignoreInEval:
193
+ # even if we won't store the polygon it still contributes to overlaps resolution
194
+ polygons_union = polygons_union.union(poly)
195
+ continue
196
+
197
+ # Take non-overlapping part of the polygon
198
+ poly_wo_overlaps = poly.difference(polygons_union)
199
+ if poly_wo_overlaps.is_empty:
200
+ continue
201
+ polygons_union = polygons_union.union(poly)
202
+
203
+ anno = {}
204
+ anno["iscrowd"] = label_name.endswith("group")
205
+ anno["category_id"] = label.id
206
+
207
+ if isinstance(poly_wo_overlaps, Polygon):
208
+ poly_list = [poly_wo_overlaps]
209
+ elif isinstance(poly_wo_overlaps, MultiPolygon):
210
+ poly_list = poly_wo_overlaps.geoms
211
+ else:
212
+ raise NotImplementedError("Unknown geometric structure {}".format(poly_wo_overlaps))
213
+
214
+ poly_coord = []
215
+ for poly_el in poly_list:
216
+ # COCO API can work only with exterior boundaries now, hence we store only them.
217
+ # TODO: store both exterior and interior boundaries once other parts of the
218
+ # codebase support holes in polygons.
219
+ poly_coord.append(list(chain(*poly_el.exterior.coords)))
220
+ anno["segmentation"] = poly_coord
221
+ (xmin, ymin, xmax, ymax) = poly_wo_overlaps.bounds
222
+
223
+ anno["bbox"] = (xmin, ymin, xmax, ymax)
224
+ anno["bbox_mode"] = BoxMode.XYXY_ABS
225
+
226
+ annos.append(anno)
227
+ else:
228
+ # See also the official annotation parsing scripts at
229
+ # https://github.com/mcordts/cityscapesScripts/blob/master/cityscapesscripts/evaluation/instances2dict.py # noqa
230
+ with PathManager.open(instance_id_file, "rb") as f:
231
+ inst_image = np.asarray(Image.open(f), order="F")
232
+ # ids < 24 are stuff labels (filtering them first is about 5% faster)
233
+ flattened_ids = np.unique(inst_image[inst_image >= 24])
234
+
235
+ ret = {
236
+ "file_name": image_file,
237
+ "image_id": os.path.basename(image_file),
238
+ "height": inst_image.shape[0],
239
+ "width": inst_image.shape[1],
240
+ }
241
+
242
+ for instance_id in flattened_ids:
243
+ # For non-crowd annotations, instance_id // 1000 is the label_id
244
+ # Crowd annotations have <1000 instance ids
245
+ label_id = instance_id // 1000 if instance_id >= 1000 else instance_id
246
+ label = id2label[label_id]
247
+ if not label.hasInstances or label.ignoreInEval:
248
+ continue
249
+
250
+ anno = {}
251
+ anno["iscrowd"] = instance_id < 1000
252
+ anno["category_id"] = label.id
253
+
254
+ mask = np.asarray(inst_image == instance_id, dtype=np.uint8, order="F")
255
+
256
+ inds = np.nonzero(mask)
257
+ ymin, ymax = inds[0].min(), inds[0].max()
258
+ xmin, xmax = inds[1].min(), inds[1].max()
259
+ anno["bbox"] = (xmin, ymin, xmax, ymax)
260
+ if xmax <= xmin or ymax <= ymin:
261
+ continue
262
+ anno["bbox_mode"] = BoxMode.XYXY_ABS
263
+ if to_polygons:
264
+ # This conversion comes from D4809743 and D5171122,
265
+ # when Mask-RCNN was first developed.
266
+ contours = cv2.findContours(mask.copy(), cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_NONE)[
267
+ -2
268
+ ]
269
+ polygons = [c.reshape(-1).tolist() for c in contours if len(c) >= 3]
270
+ # opencv's can produce invalid polygons
271
+ if len(polygons) == 0:
272
+ continue
273
+ anno["segmentation"] = polygons
274
+ else:
275
+ anno["segmentation"] = mask_util.encode(mask[:, :, None])[0]
276
+ annos.append(anno)
277
+ ret["annotations"] = annos
278
+ return ret
279
+
280
+
281
+ if __name__ == "__main__":
282
+ """
283
+ Test the cityscapes dataset loader.
284
+
285
+ Usage:
286
+ python -m detectron2.data.datasets.cityscapes \
287
+ cityscapes/leftImg8bit/train cityscapes/gtFine/train
288
+ """
289
+ import argparse
290
+
291
+ parser = argparse.ArgumentParser()
292
+ parser.add_argument("image_dir")
293
+ parser.add_argument("gt_dir")
294
+ parser.add_argument("--type", choices=["instance", "semantic"], default="instance")
295
+ args = parser.parse_args()
296
+ from annotator.oneformer.detectron2.data.catalog import Metadata
297
+ from annotator.oneformer.detectron2.utils.visualizer import Visualizer
298
+ from cityscapesscripts.helpers.labels import labels
299
+
300
+ logger = setup_logger(name=__name__)
301
+
302
+ dirname = "cityscapes-data-vis"
303
+ os.makedirs(dirname, exist_ok=True)
304
+
305
+ if args.type == "instance":
306
+ dicts = load_cityscapes_instances(
307
+ args.image_dir, args.gt_dir, from_json=True, to_polygons=True
308
+ )
309
+ logger.info("Done loading {} samples.".format(len(dicts)))
310
+
311
+ thing_classes = [k.name for k in labels if k.hasInstances and not k.ignoreInEval]
312
+ meta = Metadata().set(thing_classes=thing_classes)
313
+
314
+ else:
315
+ dicts = load_cityscapes_semantic(args.image_dir, args.gt_dir)
316
+ logger.info("Done loading {} samples.".format(len(dicts)))
317
+
318
+ stuff_classes = [k.name for k in labels if k.trainId != 255]
319
+ stuff_colors = [k.color for k in labels if k.trainId != 255]
320
+ meta = Metadata().set(stuff_classes=stuff_classes, stuff_colors=stuff_colors)
321
+
322
+ for d in dicts:
323
+ img = np.array(Image.open(PathManager.open(d["file_name"], "rb")))
324
+ visualizer = Visualizer(img, metadata=meta)
325
+ vis = visualizer.draw_dataset_dict(d)
326
+ # cv2.imshow("a", vis.get_image()[:, :, ::-1])
327
+ # cv2.waitKey()
328
+ fpath = os.path.join(dirname, os.path.basename(d["file_name"]))
329
+ vis.save(fpath)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/cityscapes_panoptic.py ADDED
@@ -0,0 +1,187 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import json
3
+ import logging
4
+ import os
5
+
6
+ from annotator.oneformer.detectron2.data import DatasetCatalog, MetadataCatalog
7
+ from annotator.oneformer.detectron2.data.datasets.builtin_meta import CITYSCAPES_CATEGORIES
8
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
9
+
10
+ """
11
+ This file contains functions to register the Cityscapes panoptic dataset to the DatasetCatalog.
12
+ """
13
+
14
+
15
+ logger = logging.getLogger(__name__)
16
+
17
+
18
+ def get_cityscapes_panoptic_files(image_dir, gt_dir, json_info):
19
+ files = []
20
+ # scan through the directory
21
+ cities = PathManager.ls(image_dir)
22
+ logger.info(f"{len(cities)} cities found in '{image_dir}'.")
23
+ image_dict = {}
24
+ for city in cities:
25
+ city_img_dir = os.path.join(image_dir, city)
26
+ for basename in PathManager.ls(city_img_dir):
27
+ image_file = os.path.join(city_img_dir, basename)
28
+
29
+ suffix = "_leftImg8bit.png"
30
+ assert basename.endswith(suffix), basename
31
+ basename = os.path.basename(basename)[: -len(suffix)]
32
+
33
+ image_dict[basename] = image_file
34
+
35
+ for ann in json_info["annotations"]:
36
+ image_file = image_dict.get(ann["image_id"], None)
37
+ assert image_file is not None, "No image {} found for annotation {}".format(
38
+ ann["image_id"], ann["file_name"]
39
+ )
40
+ label_file = os.path.join(gt_dir, ann["file_name"])
41
+ segments_info = ann["segments_info"]
42
+
43
+ files.append((image_file, label_file, segments_info))
44
+
45
+ assert len(files), "No images found in {}".format(image_dir)
46
+ assert PathManager.isfile(files[0][0]), files[0][0]
47
+ assert PathManager.isfile(files[0][1]), files[0][1]
48
+ return files
49
+
50
+
51
+ def load_cityscapes_panoptic(image_dir, gt_dir, gt_json, meta):
52
+ """
53
+ Args:
54
+ image_dir (str): path to the raw dataset. e.g., "~/cityscapes/leftImg8bit/train".
55
+ gt_dir (str): path to the raw annotations. e.g.,
56
+ "~/cityscapes/gtFine/cityscapes_panoptic_train".
57
+ gt_json (str): path to the json file. e.g.,
58
+ "~/cityscapes/gtFine/cityscapes_panoptic_train.json".
59
+ meta (dict): dictionary containing "thing_dataset_id_to_contiguous_id"
60
+ and "stuff_dataset_id_to_contiguous_id" to map category ids to
61
+ contiguous ids for training.
62
+
63
+ Returns:
64
+ list[dict]: a list of dicts in Detectron2 standard format. (See
65
+ `Using Custom Datasets </tutorials/datasets.html>`_ )
66
+ """
67
+
68
+ def _convert_category_id(segment_info, meta):
69
+ if segment_info["category_id"] in meta["thing_dataset_id_to_contiguous_id"]:
70
+ segment_info["category_id"] = meta["thing_dataset_id_to_contiguous_id"][
71
+ segment_info["category_id"]
72
+ ]
73
+ else:
74
+ segment_info["category_id"] = meta["stuff_dataset_id_to_contiguous_id"][
75
+ segment_info["category_id"]
76
+ ]
77
+ return segment_info
78
+
79
+ assert os.path.exists(
80
+ gt_json
81
+ ), "Please run `python cityscapesscripts/preparation/createPanopticImgs.py` to generate label files." # noqa
82
+ with open(gt_json) as f:
83
+ json_info = json.load(f)
84
+ files = get_cityscapes_panoptic_files(image_dir, gt_dir, json_info)
85
+ ret = []
86
+ for image_file, label_file, segments_info in files:
87
+ sem_label_file = (
88
+ image_file.replace("leftImg8bit", "gtFine").split(".")[0] + "_labelTrainIds.png"
89
+ )
90
+ segments_info = [_convert_category_id(x, meta) for x in segments_info]
91
+ ret.append(
92
+ {
93
+ "file_name": image_file,
94
+ "image_id": "_".join(
95
+ os.path.splitext(os.path.basename(image_file))[0].split("_")[:3]
96
+ ),
97
+ "sem_seg_file_name": sem_label_file,
98
+ "pan_seg_file_name": label_file,
99
+ "segments_info": segments_info,
100
+ }
101
+ )
102
+ assert len(ret), f"No images found in {image_dir}!"
103
+ assert PathManager.isfile(
104
+ ret[0]["sem_seg_file_name"]
105
+ ), "Please generate labelTrainIds.png with cityscapesscripts/preparation/createTrainIdLabelImgs.py" # noqa
106
+ assert PathManager.isfile(
107
+ ret[0]["pan_seg_file_name"]
108
+ ), "Please generate panoptic annotation with python cityscapesscripts/preparation/createPanopticImgs.py" # noqa
109
+ return ret
110
+
111
+
112
+ _RAW_CITYSCAPES_PANOPTIC_SPLITS = {
113
+ "cityscapes_fine_panoptic_train": (
114
+ "cityscapes/leftImg8bit/train",
115
+ "cityscapes/gtFine/cityscapes_panoptic_train",
116
+ "cityscapes/gtFine/cityscapes_panoptic_train.json",
117
+ ),
118
+ "cityscapes_fine_panoptic_val": (
119
+ "cityscapes/leftImg8bit/val",
120
+ "cityscapes/gtFine/cityscapes_panoptic_val",
121
+ "cityscapes/gtFine/cityscapes_panoptic_val.json",
122
+ ),
123
+ # "cityscapes_fine_panoptic_test": not supported yet
124
+ }
125
+
126
+
127
+ def register_all_cityscapes_panoptic(root):
128
+ meta = {}
129
+ # The following metadata maps contiguous id from [0, #thing categories +
130
+ # #stuff categories) to their names and colors. We have to replica of the
131
+ # same name and color under "thing_*" and "stuff_*" because the current
132
+ # visualization function in D2 handles thing and class classes differently
133
+ # due to some heuristic used in Panoptic FPN. We keep the same naming to
134
+ # enable reusing existing visualization functions.
135
+ thing_classes = [k["name"] for k in CITYSCAPES_CATEGORIES]
136
+ thing_colors = [k["color"] for k in CITYSCAPES_CATEGORIES]
137
+ stuff_classes = [k["name"] for k in CITYSCAPES_CATEGORIES]
138
+ stuff_colors = [k["color"] for k in CITYSCAPES_CATEGORIES]
139
+
140
+ meta["thing_classes"] = thing_classes
141
+ meta["thing_colors"] = thing_colors
142
+ meta["stuff_classes"] = stuff_classes
143
+ meta["stuff_colors"] = stuff_colors
144
+
145
+ # There are three types of ids in cityscapes panoptic segmentation:
146
+ # (1) category id: like semantic segmentation, it is the class id for each
147
+ # pixel. Since there are some classes not used in evaluation, the category
148
+ # id is not always contiguous and thus we have two set of category ids:
149
+ # - original category id: category id in the original dataset, mainly
150
+ # used for evaluation.
151
+ # - contiguous category id: [0, #classes), in order to train the classifier
152
+ # (2) instance id: this id is used to differentiate different instances from
153
+ # the same category. For "stuff" classes, the instance id is always 0; for
154
+ # "thing" classes, the instance id starts from 1 and 0 is reserved for
155
+ # ignored instances (e.g. crowd annotation).
156
+ # (3) panoptic id: this is the compact id that encode both category and
157
+ # instance id by: category_id * 1000 + instance_id.
158
+ thing_dataset_id_to_contiguous_id = {}
159
+ stuff_dataset_id_to_contiguous_id = {}
160
+
161
+ for k in CITYSCAPES_CATEGORIES:
162
+ if k["isthing"] == 1:
163
+ thing_dataset_id_to_contiguous_id[k["id"]] = k["trainId"]
164
+ else:
165
+ stuff_dataset_id_to_contiguous_id[k["id"]] = k["trainId"]
166
+
167
+ meta["thing_dataset_id_to_contiguous_id"] = thing_dataset_id_to_contiguous_id
168
+ meta["stuff_dataset_id_to_contiguous_id"] = stuff_dataset_id_to_contiguous_id
169
+
170
+ for key, (image_dir, gt_dir, gt_json) in _RAW_CITYSCAPES_PANOPTIC_SPLITS.items():
171
+ image_dir = os.path.join(root, image_dir)
172
+ gt_dir = os.path.join(root, gt_dir)
173
+ gt_json = os.path.join(root, gt_json)
174
+
175
+ DatasetCatalog.register(
176
+ key, lambda x=image_dir, y=gt_dir, z=gt_json: load_cityscapes_panoptic(x, y, z, meta)
177
+ )
178
+ MetadataCatalog.get(key).set(
179
+ panoptic_root=gt_dir,
180
+ image_root=image_dir,
181
+ panoptic_json=gt_json,
182
+ gt_dir=gt_dir.replace("cityscapes_panoptic_", ""),
183
+ evaluator_type="cityscapes_panoptic_seg",
184
+ ignore_label=255,
185
+ label_divisor=1000,
186
+ **meta,
187
+ )
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/coco.py ADDED
@@ -0,0 +1,539 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import contextlib
3
+ import datetime
4
+ import io
5
+ import json
6
+ import logging
7
+ import numpy as np
8
+ import os
9
+ import shutil
10
+ import pycocotools.mask as mask_util
11
+ from fvcore.common.timer import Timer
12
+ from iopath.common.file_io import file_lock
13
+ from PIL import Image
14
+
15
+ from annotator.oneformer.detectron2.structures import Boxes, BoxMode, PolygonMasks, RotatedBoxes
16
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
17
+
18
+ from .. import DatasetCatalog, MetadataCatalog
19
+
20
+ """
21
+ This file contains functions to parse COCO-format annotations into dicts in "Detectron2 format".
22
+ """
23
+
24
+
25
+ logger = logging.getLogger(__name__)
26
+
27
+ __all__ = ["load_coco_json", "load_sem_seg", "convert_to_coco_json", "register_coco_instances"]
28
+
29
+
30
+ def load_coco_json(json_file, image_root, dataset_name=None, extra_annotation_keys=None):
31
+ """
32
+ Load a json file with COCO's instances annotation format.
33
+ Currently supports instance detection, instance segmentation,
34
+ and person keypoints annotations.
35
+
36
+ Args:
37
+ json_file (str): full path to the json file in COCO instances annotation format.
38
+ image_root (str or path-like): the directory where the images in this json file exists.
39
+ dataset_name (str or None): the name of the dataset (e.g., coco_2017_train).
40
+ When provided, this function will also do the following:
41
+
42
+ * Put "thing_classes" into the metadata associated with this dataset.
43
+ * Map the category ids into a contiguous range (needed by standard dataset format),
44
+ and add "thing_dataset_id_to_contiguous_id" to the metadata associated
45
+ with this dataset.
46
+
47
+ This option should usually be provided, unless users need to load
48
+ the original json content and apply more processing manually.
49
+ extra_annotation_keys (list[str]): list of per-annotation keys that should also be
50
+ loaded into the dataset dict (besides "iscrowd", "bbox", "keypoints",
51
+ "category_id", "segmentation"). The values for these keys will be returned as-is.
52
+ For example, the densepose annotations are loaded in this way.
53
+
54
+ Returns:
55
+ list[dict]: a list of dicts in Detectron2 standard dataset dicts format (See
56
+ `Using Custom Datasets </tutorials/datasets.html>`_ ) when `dataset_name` is not None.
57
+ If `dataset_name` is None, the returned `category_ids` may be
58
+ incontiguous and may not conform to the Detectron2 standard format.
59
+
60
+ Notes:
61
+ 1. This function does not read the image files.
62
+ The results do not have the "image" field.
63
+ """
64
+ from pycocotools.coco import COCO
65
+
66
+ timer = Timer()
67
+ json_file = PathManager.get_local_path(json_file)
68
+ with contextlib.redirect_stdout(io.StringIO()):
69
+ coco_api = COCO(json_file)
70
+ if timer.seconds() > 1:
71
+ logger.info("Loading {} takes {:.2f} seconds.".format(json_file, timer.seconds()))
72
+
73
+ id_map = None
74
+ if dataset_name is not None:
75
+ meta = MetadataCatalog.get(dataset_name)
76
+ cat_ids = sorted(coco_api.getCatIds())
77
+ cats = coco_api.loadCats(cat_ids)
78
+ # The categories in a custom json file may not be sorted.
79
+ thing_classes = [c["name"] for c in sorted(cats, key=lambda x: x["id"])]
80
+ meta.thing_classes = thing_classes
81
+
82
+ # In COCO, certain category ids are artificially removed,
83
+ # and by convention they are always ignored.
84
+ # We deal with COCO's id issue and translate
85
+ # the category ids to contiguous ids in [0, 80).
86
+
87
+ # It works by looking at the "categories" field in the json, therefore
88
+ # if users' own json also have incontiguous ids, we'll
89
+ # apply this mapping as well but print a warning.
90
+ if not (min(cat_ids) == 1 and max(cat_ids) == len(cat_ids)):
91
+ if "coco" not in dataset_name:
92
+ logger.warning(
93
+ """
94
+ Category ids in annotations are not in [1, #categories]! We'll apply a mapping for you.
95
+ """
96
+ )
97
+ id_map = {v: i for i, v in enumerate(cat_ids)}
98
+ meta.thing_dataset_id_to_contiguous_id = id_map
99
+
100
+ # sort indices for reproducible results
101
+ img_ids = sorted(coco_api.imgs.keys())
102
+ # imgs is a list of dicts, each looks something like:
103
+ # {'license': 4,
104
+ # 'url': 'http://farm6.staticflickr.com/5454/9413846304_881d5e5c3b_z.jpg',
105
+ # 'file_name': 'COCO_val2014_000000001268.jpg',
106
+ # 'height': 427,
107
+ # 'width': 640,
108
+ # 'date_captured': '2013-11-17 05:57:24',
109
+ # 'id': 1268}
110
+ imgs = coco_api.loadImgs(img_ids)
111
+ # anns is a list[list[dict]], where each dict is an annotation
112
+ # record for an object. The inner list enumerates the objects in an image
113
+ # and the outer list enumerates over images. Example of anns[0]:
114
+ # [{'segmentation': [[192.81,
115
+ # 247.09,
116
+ # ...
117
+ # 219.03,
118
+ # 249.06]],
119
+ # 'area': 1035.749,
120
+ # 'iscrowd': 0,
121
+ # 'image_id': 1268,
122
+ # 'bbox': [192.81, 224.8, 74.73, 33.43],
123
+ # 'category_id': 16,
124
+ # 'id': 42986},
125
+ # ...]
126
+ anns = [coco_api.imgToAnns[img_id] for img_id in img_ids]
127
+ total_num_valid_anns = sum([len(x) for x in anns])
128
+ total_num_anns = len(coco_api.anns)
129
+ if total_num_valid_anns < total_num_anns:
130
+ logger.warning(
131
+ f"{json_file} contains {total_num_anns} annotations, but only "
132
+ f"{total_num_valid_anns} of them match to images in the file."
133
+ )
134
+
135
+ if "minival" not in json_file:
136
+ # The popular valminusminival & minival annotations for COCO2014 contain this bug.
137
+ # However the ratio of buggy annotations there is tiny and does not affect accuracy.
138
+ # Therefore we explicitly white-list them.
139
+ ann_ids = [ann["id"] for anns_per_image in anns for ann in anns_per_image]
140
+ assert len(set(ann_ids)) == len(ann_ids), "Annotation ids in '{}' are not unique!".format(
141
+ json_file
142
+ )
143
+
144
+ imgs_anns = list(zip(imgs, anns))
145
+ logger.info("Loaded {} images in COCO format from {}".format(len(imgs_anns), json_file))
146
+
147
+ dataset_dicts = []
148
+
149
+ ann_keys = ["iscrowd", "bbox", "keypoints", "category_id"] + (extra_annotation_keys or [])
150
+
151
+ num_instances_without_valid_segmentation = 0
152
+
153
+ for (img_dict, anno_dict_list) in imgs_anns:
154
+ record = {}
155
+ record["file_name"] = os.path.join(image_root, img_dict["file_name"])
156
+ record["height"] = img_dict["height"]
157
+ record["width"] = img_dict["width"]
158
+ image_id = record["image_id"] = img_dict["id"]
159
+
160
+ objs = []
161
+ for anno in anno_dict_list:
162
+ # Check that the image_id in this annotation is the same as
163
+ # the image_id we're looking at.
164
+ # This fails only when the data parsing logic or the annotation file is buggy.
165
+
166
+ # The original COCO valminusminival2014 & minival2014 annotation files
167
+ # actually contains bugs that, together with certain ways of using COCO API,
168
+ # can trigger this assertion.
169
+ assert anno["image_id"] == image_id
170
+
171
+ assert anno.get("ignore", 0) == 0, '"ignore" in COCO json file is not supported.'
172
+
173
+ obj = {key: anno[key] for key in ann_keys if key in anno}
174
+ if "bbox" in obj and len(obj["bbox"]) == 0:
175
+ raise ValueError(
176
+ f"One annotation of image {image_id} contains empty 'bbox' value! "
177
+ "This json does not have valid COCO format."
178
+ )
179
+
180
+ segm = anno.get("segmentation", None)
181
+ if segm: # either list[list[float]] or dict(RLE)
182
+ if isinstance(segm, dict):
183
+ if isinstance(segm["counts"], list):
184
+ # convert to compressed RLE
185
+ segm = mask_util.frPyObjects(segm, *segm["size"])
186
+ else:
187
+ # filter out invalid polygons (< 3 points)
188
+ segm = [poly for poly in segm if len(poly) % 2 == 0 and len(poly) >= 6]
189
+ if len(segm) == 0:
190
+ num_instances_without_valid_segmentation += 1
191
+ continue # ignore this instance
192
+ obj["segmentation"] = segm
193
+
194
+ keypts = anno.get("keypoints", None)
195
+ if keypts: # list[int]
196
+ for idx, v in enumerate(keypts):
197
+ if idx % 3 != 2:
198
+ # COCO's segmentation coordinates are floating points in [0, H or W],
199
+ # but keypoint coordinates are integers in [0, H-1 or W-1]
200
+ # Therefore we assume the coordinates are "pixel indices" and
201
+ # add 0.5 to convert to floating point coordinates.
202
+ keypts[idx] = v + 0.5
203
+ obj["keypoints"] = keypts
204
+
205
+ obj["bbox_mode"] = BoxMode.XYWH_ABS
206
+ if id_map:
207
+ annotation_category_id = obj["category_id"]
208
+ try:
209
+ obj["category_id"] = id_map[annotation_category_id]
210
+ except KeyError as e:
211
+ raise KeyError(
212
+ f"Encountered category_id={annotation_category_id} "
213
+ "but this id does not exist in 'categories' of the json file."
214
+ ) from e
215
+ objs.append(obj)
216
+ record["annotations"] = objs
217
+ dataset_dicts.append(record)
218
+
219
+ if num_instances_without_valid_segmentation > 0:
220
+ logger.warning(
221
+ "Filtered out {} instances without valid segmentation. ".format(
222
+ num_instances_without_valid_segmentation
223
+ )
224
+ + "There might be issues in your dataset generation process. Please "
225
+ "check https://detectron2.readthedocs.io/en/latest/tutorials/datasets.html carefully"
226
+ )
227
+ return dataset_dicts
228
+
229
+
230
+ def load_sem_seg(gt_root, image_root, gt_ext="png", image_ext="jpg"):
231
+ """
232
+ Load semantic segmentation datasets. All files under "gt_root" with "gt_ext" extension are
233
+ treated as ground truth annotations and all files under "image_root" with "image_ext" extension
234
+ as input images. Ground truth and input images are matched using file paths relative to
235
+ "gt_root" and "image_root" respectively without taking into account file extensions.
236
+ This works for COCO as well as some other datasets.
237
+
238
+ Args:
239
+ gt_root (str): full path to ground truth semantic segmentation files. Semantic segmentation
240
+ annotations are stored as images with integer values in pixels that represent
241
+ corresponding semantic labels.
242
+ image_root (str): the directory where the input images are.
243
+ gt_ext (str): file extension for ground truth annotations.
244
+ image_ext (str): file extension for input images.
245
+
246
+ Returns:
247
+ list[dict]:
248
+ a list of dicts in detectron2 standard format without instance-level
249
+ annotation.
250
+
251
+ Notes:
252
+ 1. This function does not read the image and ground truth files.
253
+ The results do not have the "image" and "sem_seg" fields.
254
+ """
255
+
256
+ # We match input images with ground truth based on their relative filepaths (without file
257
+ # extensions) starting from 'image_root' and 'gt_root' respectively.
258
+ def file2id(folder_path, file_path):
259
+ # extract relative path starting from `folder_path`
260
+ image_id = os.path.normpath(os.path.relpath(file_path, start=folder_path))
261
+ # remove file extension
262
+ image_id = os.path.splitext(image_id)[0]
263
+ return image_id
264
+
265
+ input_files = sorted(
266
+ (os.path.join(image_root, f) for f in PathManager.ls(image_root) if f.endswith(image_ext)),
267
+ key=lambda file_path: file2id(image_root, file_path),
268
+ )
269
+ gt_files = sorted(
270
+ (os.path.join(gt_root, f) for f in PathManager.ls(gt_root) if f.endswith(gt_ext)),
271
+ key=lambda file_path: file2id(gt_root, file_path),
272
+ )
273
+
274
+ assert len(gt_files) > 0, "No annotations found in {}.".format(gt_root)
275
+
276
+ # Use the intersection, so that val2017_100 annotations can run smoothly with val2017 images
277
+ if len(input_files) != len(gt_files):
278
+ logger.warn(
279
+ "Directory {} and {} has {} and {} files, respectively.".format(
280
+ image_root, gt_root, len(input_files), len(gt_files)
281
+ )
282
+ )
283
+ input_basenames = [os.path.basename(f)[: -len(image_ext)] for f in input_files]
284
+ gt_basenames = [os.path.basename(f)[: -len(gt_ext)] for f in gt_files]
285
+ intersect = list(set(input_basenames) & set(gt_basenames))
286
+ # sort, otherwise each worker may obtain a list[dict] in different order
287
+ intersect = sorted(intersect)
288
+ logger.warn("Will use their intersection of {} files.".format(len(intersect)))
289
+ input_files = [os.path.join(image_root, f + image_ext) for f in intersect]
290
+ gt_files = [os.path.join(gt_root, f + gt_ext) for f in intersect]
291
+
292
+ logger.info(
293
+ "Loaded {} images with semantic segmentation from {}".format(len(input_files), image_root)
294
+ )
295
+
296
+ dataset_dicts = []
297
+ for (img_path, gt_path) in zip(input_files, gt_files):
298
+ record = {}
299
+ record["file_name"] = img_path
300
+ record["sem_seg_file_name"] = gt_path
301
+ dataset_dicts.append(record)
302
+
303
+ return dataset_dicts
304
+
305
+
306
+ def convert_to_coco_dict(dataset_name):
307
+ """
308
+ Convert an instance detection/segmentation or keypoint detection dataset
309
+ in detectron2's standard format into COCO json format.
310
+
311
+ Generic dataset description can be found here:
312
+ https://detectron2.readthedocs.io/tutorials/datasets.html#register-a-dataset
313
+
314
+ COCO data format description can be found here:
315
+ http://cocodataset.org/#format-data
316
+
317
+ Args:
318
+ dataset_name (str):
319
+ name of the source dataset
320
+ Must be registered in DatastCatalog and in detectron2's standard format.
321
+ Must have corresponding metadata "thing_classes"
322
+ Returns:
323
+ coco_dict: serializable dict in COCO json format
324
+ """
325
+
326
+ dataset_dicts = DatasetCatalog.get(dataset_name)
327
+ metadata = MetadataCatalog.get(dataset_name)
328
+
329
+ # unmap the category mapping ids for COCO
330
+ if hasattr(metadata, "thing_dataset_id_to_contiguous_id"):
331
+ reverse_id_mapping = {v: k for k, v in metadata.thing_dataset_id_to_contiguous_id.items()}
332
+ reverse_id_mapper = lambda contiguous_id: reverse_id_mapping[contiguous_id] # noqa
333
+ else:
334
+ reverse_id_mapper = lambda contiguous_id: contiguous_id # noqa
335
+
336
+ categories = [
337
+ {"id": reverse_id_mapper(id), "name": name}
338
+ for id, name in enumerate(metadata.thing_classes)
339
+ ]
340
+
341
+ logger.info("Converting dataset dicts into COCO format")
342
+ coco_images = []
343
+ coco_annotations = []
344
+
345
+ for image_id, image_dict in enumerate(dataset_dicts):
346
+ coco_image = {
347
+ "id": image_dict.get("image_id", image_id),
348
+ "width": int(image_dict["width"]),
349
+ "height": int(image_dict["height"]),
350
+ "file_name": str(image_dict["file_name"]),
351
+ }
352
+ coco_images.append(coco_image)
353
+
354
+ anns_per_image = image_dict.get("annotations", [])
355
+ for annotation in anns_per_image:
356
+ # create a new dict with only COCO fields
357
+ coco_annotation = {}
358
+
359
+ # COCO requirement: XYWH box format for axis-align and XYWHA for rotated
360
+ bbox = annotation["bbox"]
361
+ if isinstance(bbox, np.ndarray):
362
+ if bbox.ndim != 1:
363
+ raise ValueError(f"bbox has to be 1-dimensional. Got shape={bbox.shape}.")
364
+ bbox = bbox.tolist()
365
+ if len(bbox) not in [4, 5]:
366
+ raise ValueError(f"bbox has to has length 4 or 5. Got {bbox}.")
367
+ from_bbox_mode = annotation["bbox_mode"]
368
+ to_bbox_mode = BoxMode.XYWH_ABS if len(bbox) == 4 else BoxMode.XYWHA_ABS
369
+ bbox = BoxMode.convert(bbox, from_bbox_mode, to_bbox_mode)
370
+
371
+ # COCO requirement: instance area
372
+ if "segmentation" in annotation:
373
+ # Computing areas for instances by counting the pixels
374
+ segmentation = annotation["segmentation"]
375
+ # TODO: check segmentation type: RLE, BinaryMask or Polygon
376
+ if isinstance(segmentation, list):
377
+ polygons = PolygonMasks([segmentation])
378
+ area = polygons.area()[0].item()
379
+ elif isinstance(segmentation, dict): # RLE
380
+ area = mask_util.area(segmentation).item()
381
+ else:
382
+ raise TypeError(f"Unknown segmentation type {type(segmentation)}!")
383
+ else:
384
+ # Computing areas using bounding boxes
385
+ if to_bbox_mode == BoxMode.XYWH_ABS:
386
+ bbox_xy = BoxMode.convert(bbox, to_bbox_mode, BoxMode.XYXY_ABS)
387
+ area = Boxes([bbox_xy]).area()[0].item()
388
+ else:
389
+ area = RotatedBoxes([bbox]).area()[0].item()
390
+
391
+ if "keypoints" in annotation:
392
+ keypoints = annotation["keypoints"] # list[int]
393
+ for idx, v in enumerate(keypoints):
394
+ if idx % 3 != 2:
395
+ # COCO's segmentation coordinates are floating points in [0, H or W],
396
+ # but keypoint coordinates are integers in [0, H-1 or W-1]
397
+ # For COCO format consistency we substract 0.5
398
+ # https://github.com/facebookresearch/detectron2/pull/175#issuecomment-551202163
399
+ keypoints[idx] = v - 0.5
400
+ if "num_keypoints" in annotation:
401
+ num_keypoints = annotation["num_keypoints"]
402
+ else:
403
+ num_keypoints = sum(kp > 0 for kp in keypoints[2::3])
404
+
405
+ # COCO requirement:
406
+ # linking annotations to images
407
+ # "id" field must start with 1
408
+ coco_annotation["id"] = len(coco_annotations) + 1
409
+ coco_annotation["image_id"] = coco_image["id"]
410
+ coco_annotation["bbox"] = [round(float(x), 3) for x in bbox]
411
+ coco_annotation["area"] = float(area)
412
+ coco_annotation["iscrowd"] = int(annotation.get("iscrowd", 0))
413
+ coco_annotation["category_id"] = int(reverse_id_mapper(annotation["category_id"]))
414
+
415
+ # Add optional fields
416
+ if "keypoints" in annotation:
417
+ coco_annotation["keypoints"] = keypoints
418
+ coco_annotation["num_keypoints"] = num_keypoints
419
+
420
+ if "segmentation" in annotation:
421
+ seg = coco_annotation["segmentation"] = annotation["segmentation"]
422
+ if isinstance(seg, dict): # RLE
423
+ counts = seg["counts"]
424
+ if not isinstance(counts, str):
425
+ # make it json-serializable
426
+ seg["counts"] = counts.decode("ascii")
427
+
428
+ coco_annotations.append(coco_annotation)
429
+
430
+ logger.info(
431
+ "Conversion finished, "
432
+ f"#images: {len(coco_images)}, #annotations: {len(coco_annotations)}"
433
+ )
434
+
435
+ info = {
436
+ "date_created": str(datetime.datetime.now()),
437
+ "description": "Automatically generated COCO json file for Detectron2.",
438
+ }
439
+ coco_dict = {"info": info, "images": coco_images, "categories": categories, "licenses": None}
440
+ if len(coco_annotations) > 0:
441
+ coco_dict["annotations"] = coco_annotations
442
+ return coco_dict
443
+
444
+
445
+ def convert_to_coco_json(dataset_name, output_file, allow_cached=True):
446
+ """
447
+ Converts dataset into COCO format and saves it to a json file.
448
+ dataset_name must be registered in DatasetCatalog and in detectron2's standard format.
449
+
450
+ Args:
451
+ dataset_name:
452
+ reference from the config file to the catalogs
453
+ must be registered in DatasetCatalog and in detectron2's standard format
454
+ output_file: path of json file that will be saved to
455
+ allow_cached: if json file is already present then skip conversion
456
+ """
457
+
458
+ # TODO: The dataset or the conversion script *may* change,
459
+ # a checksum would be useful for validating the cached data
460
+
461
+ PathManager.mkdirs(os.path.dirname(output_file))
462
+ with file_lock(output_file):
463
+ if PathManager.exists(output_file) and allow_cached:
464
+ logger.warning(
465
+ f"Using previously cached COCO format annotations at '{output_file}'. "
466
+ "You need to clear the cache file if your dataset has been modified."
467
+ )
468
+ else:
469
+ logger.info(f"Converting annotations of dataset '{dataset_name}' to COCO format ...)")
470
+ coco_dict = convert_to_coco_dict(dataset_name)
471
+
472
+ logger.info(f"Caching COCO format annotations at '{output_file}' ...")
473
+ tmp_file = output_file + ".tmp"
474
+ with PathManager.open(tmp_file, "w") as f:
475
+ json.dump(coco_dict, f)
476
+ shutil.move(tmp_file, output_file)
477
+
478
+
479
+ def register_coco_instances(name, metadata, json_file, image_root):
480
+ """
481
+ Register a dataset in COCO's json annotation format for
482
+ instance detection, instance segmentation and keypoint detection.
483
+ (i.e., Type 1 and 2 in http://cocodataset.org/#format-data.
484
+ `instances*.json` and `person_keypoints*.json` in the dataset).
485
+
486
+ This is an example of how to register a new dataset.
487
+ You can do something similar to this function, to register new datasets.
488
+
489
+ Args:
490
+ name (str): the name that identifies a dataset, e.g. "coco_2014_train".
491
+ metadata (dict): extra metadata associated with this dataset. You can
492
+ leave it as an empty dict.
493
+ json_file (str): path to the json instance annotation file.
494
+ image_root (str or path-like): directory which contains all the images.
495
+ """
496
+ assert isinstance(name, str), name
497
+ assert isinstance(json_file, (str, os.PathLike)), json_file
498
+ assert isinstance(image_root, (str, os.PathLike)), image_root
499
+ # 1. register a function which returns dicts
500
+ DatasetCatalog.register(name, lambda: load_coco_json(json_file, image_root, name))
501
+
502
+ # 2. Optionally, add metadata about this dataset,
503
+ # since they might be useful in evaluation, visualization or logging
504
+ MetadataCatalog.get(name).set(
505
+ json_file=json_file, image_root=image_root, evaluator_type="coco", **metadata
506
+ )
507
+
508
+
509
+ if __name__ == "__main__":
510
+ """
511
+ Test the COCO json dataset loader.
512
+
513
+ Usage:
514
+ python -m detectron2.data.datasets.coco \
515
+ path/to/json path/to/image_root dataset_name
516
+
517
+ "dataset_name" can be "coco_2014_minival_100", or other
518
+ pre-registered ones
519
+ """
520
+ from annotator.oneformer.detectron2.utils.logger import setup_logger
521
+ from annotator.oneformer.detectron2.utils.visualizer import Visualizer
522
+ import annotator.oneformer.detectron2.data.datasets # noqa # add pre-defined metadata
523
+ import sys
524
+
525
+ logger = setup_logger(name=__name__)
526
+ assert sys.argv[3] in DatasetCatalog.list()
527
+ meta = MetadataCatalog.get(sys.argv[3])
528
+
529
+ dicts = load_coco_json(sys.argv[1], sys.argv[2], sys.argv[3])
530
+ logger.info("Done loading {} samples.".format(len(dicts)))
531
+
532
+ dirname = "coco-data-vis"
533
+ os.makedirs(dirname, exist_ok=True)
534
+ for d in dicts:
535
+ img = np.array(Image.open(d["file_name"]))
536
+ visualizer = Visualizer(img, metadata=meta)
537
+ vis = visualizer.draw_dataset_dict(d)
538
+ fpath = os.path.join(dirname, os.path.basename(d["file_name"]))
539
+ vis.save(fpath)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/coco_panoptic.py ADDED
@@ -0,0 +1,228 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import copy
3
+ import json
4
+ import os
5
+
6
+ from annotator.oneformer.detectron2.data import DatasetCatalog, MetadataCatalog
7
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
8
+
9
+ from .coco import load_coco_json, load_sem_seg
10
+
11
+ __all__ = ["register_coco_panoptic", "register_coco_panoptic_separated"]
12
+
13
+
14
+ def load_coco_panoptic_json(json_file, image_dir, gt_dir, meta):
15
+ """
16
+ Args:
17
+ image_dir (str): path to the raw dataset. e.g., "~/coco/train2017".
18
+ gt_dir (str): path to the raw annotations. e.g., "~/coco/panoptic_train2017".
19
+ json_file (str): path to the json file. e.g., "~/coco/annotations/panoptic_train2017.json".
20
+
21
+ Returns:
22
+ list[dict]: a list of dicts in Detectron2 standard format. (See
23
+ `Using Custom Datasets </tutorials/datasets.html>`_ )
24
+ """
25
+
26
+ def _convert_category_id(segment_info, meta):
27
+ if segment_info["category_id"] in meta["thing_dataset_id_to_contiguous_id"]:
28
+ segment_info["category_id"] = meta["thing_dataset_id_to_contiguous_id"][
29
+ segment_info["category_id"]
30
+ ]
31
+ segment_info["isthing"] = True
32
+ else:
33
+ segment_info["category_id"] = meta["stuff_dataset_id_to_contiguous_id"][
34
+ segment_info["category_id"]
35
+ ]
36
+ segment_info["isthing"] = False
37
+ return segment_info
38
+
39
+ with PathManager.open(json_file) as f:
40
+ json_info = json.load(f)
41
+
42
+ ret = []
43
+ for ann in json_info["annotations"]:
44
+ image_id = int(ann["image_id"])
45
+ # TODO: currently we assume image and label has the same filename but
46
+ # different extension, and images have extension ".jpg" for COCO. Need
47
+ # to make image extension a user-provided argument if we extend this
48
+ # function to support other COCO-like datasets.
49
+ image_file = os.path.join(image_dir, os.path.splitext(ann["file_name"])[0] + ".jpg")
50
+ label_file = os.path.join(gt_dir, ann["file_name"])
51
+ segments_info = [_convert_category_id(x, meta) for x in ann["segments_info"]]
52
+ ret.append(
53
+ {
54
+ "file_name": image_file,
55
+ "image_id": image_id,
56
+ "pan_seg_file_name": label_file,
57
+ "segments_info": segments_info,
58
+ }
59
+ )
60
+ assert len(ret), f"No images found in {image_dir}!"
61
+ assert PathManager.isfile(ret[0]["file_name"]), ret[0]["file_name"]
62
+ assert PathManager.isfile(ret[0]["pan_seg_file_name"]), ret[0]["pan_seg_file_name"]
63
+ return ret
64
+
65
+
66
+ def register_coco_panoptic(
67
+ name, metadata, image_root, panoptic_root, panoptic_json, instances_json=None
68
+ ):
69
+ """
70
+ Register a "standard" version of COCO panoptic segmentation dataset named `name`.
71
+ The dictionaries in this registered dataset follows detectron2's standard format.
72
+ Hence it's called "standard".
73
+
74
+ Args:
75
+ name (str): the name that identifies a dataset,
76
+ e.g. "coco_2017_train_panoptic"
77
+ metadata (dict): extra metadata associated with this dataset.
78
+ image_root (str): directory which contains all the images
79
+ panoptic_root (str): directory which contains panoptic annotation images in COCO format
80
+ panoptic_json (str): path to the json panoptic annotation file in COCO format
81
+ sem_seg_root (none): not used, to be consistent with
82
+ `register_coco_panoptic_separated`.
83
+ instances_json (str): path to the json instance annotation file
84
+ """
85
+ panoptic_name = name
86
+ DatasetCatalog.register(
87
+ panoptic_name,
88
+ lambda: load_coco_panoptic_json(panoptic_json, image_root, panoptic_root, metadata),
89
+ )
90
+ MetadataCatalog.get(panoptic_name).set(
91
+ panoptic_root=panoptic_root,
92
+ image_root=image_root,
93
+ panoptic_json=panoptic_json,
94
+ json_file=instances_json,
95
+ evaluator_type="coco_panoptic_seg",
96
+ ignore_label=255,
97
+ label_divisor=1000,
98
+ **metadata,
99
+ )
100
+
101
+
102
+ def register_coco_panoptic_separated(
103
+ name, metadata, image_root, panoptic_root, panoptic_json, sem_seg_root, instances_json
104
+ ):
105
+ """
106
+ Register a "separated" version of COCO panoptic segmentation dataset named `name`.
107
+ The annotations in this registered dataset will contain both instance annotations and
108
+ semantic annotations, each with its own contiguous ids. Hence it's called "separated".
109
+
110
+ It follows the setting used by the PanopticFPN paper:
111
+
112
+ 1. The instance annotations directly come from polygons in the COCO
113
+ instances annotation task, rather than from the masks in the COCO panoptic annotations.
114
+
115
+ The two format have small differences:
116
+ Polygons in the instance annotations may have overlaps.
117
+ The mask annotations are produced by labeling the overlapped polygons
118
+ with depth ordering.
119
+
120
+ 2. The semantic annotations are converted from panoptic annotations, where
121
+ all "things" are assigned a semantic id of 0.
122
+ All semantic categories will therefore have ids in contiguous
123
+ range [1, #stuff_categories].
124
+
125
+ This function will also register a pure semantic segmentation dataset
126
+ named ``name + '_stuffonly'``.
127
+
128
+ Args:
129
+ name (str): the name that identifies a dataset,
130
+ e.g. "coco_2017_train_panoptic"
131
+ metadata (dict): extra metadata associated with this dataset.
132
+ image_root (str): directory which contains all the images
133
+ panoptic_root (str): directory which contains panoptic annotation images
134
+ panoptic_json (str): path to the json panoptic annotation file
135
+ sem_seg_root (str): directory which contains all the ground truth segmentation annotations.
136
+ instances_json (str): path to the json instance annotation file
137
+ """
138
+ panoptic_name = name + "_separated"
139
+ DatasetCatalog.register(
140
+ panoptic_name,
141
+ lambda: merge_to_panoptic(
142
+ load_coco_json(instances_json, image_root, panoptic_name),
143
+ load_sem_seg(sem_seg_root, image_root),
144
+ ),
145
+ )
146
+ MetadataCatalog.get(panoptic_name).set(
147
+ panoptic_root=panoptic_root,
148
+ image_root=image_root,
149
+ panoptic_json=panoptic_json,
150
+ sem_seg_root=sem_seg_root,
151
+ json_file=instances_json, # TODO rename
152
+ evaluator_type="coco_panoptic_seg",
153
+ ignore_label=255,
154
+ **metadata,
155
+ )
156
+
157
+ semantic_name = name + "_stuffonly"
158
+ DatasetCatalog.register(semantic_name, lambda: load_sem_seg(sem_seg_root, image_root))
159
+ MetadataCatalog.get(semantic_name).set(
160
+ sem_seg_root=sem_seg_root,
161
+ image_root=image_root,
162
+ evaluator_type="sem_seg",
163
+ ignore_label=255,
164
+ **metadata,
165
+ )
166
+
167
+
168
+ def merge_to_panoptic(detection_dicts, sem_seg_dicts):
169
+ """
170
+ Create dataset dicts for panoptic segmentation, by
171
+ merging two dicts using "file_name" field to match their entries.
172
+
173
+ Args:
174
+ detection_dicts (list[dict]): lists of dicts for object detection or instance segmentation.
175
+ sem_seg_dicts (list[dict]): lists of dicts for semantic segmentation.
176
+
177
+ Returns:
178
+ list[dict] (one per input image): Each dict contains all (key, value) pairs from dicts in
179
+ both detection_dicts and sem_seg_dicts that correspond to the same image.
180
+ The function assumes that the same key in different dicts has the same value.
181
+ """
182
+ results = []
183
+ sem_seg_file_to_entry = {x["file_name"]: x for x in sem_seg_dicts}
184
+ assert len(sem_seg_file_to_entry) > 0
185
+
186
+ for det_dict in detection_dicts:
187
+ dic = copy.copy(det_dict)
188
+ dic.update(sem_seg_file_to_entry[dic["file_name"]])
189
+ results.append(dic)
190
+ return results
191
+
192
+
193
+ if __name__ == "__main__":
194
+ """
195
+ Test the COCO panoptic dataset loader.
196
+
197
+ Usage:
198
+ python -m detectron2.data.datasets.coco_panoptic \
199
+ path/to/image_root path/to/panoptic_root path/to/panoptic_json dataset_name 10
200
+
201
+ "dataset_name" can be "coco_2017_train_panoptic", or other
202
+ pre-registered ones
203
+ """
204
+ from annotator.oneformer.detectron2.utils.logger import setup_logger
205
+ from annotator.oneformer.detectron2.utils.visualizer import Visualizer
206
+ import annotator.oneformer.detectron2.data.datasets # noqa # add pre-defined metadata
207
+ import sys
208
+ from PIL import Image
209
+ import numpy as np
210
+
211
+ logger = setup_logger(name=__name__)
212
+ assert sys.argv[4] in DatasetCatalog.list()
213
+ meta = MetadataCatalog.get(sys.argv[4])
214
+
215
+ dicts = load_coco_panoptic_json(sys.argv[3], sys.argv[1], sys.argv[2], meta.as_dict())
216
+ logger.info("Done loading {} samples.".format(len(dicts)))
217
+
218
+ dirname = "coco-data-vis"
219
+ os.makedirs(dirname, exist_ok=True)
220
+ num_imgs_to_vis = int(sys.argv[5])
221
+ for i, d in enumerate(dicts):
222
+ img = np.array(Image.open(d["file_name"]))
223
+ visualizer = Visualizer(img, metadata=meta)
224
+ vis = visualizer.draw_dataset_dict(d)
225
+ fpath = os.path.join(dirname, os.path.basename(d["file_name"]))
226
+ vis.save(fpath)
227
+ if i + 1 >= num_imgs_to_vis:
228
+ break
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis.py ADDED
@@ -0,0 +1,241 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import logging
3
+ import os
4
+ from fvcore.common.timer import Timer
5
+
6
+ from annotator.oneformer.detectron2.data import DatasetCatalog, MetadataCatalog
7
+ from annotator.oneformer.detectron2.structures import BoxMode
8
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
9
+
10
+ from .builtin_meta import _get_coco_instances_meta
11
+ from .lvis_v0_5_categories import LVIS_CATEGORIES as LVIS_V0_5_CATEGORIES
12
+ from .lvis_v1_categories import LVIS_CATEGORIES as LVIS_V1_CATEGORIES
13
+ from .lvis_v1_category_image_count import LVIS_CATEGORY_IMAGE_COUNT as LVIS_V1_CATEGORY_IMAGE_COUNT
14
+
15
+ """
16
+ This file contains functions to parse LVIS-format annotations into dicts in the
17
+ "Detectron2 format".
18
+ """
19
+
20
+ logger = logging.getLogger(__name__)
21
+
22
+ __all__ = ["load_lvis_json", "register_lvis_instances", "get_lvis_instances_meta"]
23
+
24
+
25
+ def register_lvis_instances(name, metadata, json_file, image_root):
26
+ """
27
+ Register a dataset in LVIS's json annotation format for instance detection and segmentation.
28
+
29
+ Args:
30
+ name (str): a name that identifies the dataset, e.g. "lvis_v0.5_train".
31
+ metadata (dict): extra metadata associated with this dataset. It can be an empty dict.
32
+ json_file (str): path to the json instance annotation file.
33
+ image_root (str or path-like): directory which contains all the images.
34
+ """
35
+ DatasetCatalog.register(name, lambda: load_lvis_json(json_file, image_root, name))
36
+ MetadataCatalog.get(name).set(
37
+ json_file=json_file, image_root=image_root, evaluator_type="lvis", **metadata
38
+ )
39
+
40
+
41
+ def load_lvis_json(json_file, image_root, dataset_name=None, extra_annotation_keys=None):
42
+ """
43
+ Load a json file in LVIS's annotation format.
44
+
45
+ Args:
46
+ json_file (str): full path to the LVIS json annotation file.
47
+ image_root (str): the directory where the images in this json file exists.
48
+ dataset_name (str): the name of the dataset (e.g., "lvis_v0.5_train").
49
+ If provided, this function will put "thing_classes" into the metadata
50
+ associated with this dataset.
51
+ extra_annotation_keys (list[str]): list of per-annotation keys that should also be
52
+ loaded into the dataset dict (besides "bbox", "bbox_mode", "category_id",
53
+ "segmentation"). The values for these keys will be returned as-is.
54
+
55
+ Returns:
56
+ list[dict]: a list of dicts in Detectron2 standard format. (See
57
+ `Using Custom Datasets </tutorials/datasets.html>`_ )
58
+
59
+ Notes:
60
+ 1. This function does not read the image files.
61
+ The results do not have the "image" field.
62
+ """
63
+ from lvis import LVIS
64
+
65
+ json_file = PathManager.get_local_path(json_file)
66
+
67
+ timer = Timer()
68
+ lvis_api = LVIS(json_file)
69
+ if timer.seconds() > 1:
70
+ logger.info("Loading {} takes {:.2f} seconds.".format(json_file, timer.seconds()))
71
+
72
+ if dataset_name is not None:
73
+ meta = get_lvis_instances_meta(dataset_name)
74
+ MetadataCatalog.get(dataset_name).set(**meta)
75
+
76
+ # sort indices for reproducible results
77
+ img_ids = sorted(lvis_api.imgs.keys())
78
+ # imgs is a list of dicts, each looks something like:
79
+ # {'license': 4,
80
+ # 'url': 'http://farm6.staticflickr.com/5454/9413846304_881d5e5c3b_z.jpg',
81
+ # 'file_name': 'COCO_val2014_000000001268.jpg',
82
+ # 'height': 427,
83
+ # 'width': 640,
84
+ # 'date_captured': '2013-11-17 05:57:24',
85
+ # 'id': 1268}
86
+ imgs = lvis_api.load_imgs(img_ids)
87
+ # anns is a list[list[dict]], where each dict is an annotation
88
+ # record for an object. The inner list enumerates the objects in an image
89
+ # and the outer list enumerates over images. Example of anns[0]:
90
+ # [{'segmentation': [[192.81,
91
+ # 247.09,
92
+ # ...
93
+ # 219.03,
94
+ # 249.06]],
95
+ # 'area': 1035.749,
96
+ # 'image_id': 1268,
97
+ # 'bbox': [192.81, 224.8, 74.73, 33.43],
98
+ # 'category_id': 16,
99
+ # 'id': 42986},
100
+ # ...]
101
+ anns = [lvis_api.img_ann_map[img_id] for img_id in img_ids]
102
+
103
+ # Sanity check that each annotation has a unique id
104
+ ann_ids = [ann["id"] for anns_per_image in anns for ann in anns_per_image]
105
+ assert len(set(ann_ids)) == len(ann_ids), "Annotation ids in '{}' are not unique".format(
106
+ json_file
107
+ )
108
+
109
+ imgs_anns = list(zip(imgs, anns))
110
+
111
+ logger.info("Loaded {} images in the LVIS format from {}".format(len(imgs_anns), json_file))
112
+
113
+ if extra_annotation_keys:
114
+ logger.info(
115
+ "The following extra annotation keys will be loaded: {} ".format(extra_annotation_keys)
116
+ )
117
+ else:
118
+ extra_annotation_keys = []
119
+
120
+ def get_file_name(img_root, img_dict):
121
+ # Determine the path including the split folder ("train2017", "val2017", "test2017") from
122
+ # the coco_url field. Example:
123
+ # 'coco_url': 'http://images.cocodataset.org/train2017/000000155379.jpg'
124
+ split_folder, file_name = img_dict["coco_url"].split("/")[-2:]
125
+ return os.path.join(img_root + split_folder, file_name)
126
+
127
+ dataset_dicts = []
128
+
129
+ for (img_dict, anno_dict_list) in imgs_anns:
130
+ record = {}
131
+ record["file_name"] = get_file_name(image_root, img_dict)
132
+ record["height"] = img_dict["height"]
133
+ record["width"] = img_dict["width"]
134
+ record["not_exhaustive_category_ids"] = img_dict.get("not_exhaustive_category_ids", [])
135
+ record["neg_category_ids"] = img_dict.get("neg_category_ids", [])
136
+ image_id = record["image_id"] = img_dict["id"]
137
+
138
+ objs = []
139
+ for anno in anno_dict_list:
140
+ # Check that the image_id in this annotation is the same as
141
+ # the image_id we're looking at.
142
+ # This fails only when the data parsing logic or the annotation file is buggy.
143
+ assert anno["image_id"] == image_id
144
+ obj = {"bbox": anno["bbox"], "bbox_mode": BoxMode.XYWH_ABS}
145
+ # LVIS data loader can be used to load COCO dataset categories. In this case `meta`
146
+ # variable will have a field with COCO-specific category mapping.
147
+ if dataset_name is not None and "thing_dataset_id_to_contiguous_id" in meta:
148
+ obj["category_id"] = meta["thing_dataset_id_to_contiguous_id"][anno["category_id"]]
149
+ else:
150
+ obj["category_id"] = anno["category_id"] - 1 # Convert 1-indexed to 0-indexed
151
+ segm = anno["segmentation"] # list[list[float]]
152
+ # filter out invalid polygons (< 3 points)
153
+ valid_segm = [poly for poly in segm if len(poly) % 2 == 0 and len(poly) >= 6]
154
+ assert len(segm) == len(
155
+ valid_segm
156
+ ), "Annotation contains an invalid polygon with < 3 points"
157
+ assert len(segm) > 0
158
+ obj["segmentation"] = segm
159
+ for extra_ann_key in extra_annotation_keys:
160
+ obj[extra_ann_key] = anno[extra_ann_key]
161
+ objs.append(obj)
162
+ record["annotations"] = objs
163
+ dataset_dicts.append(record)
164
+
165
+ return dataset_dicts
166
+
167
+
168
+ def get_lvis_instances_meta(dataset_name):
169
+ """
170
+ Load LVIS metadata.
171
+
172
+ Args:
173
+ dataset_name (str): LVIS dataset name without the split name (e.g., "lvis_v0.5").
174
+
175
+ Returns:
176
+ dict: LVIS metadata with keys: thing_classes
177
+ """
178
+ if "cocofied" in dataset_name:
179
+ return _get_coco_instances_meta()
180
+ if "v0.5" in dataset_name:
181
+ return _get_lvis_instances_meta_v0_5()
182
+ elif "v1" in dataset_name:
183
+ return _get_lvis_instances_meta_v1()
184
+ raise ValueError("No built-in metadata for dataset {}".format(dataset_name))
185
+
186
+
187
+ def _get_lvis_instances_meta_v0_5():
188
+ assert len(LVIS_V0_5_CATEGORIES) == 1230
189
+ cat_ids = [k["id"] for k in LVIS_V0_5_CATEGORIES]
190
+ assert min(cat_ids) == 1 and max(cat_ids) == len(
191
+ cat_ids
192
+ ), "Category ids are not in [1, #categories], as expected"
193
+ # Ensure that the category list is sorted by id
194
+ lvis_categories = sorted(LVIS_V0_5_CATEGORIES, key=lambda x: x["id"])
195
+ thing_classes = [k["synonyms"][0] for k in lvis_categories]
196
+ meta = {"thing_classes": thing_classes}
197
+ return meta
198
+
199
+
200
+ def _get_lvis_instances_meta_v1():
201
+ assert len(LVIS_V1_CATEGORIES) == 1203
202
+ cat_ids = [k["id"] for k in LVIS_V1_CATEGORIES]
203
+ assert min(cat_ids) == 1 and max(cat_ids) == len(
204
+ cat_ids
205
+ ), "Category ids are not in [1, #categories], as expected"
206
+ # Ensure that the category list is sorted by id
207
+ lvis_categories = sorted(LVIS_V1_CATEGORIES, key=lambda x: x["id"])
208
+ thing_classes = [k["synonyms"][0] for k in lvis_categories]
209
+ meta = {"thing_classes": thing_classes, "class_image_count": LVIS_V1_CATEGORY_IMAGE_COUNT}
210
+ return meta
211
+
212
+
213
+ if __name__ == "__main__":
214
+ """
215
+ Test the LVIS json dataset loader.
216
+
217
+ Usage:
218
+ python -m detectron2.data.datasets.lvis \
219
+ path/to/json path/to/image_root dataset_name vis_limit
220
+ """
221
+ import sys
222
+ import numpy as np
223
+ from annotator.oneformer.detectron2.utils.logger import setup_logger
224
+ from PIL import Image
225
+ import annotator.oneformer.detectron2.data.datasets # noqa # add pre-defined metadata
226
+ from annotator.oneformer.detectron2.utils.visualizer import Visualizer
227
+
228
+ logger = setup_logger(name=__name__)
229
+ meta = MetadataCatalog.get(sys.argv[3])
230
+
231
+ dicts = load_lvis_json(sys.argv[1], sys.argv[2], sys.argv[3])
232
+ logger.info("Done loading {} samples.".format(len(dicts)))
233
+
234
+ dirname = "lvis-data-vis"
235
+ os.makedirs(dirname, exist_ok=True)
236
+ for d in dicts[: int(sys.argv[4])]:
237
+ img = np.array(Image.open(d["file_name"]))
238
+ visualizer = Visualizer(img, metadata=meta)
239
+ vis = visualizer.draw_dataset_dict(d)
240
+ fpath = os.path.join(dirname, os.path.basename(d["file_name"]))
241
+ vis.save(fpath)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis_v0_5_categories.py ADDED
The diff for this file is too large to render. See raw diff
 
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis_v1_categories.py ADDED
The diff for this file is too large to render. See raw diff
 
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/lvis_v1_category_image_count.py ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ # Autogen with
3
+ # with open("lvis_v1_train.json", "r") as f:
4
+ # a = json.load(f)
5
+ # c = a["categories"]
6
+ # for x in c:
7
+ # del x["name"]
8
+ # del x["instance_count"]
9
+ # del x["def"]
10
+ # del x["synonyms"]
11
+ # del x["frequency"]
12
+ # del x["synset"]
13
+ # LVIS_CATEGORY_IMAGE_COUNT = repr(c) + " # noqa"
14
+ # with open("/tmp/lvis_category_image_count.py", "wt") as f:
15
+ # f.write(f"LVIS_CATEGORY_IMAGE_COUNT = {LVIS_CATEGORY_IMAGE_COUNT}")
16
+ # Then paste the contents of that file below
17
+
18
+ # fmt: off
19
+ LVIS_CATEGORY_IMAGE_COUNT = [{'id': 1, 'image_count': 64}, {'id': 2, 'image_count': 364}, {'id': 3, 'image_count': 1911}, {'id': 4, 'image_count': 149}, {'id': 5, 'image_count': 29}, {'id': 6, 'image_count': 26}, {'id': 7, 'image_count': 59}, {'id': 8, 'image_count': 22}, {'id': 9, 'image_count': 12}, {'id': 10, 'image_count': 28}, {'id': 11, 'image_count': 505}, {'id': 12, 'image_count': 1207}, {'id': 13, 'image_count': 4}, {'id': 14, 'image_count': 10}, {'id': 15, 'image_count': 500}, {'id': 16, 'image_count': 33}, {'id': 17, 'image_count': 3}, {'id': 18, 'image_count': 44}, {'id': 19, 'image_count': 561}, {'id': 20, 'image_count': 8}, {'id': 21, 'image_count': 9}, {'id': 22, 'image_count': 33}, {'id': 23, 'image_count': 1883}, {'id': 24, 'image_count': 98}, {'id': 25, 'image_count': 70}, {'id': 26, 'image_count': 46}, {'id': 27, 'image_count': 117}, {'id': 28, 'image_count': 41}, {'id': 29, 'image_count': 1395}, {'id': 30, 'image_count': 7}, {'id': 31, 'image_count': 1}, {'id': 32, 'image_count': 314}, {'id': 33, 'image_count': 31}, {'id': 34, 'image_count': 1905}, {'id': 35, 'image_count': 1859}, {'id': 36, 'image_count': 1623}, {'id': 37, 'image_count': 47}, {'id': 38, 'image_count': 3}, {'id': 39, 'image_count': 3}, {'id': 40, 'image_count': 1}, {'id': 41, 'image_count': 305}, {'id': 42, 'image_count': 6}, {'id': 43, 'image_count': 210}, {'id': 44, 'image_count': 36}, {'id': 45, 'image_count': 1787}, {'id': 46, 'image_count': 17}, {'id': 47, 'image_count': 51}, {'id': 48, 'image_count': 138}, {'id': 49, 'image_count': 3}, {'id': 50, 'image_count': 1470}, {'id': 51, 'image_count': 3}, {'id': 52, 'image_count': 2}, {'id': 53, 'image_count': 186}, {'id': 54, 'image_count': 76}, {'id': 55, 'image_count': 26}, {'id': 56, 'image_count': 303}, {'id': 57, 'image_count': 738}, {'id': 58, 'image_count': 1799}, {'id': 59, 'image_count': 1934}, {'id': 60, 'image_count': 1609}, {'id': 61, 'image_count': 1622}, {'id': 62, 'image_count': 41}, {'id': 63, 'image_count': 4}, {'id': 64, 'image_count': 11}, {'id': 65, 'image_count': 270}, {'id': 66, 'image_count': 349}, {'id': 67, 'image_count': 42}, {'id': 68, 'image_count': 823}, {'id': 69, 'image_count': 6}, {'id': 70, 'image_count': 48}, {'id': 71, 'image_count': 3}, {'id': 72, 'image_count': 42}, {'id': 73, 'image_count': 24}, {'id': 74, 'image_count': 16}, {'id': 75, 'image_count': 605}, {'id': 76, 'image_count': 646}, {'id': 77, 'image_count': 1765}, {'id': 78, 'image_count': 2}, {'id': 79, 'image_count': 125}, {'id': 80, 'image_count': 1420}, {'id': 81, 'image_count': 140}, {'id': 82, 'image_count': 4}, {'id': 83, 'image_count': 322}, {'id': 84, 'image_count': 60}, {'id': 85, 'image_count': 2}, {'id': 86, 'image_count': 231}, {'id': 87, 'image_count': 333}, {'id': 88, 'image_count': 1941}, {'id': 89, 'image_count': 367}, {'id': 90, 'image_count': 1922}, {'id': 91, 'image_count': 18}, {'id': 92, 'image_count': 81}, {'id': 93, 'image_count': 1}, {'id': 94, 'image_count': 1852}, {'id': 95, 'image_count': 430}, {'id': 96, 'image_count': 247}, {'id': 97, 'image_count': 94}, {'id': 98, 'image_count': 21}, {'id': 99, 'image_count': 1821}, {'id': 100, 'image_count': 16}, {'id': 101, 'image_count': 12}, {'id': 102, 'image_count': 25}, {'id': 103, 'image_count': 41}, {'id': 104, 'image_count': 244}, {'id': 105, 'image_count': 7}, {'id': 106, 'image_count': 1}, {'id': 107, 'image_count': 40}, {'id': 108, 'image_count': 40}, {'id': 109, 'image_count': 104}, {'id': 110, 'image_count': 1671}, {'id': 111, 'image_count': 49}, {'id': 112, 'image_count': 243}, {'id': 113, 'image_count': 2}, {'id': 114, 'image_count': 242}, {'id': 115, 'image_count': 271}, {'id': 116, 'image_count': 104}, {'id': 117, 'image_count': 8}, {'id': 118, 'image_count': 1758}, {'id': 119, 'image_count': 1}, {'id': 120, 'image_count': 48}, {'id': 121, 'image_count': 14}, {'id': 122, 'image_count': 40}, {'id': 123, 'image_count': 1}, {'id': 124, 'image_count': 37}, {'id': 125, 'image_count': 1510}, {'id': 126, 'image_count': 6}, {'id': 127, 'image_count': 1903}, {'id': 128, 'image_count': 70}, {'id': 129, 'image_count': 86}, {'id': 130, 'image_count': 7}, {'id': 131, 'image_count': 5}, {'id': 132, 'image_count': 1406}, {'id': 133, 'image_count': 1901}, {'id': 134, 'image_count': 15}, {'id': 135, 'image_count': 28}, {'id': 136, 'image_count': 6}, {'id': 137, 'image_count': 494}, {'id': 138, 'image_count': 234}, {'id': 139, 'image_count': 1922}, {'id': 140, 'image_count': 1}, {'id': 141, 'image_count': 35}, {'id': 142, 'image_count': 5}, {'id': 143, 'image_count': 1828}, {'id': 144, 'image_count': 8}, {'id': 145, 'image_count': 63}, {'id': 146, 'image_count': 1668}, {'id': 147, 'image_count': 4}, {'id': 148, 'image_count': 95}, {'id': 149, 'image_count': 17}, {'id': 150, 'image_count': 1567}, {'id': 151, 'image_count': 2}, {'id': 152, 'image_count': 103}, {'id': 153, 'image_count': 50}, {'id': 154, 'image_count': 1309}, {'id': 155, 'image_count': 6}, {'id': 156, 'image_count': 92}, {'id': 157, 'image_count': 19}, {'id': 158, 'image_count': 37}, {'id': 159, 'image_count': 4}, {'id': 160, 'image_count': 709}, {'id': 161, 'image_count': 9}, {'id': 162, 'image_count': 82}, {'id': 163, 'image_count': 15}, {'id': 164, 'image_count': 3}, {'id': 165, 'image_count': 61}, {'id': 166, 'image_count': 51}, {'id': 167, 'image_count': 5}, {'id': 168, 'image_count': 13}, {'id': 169, 'image_count': 642}, {'id': 170, 'image_count': 24}, {'id': 171, 'image_count': 255}, {'id': 172, 'image_count': 9}, {'id': 173, 'image_count': 1808}, {'id': 174, 'image_count': 31}, {'id': 175, 'image_count': 158}, {'id': 176, 'image_count': 80}, {'id': 177, 'image_count': 1884}, {'id': 178, 'image_count': 158}, {'id': 179, 'image_count': 2}, {'id': 180, 'image_count': 12}, {'id': 181, 'image_count': 1659}, {'id': 182, 'image_count': 7}, {'id': 183, 'image_count': 834}, {'id': 184, 'image_count': 57}, {'id': 185, 'image_count': 174}, {'id': 186, 'image_count': 95}, {'id': 187, 'image_count': 27}, {'id': 188, 'image_count': 22}, {'id': 189, 'image_count': 1391}, {'id': 190, 'image_count': 90}, {'id': 191, 'image_count': 40}, {'id': 192, 'image_count': 445}, {'id': 193, 'image_count': 21}, {'id': 194, 'image_count': 1132}, {'id': 195, 'image_count': 177}, {'id': 196, 'image_count': 4}, {'id': 197, 'image_count': 17}, {'id': 198, 'image_count': 84}, {'id': 199, 'image_count': 55}, {'id': 200, 'image_count': 30}, {'id': 201, 'image_count': 25}, {'id': 202, 'image_count': 2}, {'id': 203, 'image_count': 125}, {'id': 204, 'image_count': 1135}, {'id': 205, 'image_count': 19}, {'id': 206, 'image_count': 72}, {'id': 207, 'image_count': 1926}, {'id': 208, 'image_count': 159}, {'id': 209, 'image_count': 7}, {'id': 210, 'image_count': 1}, {'id': 211, 'image_count': 13}, {'id': 212, 'image_count': 35}, {'id': 213, 'image_count': 18}, {'id': 214, 'image_count': 8}, {'id': 215, 'image_count': 6}, {'id': 216, 'image_count': 35}, {'id': 217, 'image_count': 1222}, {'id': 218, 'image_count': 103}, {'id': 219, 'image_count': 28}, {'id': 220, 'image_count': 63}, {'id': 221, 'image_count': 28}, {'id': 222, 'image_count': 5}, {'id': 223, 'image_count': 7}, {'id': 224, 'image_count': 14}, {'id': 225, 'image_count': 1918}, {'id': 226, 'image_count': 133}, {'id': 227, 'image_count': 16}, {'id': 228, 'image_count': 27}, {'id': 229, 'image_count': 110}, {'id': 230, 'image_count': 1895}, {'id': 231, 'image_count': 4}, {'id': 232, 'image_count': 1927}, {'id': 233, 'image_count': 8}, {'id': 234, 'image_count': 1}, {'id': 235, 'image_count': 263}, {'id': 236, 'image_count': 10}, {'id': 237, 'image_count': 2}, {'id': 238, 'image_count': 3}, {'id': 239, 'image_count': 87}, {'id': 240, 'image_count': 9}, {'id': 241, 'image_count': 71}, {'id': 242, 'image_count': 13}, {'id': 243, 'image_count': 18}, {'id': 244, 'image_count': 2}, {'id': 245, 'image_count': 5}, {'id': 246, 'image_count': 45}, {'id': 247, 'image_count': 1}, {'id': 248, 'image_count': 23}, {'id': 249, 'image_count': 32}, {'id': 250, 'image_count': 4}, {'id': 251, 'image_count': 1}, {'id': 252, 'image_count': 858}, {'id': 253, 'image_count': 661}, {'id': 254, 'image_count': 168}, {'id': 255, 'image_count': 210}, {'id': 256, 'image_count': 65}, {'id': 257, 'image_count': 4}, {'id': 258, 'image_count': 2}, {'id': 259, 'image_count': 159}, {'id': 260, 'image_count': 31}, {'id': 261, 'image_count': 811}, {'id': 262, 'image_count': 1}, {'id': 263, 'image_count': 42}, {'id': 264, 'image_count': 27}, {'id': 265, 'image_count': 2}, {'id': 266, 'image_count': 5}, {'id': 267, 'image_count': 95}, {'id': 268, 'image_count': 32}, {'id': 269, 'image_count': 1}, {'id': 270, 'image_count': 1}, {'id': 271, 'image_count': 1844}, {'id': 272, 'image_count': 897}, {'id': 273, 'image_count': 31}, {'id': 274, 'image_count': 23}, {'id': 275, 'image_count': 1}, {'id': 276, 'image_count': 202}, {'id': 277, 'image_count': 746}, {'id': 278, 'image_count': 44}, {'id': 279, 'image_count': 14}, {'id': 280, 'image_count': 26}, {'id': 281, 'image_count': 1}, {'id': 282, 'image_count': 2}, {'id': 283, 'image_count': 25}, {'id': 284, 'image_count': 238}, {'id': 285, 'image_count': 592}, {'id': 286, 'image_count': 26}, {'id': 287, 'image_count': 5}, {'id': 288, 'image_count': 42}, {'id': 289, 'image_count': 13}, {'id': 290, 'image_count': 46}, {'id': 291, 'image_count': 1}, {'id': 292, 'image_count': 8}, {'id': 293, 'image_count': 34}, {'id': 294, 'image_count': 5}, {'id': 295, 'image_count': 1}, {'id': 296, 'image_count': 1871}, {'id': 297, 'image_count': 717}, {'id': 298, 'image_count': 1010}, {'id': 299, 'image_count': 679}, {'id': 300, 'image_count': 3}, {'id': 301, 'image_count': 4}, {'id': 302, 'image_count': 1}, {'id': 303, 'image_count': 166}, {'id': 304, 'image_count': 2}, {'id': 305, 'image_count': 266}, {'id': 306, 'image_count': 101}, {'id': 307, 'image_count': 6}, {'id': 308, 'image_count': 14}, {'id': 309, 'image_count': 133}, {'id': 310, 'image_count': 2}, {'id': 311, 'image_count': 38}, {'id': 312, 'image_count': 95}, {'id': 313, 'image_count': 1}, {'id': 314, 'image_count': 12}, {'id': 315, 'image_count': 49}, {'id': 316, 'image_count': 5}, {'id': 317, 'image_count': 5}, {'id': 318, 'image_count': 16}, {'id': 319, 'image_count': 216}, {'id': 320, 'image_count': 12}, {'id': 321, 'image_count': 1}, {'id': 322, 'image_count': 54}, {'id': 323, 'image_count': 5}, {'id': 324, 'image_count': 245}, {'id': 325, 'image_count': 12}, {'id': 326, 'image_count': 7}, {'id': 327, 'image_count': 35}, {'id': 328, 'image_count': 36}, {'id': 329, 'image_count': 32}, {'id': 330, 'image_count': 1027}, {'id': 331, 'image_count': 10}, {'id': 332, 'image_count': 12}, {'id': 333, 'image_count': 1}, {'id': 334, 'image_count': 67}, {'id': 335, 'image_count': 71}, {'id': 336, 'image_count': 30}, {'id': 337, 'image_count': 48}, {'id': 338, 'image_count': 249}, {'id': 339, 'image_count': 13}, {'id': 340, 'image_count': 29}, {'id': 341, 'image_count': 14}, {'id': 342, 'image_count': 236}, {'id': 343, 'image_count': 15}, {'id': 344, 'image_count': 1521}, {'id': 345, 'image_count': 25}, {'id': 346, 'image_count': 249}, {'id': 347, 'image_count': 139}, {'id': 348, 'image_count': 2}, {'id': 349, 'image_count': 2}, {'id': 350, 'image_count': 1890}, {'id': 351, 'image_count': 1240}, {'id': 352, 'image_count': 1}, {'id': 353, 'image_count': 9}, {'id': 354, 'image_count': 1}, {'id': 355, 'image_count': 3}, {'id': 356, 'image_count': 11}, {'id': 357, 'image_count': 4}, {'id': 358, 'image_count': 236}, {'id': 359, 'image_count': 44}, {'id': 360, 'image_count': 19}, {'id': 361, 'image_count': 1100}, {'id': 362, 'image_count': 7}, {'id': 363, 'image_count': 69}, {'id': 364, 'image_count': 2}, {'id': 365, 'image_count': 8}, {'id': 366, 'image_count': 5}, {'id': 367, 'image_count': 227}, {'id': 368, 'image_count': 6}, {'id': 369, 'image_count': 106}, {'id': 370, 'image_count': 81}, {'id': 371, 'image_count': 17}, {'id': 372, 'image_count': 134}, {'id': 373, 'image_count': 312}, {'id': 374, 'image_count': 8}, {'id': 375, 'image_count': 271}, {'id': 376, 'image_count': 2}, {'id': 377, 'image_count': 103}, {'id': 378, 'image_count': 1938}, {'id': 379, 'image_count': 574}, {'id': 380, 'image_count': 120}, {'id': 381, 'image_count': 2}, {'id': 382, 'image_count': 2}, {'id': 383, 'image_count': 13}, {'id': 384, 'image_count': 29}, {'id': 385, 'image_count': 1710}, {'id': 386, 'image_count': 66}, {'id': 387, 'image_count': 1008}, {'id': 388, 'image_count': 1}, {'id': 389, 'image_count': 3}, {'id': 390, 'image_count': 1942}, {'id': 391, 'image_count': 19}, {'id': 392, 'image_count': 1488}, {'id': 393, 'image_count': 46}, {'id': 394, 'image_count': 106}, {'id': 395, 'image_count': 115}, {'id': 396, 'image_count': 19}, {'id': 397, 'image_count': 2}, {'id': 398, 'image_count': 1}, {'id': 399, 'image_count': 28}, {'id': 400, 'image_count': 9}, {'id': 401, 'image_count': 192}, {'id': 402, 'image_count': 12}, {'id': 403, 'image_count': 21}, {'id': 404, 'image_count': 247}, {'id': 405, 'image_count': 6}, {'id': 406, 'image_count': 64}, {'id': 407, 'image_count': 7}, {'id': 408, 'image_count': 40}, {'id': 409, 'image_count': 542}, {'id': 410, 'image_count': 2}, {'id': 411, 'image_count': 1898}, {'id': 412, 'image_count': 36}, {'id': 413, 'image_count': 4}, {'id': 414, 'image_count': 1}, {'id': 415, 'image_count': 191}, {'id': 416, 'image_count': 6}, {'id': 417, 'image_count': 41}, {'id': 418, 'image_count': 39}, {'id': 419, 'image_count': 46}, {'id': 420, 'image_count': 1}, {'id': 421, 'image_count': 1451}, {'id': 422, 'image_count': 1878}, {'id': 423, 'image_count': 11}, {'id': 424, 'image_count': 82}, {'id': 425, 'image_count': 18}, {'id': 426, 'image_count': 1}, {'id': 427, 'image_count': 7}, {'id': 428, 'image_count': 3}, {'id': 429, 'image_count': 575}, {'id': 430, 'image_count': 1907}, {'id': 431, 'image_count': 8}, {'id': 432, 'image_count': 4}, {'id': 433, 'image_count': 32}, {'id': 434, 'image_count': 11}, {'id': 435, 'image_count': 4}, {'id': 436, 'image_count': 54}, {'id': 437, 'image_count': 202}, {'id': 438, 'image_count': 32}, {'id': 439, 'image_count': 3}, {'id': 440, 'image_count': 130}, {'id': 441, 'image_count': 119}, {'id': 442, 'image_count': 141}, {'id': 443, 'image_count': 29}, {'id': 444, 'image_count': 525}, {'id': 445, 'image_count': 1323}, {'id': 446, 'image_count': 2}, {'id': 447, 'image_count': 113}, {'id': 448, 'image_count': 16}, {'id': 449, 'image_count': 7}, {'id': 450, 'image_count': 35}, {'id': 451, 'image_count': 1908}, {'id': 452, 'image_count': 353}, {'id': 453, 'image_count': 18}, {'id': 454, 'image_count': 14}, {'id': 455, 'image_count': 77}, {'id': 456, 'image_count': 8}, {'id': 457, 'image_count': 37}, {'id': 458, 'image_count': 1}, {'id': 459, 'image_count': 346}, {'id': 460, 'image_count': 19}, {'id': 461, 'image_count': 1779}, {'id': 462, 'image_count': 23}, {'id': 463, 'image_count': 25}, {'id': 464, 'image_count': 67}, {'id': 465, 'image_count': 19}, {'id': 466, 'image_count': 28}, {'id': 467, 'image_count': 4}, {'id': 468, 'image_count': 27}, {'id': 469, 'image_count': 1861}, {'id': 470, 'image_count': 11}, {'id': 471, 'image_count': 13}, {'id': 472, 'image_count': 13}, {'id': 473, 'image_count': 32}, {'id': 474, 'image_count': 1767}, {'id': 475, 'image_count': 42}, {'id': 476, 'image_count': 17}, {'id': 477, 'image_count': 128}, {'id': 478, 'image_count': 1}, {'id': 479, 'image_count': 9}, {'id': 480, 'image_count': 10}, {'id': 481, 'image_count': 4}, {'id': 482, 'image_count': 9}, {'id': 483, 'image_count': 18}, {'id': 484, 'image_count': 41}, {'id': 485, 'image_count': 28}, {'id': 486, 'image_count': 3}, {'id': 487, 'image_count': 65}, {'id': 488, 'image_count': 9}, {'id': 489, 'image_count': 23}, {'id': 490, 'image_count': 24}, {'id': 491, 'image_count': 1}, {'id': 492, 'image_count': 2}, {'id': 493, 'image_count': 59}, {'id': 494, 'image_count': 48}, {'id': 495, 'image_count': 17}, {'id': 496, 'image_count': 1877}, {'id': 497, 'image_count': 18}, {'id': 498, 'image_count': 1920}, {'id': 499, 'image_count': 50}, {'id': 500, 'image_count': 1890}, {'id': 501, 'image_count': 99}, {'id': 502, 'image_count': 1530}, {'id': 503, 'image_count': 3}, {'id': 504, 'image_count': 11}, {'id': 505, 'image_count': 19}, {'id': 506, 'image_count': 3}, {'id': 507, 'image_count': 63}, {'id': 508, 'image_count': 5}, {'id': 509, 'image_count': 6}, {'id': 510, 'image_count': 233}, {'id': 511, 'image_count': 54}, {'id': 512, 'image_count': 36}, {'id': 513, 'image_count': 10}, {'id': 514, 'image_count': 124}, {'id': 515, 'image_count': 101}, {'id': 516, 'image_count': 3}, {'id': 517, 'image_count': 363}, {'id': 518, 'image_count': 3}, {'id': 519, 'image_count': 30}, {'id': 520, 'image_count': 18}, {'id': 521, 'image_count': 199}, {'id': 522, 'image_count': 97}, {'id': 523, 'image_count': 32}, {'id': 524, 'image_count': 121}, {'id': 525, 'image_count': 16}, {'id': 526, 'image_count': 12}, {'id': 527, 'image_count': 2}, {'id': 528, 'image_count': 214}, {'id': 529, 'image_count': 48}, {'id': 530, 'image_count': 26}, {'id': 531, 'image_count': 13}, {'id': 532, 'image_count': 4}, {'id': 533, 'image_count': 11}, {'id': 534, 'image_count': 123}, {'id': 535, 'image_count': 7}, {'id': 536, 'image_count': 200}, {'id': 537, 'image_count': 91}, {'id': 538, 'image_count': 9}, {'id': 539, 'image_count': 72}, {'id': 540, 'image_count': 1886}, {'id': 541, 'image_count': 4}, {'id': 542, 'image_count': 1}, {'id': 543, 'image_count': 1}, {'id': 544, 'image_count': 1932}, {'id': 545, 'image_count': 4}, {'id': 546, 'image_count': 56}, {'id': 547, 'image_count': 854}, {'id': 548, 'image_count': 755}, {'id': 549, 'image_count': 1843}, {'id': 550, 'image_count': 96}, {'id': 551, 'image_count': 7}, {'id': 552, 'image_count': 74}, {'id': 553, 'image_count': 66}, {'id': 554, 'image_count': 57}, {'id': 555, 'image_count': 44}, {'id': 556, 'image_count': 1905}, {'id': 557, 'image_count': 4}, {'id': 558, 'image_count': 90}, {'id': 559, 'image_count': 1635}, {'id': 560, 'image_count': 8}, {'id': 561, 'image_count': 5}, {'id': 562, 'image_count': 50}, {'id': 563, 'image_count': 545}, {'id': 564, 'image_count': 20}, {'id': 565, 'image_count': 193}, {'id': 566, 'image_count': 285}, {'id': 567, 'image_count': 3}, {'id': 568, 'image_count': 1}, {'id': 569, 'image_count': 1904}, {'id': 570, 'image_count': 294}, {'id': 571, 'image_count': 3}, {'id': 572, 'image_count': 5}, {'id': 573, 'image_count': 24}, {'id': 574, 'image_count': 2}, {'id': 575, 'image_count': 2}, {'id': 576, 'image_count': 16}, {'id': 577, 'image_count': 8}, {'id': 578, 'image_count': 154}, {'id': 579, 'image_count': 66}, {'id': 580, 'image_count': 1}, {'id': 581, 'image_count': 24}, {'id': 582, 'image_count': 1}, {'id': 583, 'image_count': 4}, {'id': 584, 'image_count': 75}, {'id': 585, 'image_count': 6}, {'id': 586, 'image_count': 126}, {'id': 587, 'image_count': 24}, {'id': 588, 'image_count': 22}, {'id': 589, 'image_count': 1872}, {'id': 590, 'image_count': 16}, {'id': 591, 'image_count': 423}, {'id': 592, 'image_count': 1927}, {'id': 593, 'image_count': 38}, {'id': 594, 'image_count': 3}, {'id': 595, 'image_count': 1945}, {'id': 596, 'image_count': 35}, {'id': 597, 'image_count': 1}, {'id': 598, 'image_count': 13}, {'id': 599, 'image_count': 9}, {'id': 600, 'image_count': 14}, {'id': 601, 'image_count': 37}, {'id': 602, 'image_count': 3}, {'id': 603, 'image_count': 4}, {'id': 604, 'image_count': 100}, {'id': 605, 'image_count': 195}, {'id': 606, 'image_count': 1}, {'id': 607, 'image_count': 12}, {'id': 608, 'image_count': 24}, {'id': 609, 'image_count': 489}, {'id': 610, 'image_count': 10}, {'id': 611, 'image_count': 1689}, {'id': 612, 'image_count': 42}, {'id': 613, 'image_count': 81}, {'id': 614, 'image_count': 894}, {'id': 615, 'image_count': 1868}, {'id': 616, 'image_count': 7}, {'id': 617, 'image_count': 1567}, {'id': 618, 'image_count': 10}, {'id': 619, 'image_count': 8}, {'id': 620, 'image_count': 7}, {'id': 621, 'image_count': 629}, {'id': 622, 'image_count': 89}, {'id': 623, 'image_count': 15}, {'id': 624, 'image_count': 134}, {'id': 625, 'image_count': 4}, {'id': 626, 'image_count': 1802}, {'id': 627, 'image_count': 595}, {'id': 628, 'image_count': 1210}, {'id': 629, 'image_count': 48}, {'id': 630, 'image_count': 418}, {'id': 631, 'image_count': 1846}, {'id': 632, 'image_count': 5}, {'id': 633, 'image_count': 221}, {'id': 634, 'image_count': 10}, {'id': 635, 'image_count': 7}, {'id': 636, 'image_count': 76}, {'id': 637, 'image_count': 22}, {'id': 638, 'image_count': 10}, {'id': 639, 'image_count': 341}, {'id': 640, 'image_count': 1}, {'id': 641, 'image_count': 705}, {'id': 642, 'image_count': 1900}, {'id': 643, 'image_count': 188}, {'id': 644, 'image_count': 227}, {'id': 645, 'image_count': 861}, {'id': 646, 'image_count': 6}, {'id': 647, 'image_count': 115}, {'id': 648, 'image_count': 5}, {'id': 649, 'image_count': 43}, {'id': 650, 'image_count': 14}, {'id': 651, 'image_count': 6}, {'id': 652, 'image_count': 15}, {'id': 653, 'image_count': 1167}, {'id': 654, 'image_count': 15}, {'id': 655, 'image_count': 994}, {'id': 656, 'image_count': 28}, {'id': 657, 'image_count': 2}, {'id': 658, 'image_count': 338}, {'id': 659, 'image_count': 334}, {'id': 660, 'image_count': 15}, {'id': 661, 'image_count': 102}, {'id': 662, 'image_count': 1}, {'id': 663, 'image_count': 8}, {'id': 664, 'image_count': 1}, {'id': 665, 'image_count': 1}, {'id': 666, 'image_count': 28}, {'id': 667, 'image_count': 91}, {'id': 668, 'image_count': 260}, {'id': 669, 'image_count': 131}, {'id': 670, 'image_count': 128}, {'id': 671, 'image_count': 3}, {'id': 672, 'image_count': 10}, {'id': 673, 'image_count': 39}, {'id': 674, 'image_count': 2}, {'id': 675, 'image_count': 925}, {'id': 676, 'image_count': 354}, {'id': 677, 'image_count': 31}, {'id': 678, 'image_count': 10}, {'id': 679, 'image_count': 215}, {'id': 680, 'image_count': 71}, {'id': 681, 'image_count': 43}, {'id': 682, 'image_count': 28}, {'id': 683, 'image_count': 34}, {'id': 684, 'image_count': 16}, {'id': 685, 'image_count': 273}, {'id': 686, 'image_count': 2}, {'id': 687, 'image_count': 999}, {'id': 688, 'image_count': 4}, {'id': 689, 'image_count': 107}, {'id': 690, 'image_count': 2}, {'id': 691, 'image_count': 1}, {'id': 692, 'image_count': 454}, {'id': 693, 'image_count': 9}, {'id': 694, 'image_count': 1901}, {'id': 695, 'image_count': 61}, {'id': 696, 'image_count': 91}, {'id': 697, 'image_count': 46}, {'id': 698, 'image_count': 1402}, {'id': 699, 'image_count': 74}, {'id': 700, 'image_count': 421}, {'id': 701, 'image_count': 226}, {'id': 702, 'image_count': 10}, {'id': 703, 'image_count': 1720}, {'id': 704, 'image_count': 261}, {'id': 705, 'image_count': 1337}, {'id': 706, 'image_count': 293}, {'id': 707, 'image_count': 62}, {'id': 708, 'image_count': 814}, {'id': 709, 'image_count': 407}, {'id': 710, 'image_count': 6}, {'id': 711, 'image_count': 16}, {'id': 712, 'image_count': 7}, {'id': 713, 'image_count': 1791}, {'id': 714, 'image_count': 2}, {'id': 715, 'image_count': 1915}, {'id': 716, 'image_count': 1940}, {'id': 717, 'image_count': 13}, {'id': 718, 'image_count': 16}, {'id': 719, 'image_count': 448}, {'id': 720, 'image_count': 12}, {'id': 721, 'image_count': 18}, {'id': 722, 'image_count': 4}, {'id': 723, 'image_count': 71}, {'id': 724, 'image_count': 189}, {'id': 725, 'image_count': 74}, {'id': 726, 'image_count': 103}, {'id': 727, 'image_count': 3}, {'id': 728, 'image_count': 110}, {'id': 729, 'image_count': 5}, {'id': 730, 'image_count': 9}, {'id': 731, 'image_count': 15}, {'id': 732, 'image_count': 25}, {'id': 733, 'image_count': 7}, {'id': 734, 'image_count': 647}, {'id': 735, 'image_count': 824}, {'id': 736, 'image_count': 100}, {'id': 737, 'image_count': 47}, {'id': 738, 'image_count': 121}, {'id': 739, 'image_count': 731}, {'id': 740, 'image_count': 73}, {'id': 741, 'image_count': 49}, {'id': 742, 'image_count': 23}, {'id': 743, 'image_count': 4}, {'id': 744, 'image_count': 62}, {'id': 745, 'image_count': 118}, {'id': 746, 'image_count': 99}, {'id': 747, 'image_count': 40}, {'id': 748, 'image_count': 1036}, {'id': 749, 'image_count': 105}, {'id': 750, 'image_count': 21}, {'id': 751, 'image_count': 229}, {'id': 752, 'image_count': 7}, {'id': 753, 'image_count': 72}, {'id': 754, 'image_count': 9}, {'id': 755, 'image_count': 10}, {'id': 756, 'image_count': 328}, {'id': 757, 'image_count': 468}, {'id': 758, 'image_count': 1}, {'id': 759, 'image_count': 2}, {'id': 760, 'image_count': 24}, {'id': 761, 'image_count': 11}, {'id': 762, 'image_count': 72}, {'id': 763, 'image_count': 17}, {'id': 764, 'image_count': 10}, {'id': 765, 'image_count': 17}, {'id': 766, 'image_count': 489}, {'id': 767, 'image_count': 47}, {'id': 768, 'image_count': 93}, {'id': 769, 'image_count': 1}, {'id': 770, 'image_count': 12}, {'id': 771, 'image_count': 228}, {'id': 772, 'image_count': 5}, {'id': 773, 'image_count': 76}, {'id': 774, 'image_count': 71}, {'id': 775, 'image_count': 30}, {'id': 776, 'image_count': 109}, {'id': 777, 'image_count': 14}, {'id': 778, 'image_count': 1}, {'id': 779, 'image_count': 8}, {'id': 780, 'image_count': 26}, {'id': 781, 'image_count': 339}, {'id': 782, 'image_count': 153}, {'id': 783, 'image_count': 2}, {'id': 784, 'image_count': 3}, {'id': 785, 'image_count': 8}, {'id': 786, 'image_count': 47}, {'id': 787, 'image_count': 8}, {'id': 788, 'image_count': 6}, {'id': 789, 'image_count': 116}, {'id': 790, 'image_count': 69}, {'id': 791, 'image_count': 13}, {'id': 792, 'image_count': 6}, {'id': 793, 'image_count': 1928}, {'id': 794, 'image_count': 79}, {'id': 795, 'image_count': 14}, {'id': 796, 'image_count': 7}, {'id': 797, 'image_count': 20}, {'id': 798, 'image_count': 114}, {'id': 799, 'image_count': 221}, {'id': 800, 'image_count': 502}, {'id': 801, 'image_count': 62}, {'id': 802, 'image_count': 87}, {'id': 803, 'image_count': 4}, {'id': 804, 'image_count': 1912}, {'id': 805, 'image_count': 7}, {'id': 806, 'image_count': 186}, {'id': 807, 'image_count': 18}, {'id': 808, 'image_count': 4}, {'id': 809, 'image_count': 3}, {'id': 810, 'image_count': 7}, {'id': 811, 'image_count': 1413}, {'id': 812, 'image_count': 7}, {'id': 813, 'image_count': 12}, {'id': 814, 'image_count': 248}, {'id': 815, 'image_count': 4}, {'id': 816, 'image_count': 1881}, {'id': 817, 'image_count': 529}, {'id': 818, 'image_count': 1932}, {'id': 819, 'image_count': 50}, {'id': 820, 'image_count': 3}, {'id': 821, 'image_count': 28}, {'id': 822, 'image_count': 10}, {'id': 823, 'image_count': 5}, {'id': 824, 'image_count': 5}, {'id': 825, 'image_count': 18}, {'id': 826, 'image_count': 14}, {'id': 827, 'image_count': 1890}, {'id': 828, 'image_count': 660}, {'id': 829, 'image_count': 8}, {'id': 830, 'image_count': 25}, {'id': 831, 'image_count': 10}, {'id': 832, 'image_count': 218}, {'id': 833, 'image_count': 36}, {'id': 834, 'image_count': 16}, {'id': 835, 'image_count': 808}, {'id': 836, 'image_count': 479}, {'id': 837, 'image_count': 1404}, {'id': 838, 'image_count': 307}, {'id': 839, 'image_count': 57}, {'id': 840, 'image_count': 28}, {'id': 841, 'image_count': 80}, {'id': 842, 'image_count': 11}, {'id': 843, 'image_count': 92}, {'id': 844, 'image_count': 20}, {'id': 845, 'image_count': 194}, {'id': 846, 'image_count': 23}, {'id': 847, 'image_count': 52}, {'id': 848, 'image_count': 673}, {'id': 849, 'image_count': 2}, {'id': 850, 'image_count': 2}, {'id': 851, 'image_count': 1}, {'id': 852, 'image_count': 2}, {'id': 853, 'image_count': 8}, {'id': 854, 'image_count': 80}, {'id': 855, 'image_count': 3}, {'id': 856, 'image_count': 3}, {'id': 857, 'image_count': 15}, {'id': 858, 'image_count': 2}, {'id': 859, 'image_count': 10}, {'id': 860, 'image_count': 386}, {'id': 861, 'image_count': 65}, {'id': 862, 'image_count': 3}, {'id': 863, 'image_count': 35}, {'id': 864, 'image_count': 5}, {'id': 865, 'image_count': 180}, {'id': 866, 'image_count': 99}, {'id': 867, 'image_count': 49}, {'id': 868, 'image_count': 28}, {'id': 869, 'image_count': 1}, {'id': 870, 'image_count': 52}, {'id': 871, 'image_count': 36}, {'id': 872, 'image_count': 70}, {'id': 873, 'image_count': 6}, {'id': 874, 'image_count': 29}, {'id': 875, 'image_count': 24}, {'id': 876, 'image_count': 1115}, {'id': 877, 'image_count': 61}, {'id': 878, 'image_count': 18}, {'id': 879, 'image_count': 18}, {'id': 880, 'image_count': 665}, {'id': 881, 'image_count': 1096}, {'id': 882, 'image_count': 29}, {'id': 883, 'image_count': 8}, {'id': 884, 'image_count': 14}, {'id': 885, 'image_count': 1622}, {'id': 886, 'image_count': 2}, {'id': 887, 'image_count': 3}, {'id': 888, 'image_count': 32}, {'id': 889, 'image_count': 55}, {'id': 890, 'image_count': 1}, {'id': 891, 'image_count': 10}, {'id': 892, 'image_count': 10}, {'id': 893, 'image_count': 47}, {'id': 894, 'image_count': 3}, {'id': 895, 'image_count': 29}, {'id': 896, 'image_count': 342}, {'id': 897, 'image_count': 25}, {'id': 898, 'image_count': 1469}, {'id': 899, 'image_count': 521}, {'id': 900, 'image_count': 347}, {'id': 901, 'image_count': 35}, {'id': 902, 'image_count': 7}, {'id': 903, 'image_count': 207}, {'id': 904, 'image_count': 108}, {'id': 905, 'image_count': 2}, {'id': 906, 'image_count': 34}, {'id': 907, 'image_count': 12}, {'id': 908, 'image_count': 10}, {'id': 909, 'image_count': 13}, {'id': 910, 'image_count': 361}, {'id': 911, 'image_count': 1023}, {'id': 912, 'image_count': 782}, {'id': 913, 'image_count': 2}, {'id': 914, 'image_count': 5}, {'id': 915, 'image_count': 247}, {'id': 916, 'image_count': 221}, {'id': 917, 'image_count': 4}, {'id': 918, 'image_count': 8}, {'id': 919, 'image_count': 158}, {'id': 920, 'image_count': 3}, {'id': 921, 'image_count': 752}, {'id': 922, 'image_count': 64}, {'id': 923, 'image_count': 707}, {'id': 924, 'image_count': 143}, {'id': 925, 'image_count': 1}, {'id': 926, 'image_count': 49}, {'id': 927, 'image_count': 126}, {'id': 928, 'image_count': 76}, {'id': 929, 'image_count': 11}, {'id': 930, 'image_count': 11}, {'id': 931, 'image_count': 4}, {'id': 932, 'image_count': 39}, {'id': 933, 'image_count': 11}, {'id': 934, 'image_count': 13}, {'id': 935, 'image_count': 91}, {'id': 936, 'image_count': 14}, {'id': 937, 'image_count': 5}, {'id': 938, 'image_count': 3}, {'id': 939, 'image_count': 10}, {'id': 940, 'image_count': 18}, {'id': 941, 'image_count': 9}, {'id': 942, 'image_count': 6}, {'id': 943, 'image_count': 951}, {'id': 944, 'image_count': 2}, {'id': 945, 'image_count': 1}, {'id': 946, 'image_count': 19}, {'id': 947, 'image_count': 1942}, {'id': 948, 'image_count': 1916}, {'id': 949, 'image_count': 139}, {'id': 950, 'image_count': 43}, {'id': 951, 'image_count': 1969}, {'id': 952, 'image_count': 5}, {'id': 953, 'image_count': 134}, {'id': 954, 'image_count': 74}, {'id': 955, 'image_count': 381}, {'id': 956, 'image_count': 1}, {'id': 957, 'image_count': 381}, {'id': 958, 'image_count': 6}, {'id': 959, 'image_count': 1826}, {'id': 960, 'image_count': 28}, {'id': 961, 'image_count': 1635}, {'id': 962, 'image_count': 1967}, {'id': 963, 'image_count': 16}, {'id': 964, 'image_count': 1926}, {'id': 965, 'image_count': 1789}, {'id': 966, 'image_count': 401}, {'id': 967, 'image_count': 1968}, {'id': 968, 'image_count': 1167}, {'id': 969, 'image_count': 1}, {'id': 970, 'image_count': 56}, {'id': 971, 'image_count': 17}, {'id': 972, 'image_count': 1}, {'id': 973, 'image_count': 58}, {'id': 974, 'image_count': 9}, {'id': 975, 'image_count': 8}, {'id': 976, 'image_count': 1124}, {'id': 977, 'image_count': 31}, {'id': 978, 'image_count': 16}, {'id': 979, 'image_count': 491}, {'id': 980, 'image_count': 432}, {'id': 981, 'image_count': 1945}, {'id': 982, 'image_count': 1899}, {'id': 983, 'image_count': 5}, {'id': 984, 'image_count': 28}, {'id': 985, 'image_count': 7}, {'id': 986, 'image_count': 146}, {'id': 987, 'image_count': 1}, {'id': 988, 'image_count': 25}, {'id': 989, 'image_count': 22}, {'id': 990, 'image_count': 1}, {'id': 991, 'image_count': 10}, {'id': 992, 'image_count': 9}, {'id': 993, 'image_count': 308}, {'id': 994, 'image_count': 4}, {'id': 995, 'image_count': 1969}, {'id': 996, 'image_count': 45}, {'id': 997, 'image_count': 12}, {'id': 998, 'image_count': 1}, {'id': 999, 'image_count': 85}, {'id': 1000, 'image_count': 1127}, {'id': 1001, 'image_count': 11}, {'id': 1002, 'image_count': 60}, {'id': 1003, 'image_count': 1}, {'id': 1004, 'image_count': 16}, {'id': 1005, 'image_count': 1}, {'id': 1006, 'image_count': 65}, {'id': 1007, 'image_count': 13}, {'id': 1008, 'image_count': 655}, {'id': 1009, 'image_count': 51}, {'id': 1010, 'image_count': 1}, {'id': 1011, 'image_count': 673}, {'id': 1012, 'image_count': 5}, {'id': 1013, 'image_count': 36}, {'id': 1014, 'image_count': 54}, {'id': 1015, 'image_count': 5}, {'id': 1016, 'image_count': 8}, {'id': 1017, 'image_count': 305}, {'id': 1018, 'image_count': 297}, {'id': 1019, 'image_count': 1053}, {'id': 1020, 'image_count': 223}, {'id': 1021, 'image_count': 1037}, {'id': 1022, 'image_count': 63}, {'id': 1023, 'image_count': 1881}, {'id': 1024, 'image_count': 507}, {'id': 1025, 'image_count': 333}, {'id': 1026, 'image_count': 1911}, {'id': 1027, 'image_count': 1765}, {'id': 1028, 'image_count': 1}, {'id': 1029, 'image_count': 5}, {'id': 1030, 'image_count': 1}, {'id': 1031, 'image_count': 9}, {'id': 1032, 'image_count': 2}, {'id': 1033, 'image_count': 151}, {'id': 1034, 'image_count': 82}, {'id': 1035, 'image_count': 1931}, {'id': 1036, 'image_count': 41}, {'id': 1037, 'image_count': 1895}, {'id': 1038, 'image_count': 24}, {'id': 1039, 'image_count': 22}, {'id': 1040, 'image_count': 35}, {'id': 1041, 'image_count': 69}, {'id': 1042, 'image_count': 962}, {'id': 1043, 'image_count': 588}, {'id': 1044, 'image_count': 21}, {'id': 1045, 'image_count': 825}, {'id': 1046, 'image_count': 52}, {'id': 1047, 'image_count': 5}, {'id': 1048, 'image_count': 5}, {'id': 1049, 'image_count': 5}, {'id': 1050, 'image_count': 1860}, {'id': 1051, 'image_count': 56}, {'id': 1052, 'image_count': 1582}, {'id': 1053, 'image_count': 7}, {'id': 1054, 'image_count': 2}, {'id': 1055, 'image_count': 1562}, {'id': 1056, 'image_count': 1885}, {'id': 1057, 'image_count': 1}, {'id': 1058, 'image_count': 5}, {'id': 1059, 'image_count': 137}, {'id': 1060, 'image_count': 1094}, {'id': 1061, 'image_count': 134}, {'id': 1062, 'image_count': 29}, {'id': 1063, 'image_count': 22}, {'id': 1064, 'image_count': 522}, {'id': 1065, 'image_count': 50}, {'id': 1066, 'image_count': 68}, {'id': 1067, 'image_count': 16}, {'id': 1068, 'image_count': 40}, {'id': 1069, 'image_count': 35}, {'id': 1070, 'image_count': 135}, {'id': 1071, 'image_count': 1413}, {'id': 1072, 'image_count': 772}, {'id': 1073, 'image_count': 50}, {'id': 1074, 'image_count': 1015}, {'id': 1075, 'image_count': 1}, {'id': 1076, 'image_count': 65}, {'id': 1077, 'image_count': 1900}, {'id': 1078, 'image_count': 1302}, {'id': 1079, 'image_count': 1977}, {'id': 1080, 'image_count': 2}, {'id': 1081, 'image_count': 29}, {'id': 1082, 'image_count': 36}, {'id': 1083, 'image_count': 138}, {'id': 1084, 'image_count': 4}, {'id': 1085, 'image_count': 67}, {'id': 1086, 'image_count': 26}, {'id': 1087, 'image_count': 25}, {'id': 1088, 'image_count': 33}, {'id': 1089, 'image_count': 37}, {'id': 1090, 'image_count': 50}, {'id': 1091, 'image_count': 270}, {'id': 1092, 'image_count': 12}, {'id': 1093, 'image_count': 316}, {'id': 1094, 'image_count': 41}, {'id': 1095, 'image_count': 224}, {'id': 1096, 'image_count': 105}, {'id': 1097, 'image_count': 1925}, {'id': 1098, 'image_count': 1021}, {'id': 1099, 'image_count': 1213}, {'id': 1100, 'image_count': 172}, {'id': 1101, 'image_count': 28}, {'id': 1102, 'image_count': 745}, {'id': 1103, 'image_count': 187}, {'id': 1104, 'image_count': 147}, {'id': 1105, 'image_count': 136}, {'id': 1106, 'image_count': 34}, {'id': 1107, 'image_count': 41}, {'id': 1108, 'image_count': 636}, {'id': 1109, 'image_count': 570}, {'id': 1110, 'image_count': 1149}, {'id': 1111, 'image_count': 61}, {'id': 1112, 'image_count': 1890}, {'id': 1113, 'image_count': 18}, {'id': 1114, 'image_count': 143}, {'id': 1115, 'image_count': 1517}, {'id': 1116, 'image_count': 7}, {'id': 1117, 'image_count': 943}, {'id': 1118, 'image_count': 6}, {'id': 1119, 'image_count': 1}, {'id': 1120, 'image_count': 11}, {'id': 1121, 'image_count': 101}, {'id': 1122, 'image_count': 1909}, {'id': 1123, 'image_count': 800}, {'id': 1124, 'image_count': 1}, {'id': 1125, 'image_count': 44}, {'id': 1126, 'image_count': 3}, {'id': 1127, 'image_count': 44}, {'id': 1128, 'image_count': 31}, {'id': 1129, 'image_count': 7}, {'id': 1130, 'image_count': 20}, {'id': 1131, 'image_count': 11}, {'id': 1132, 'image_count': 13}, {'id': 1133, 'image_count': 1924}, {'id': 1134, 'image_count': 113}, {'id': 1135, 'image_count': 2}, {'id': 1136, 'image_count': 139}, {'id': 1137, 'image_count': 12}, {'id': 1138, 'image_count': 37}, {'id': 1139, 'image_count': 1866}, {'id': 1140, 'image_count': 47}, {'id': 1141, 'image_count': 1468}, {'id': 1142, 'image_count': 729}, {'id': 1143, 'image_count': 24}, {'id': 1144, 'image_count': 1}, {'id': 1145, 'image_count': 10}, {'id': 1146, 'image_count': 3}, {'id': 1147, 'image_count': 14}, {'id': 1148, 'image_count': 4}, {'id': 1149, 'image_count': 29}, {'id': 1150, 'image_count': 4}, {'id': 1151, 'image_count': 70}, {'id': 1152, 'image_count': 46}, {'id': 1153, 'image_count': 14}, {'id': 1154, 'image_count': 48}, {'id': 1155, 'image_count': 1855}, {'id': 1156, 'image_count': 113}, {'id': 1157, 'image_count': 1}, {'id': 1158, 'image_count': 1}, {'id': 1159, 'image_count': 10}, {'id': 1160, 'image_count': 54}, {'id': 1161, 'image_count': 1923}, {'id': 1162, 'image_count': 630}, {'id': 1163, 'image_count': 31}, {'id': 1164, 'image_count': 69}, {'id': 1165, 'image_count': 7}, {'id': 1166, 'image_count': 11}, {'id': 1167, 'image_count': 1}, {'id': 1168, 'image_count': 30}, {'id': 1169, 'image_count': 50}, {'id': 1170, 'image_count': 45}, {'id': 1171, 'image_count': 28}, {'id': 1172, 'image_count': 114}, {'id': 1173, 'image_count': 193}, {'id': 1174, 'image_count': 21}, {'id': 1175, 'image_count': 91}, {'id': 1176, 'image_count': 31}, {'id': 1177, 'image_count': 1469}, {'id': 1178, 'image_count': 1924}, {'id': 1179, 'image_count': 87}, {'id': 1180, 'image_count': 77}, {'id': 1181, 'image_count': 11}, {'id': 1182, 'image_count': 47}, {'id': 1183, 'image_count': 21}, {'id': 1184, 'image_count': 47}, {'id': 1185, 'image_count': 70}, {'id': 1186, 'image_count': 1838}, {'id': 1187, 'image_count': 19}, {'id': 1188, 'image_count': 531}, {'id': 1189, 'image_count': 11}, {'id': 1190, 'image_count': 941}, {'id': 1191, 'image_count': 113}, {'id': 1192, 'image_count': 26}, {'id': 1193, 'image_count': 5}, {'id': 1194, 'image_count': 56}, {'id': 1195, 'image_count': 73}, {'id': 1196, 'image_count': 32}, {'id': 1197, 'image_count': 128}, {'id': 1198, 'image_count': 623}, {'id': 1199, 'image_count': 12}, {'id': 1200, 'image_count': 52}, {'id': 1201, 'image_count': 11}, {'id': 1202, 'image_count': 1674}, {'id': 1203, 'image_count': 81}] # noqa
20
+ # fmt: on
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/pascal_voc.py ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ import numpy as np
5
+ import os
6
+ import xml.etree.ElementTree as ET
7
+ from typing import List, Tuple, Union
8
+
9
+ from annotator.oneformer.detectron2.data import DatasetCatalog, MetadataCatalog
10
+ from annotator.oneformer.detectron2.structures import BoxMode
11
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
12
+
13
+ __all__ = ["load_voc_instances", "register_pascal_voc"]
14
+
15
+
16
+ # fmt: off
17
+ CLASS_NAMES = (
18
+ "aeroplane", "bicycle", "bird", "boat", "bottle", "bus", "car", "cat",
19
+ "chair", "cow", "diningtable", "dog", "horse", "motorbike", "person",
20
+ "pottedplant", "sheep", "sofa", "train", "tvmonitor"
21
+ )
22
+ # fmt: on
23
+
24
+
25
+ def load_voc_instances(dirname: str, split: str, class_names: Union[List[str], Tuple[str, ...]]):
26
+ """
27
+ Load Pascal VOC detection annotations to Detectron2 format.
28
+
29
+ Args:
30
+ dirname: Contain "Annotations", "ImageSets", "JPEGImages"
31
+ split (str): one of "train", "test", "val", "trainval"
32
+ class_names: list or tuple of class names
33
+ """
34
+ with PathManager.open(os.path.join(dirname, "ImageSets", "Main", split + ".txt")) as f:
35
+ fileids = np.loadtxt(f, dtype=np.str)
36
+
37
+ # Needs to read many small annotation files. Makes sense at local
38
+ annotation_dirname = PathManager.get_local_path(os.path.join(dirname, "Annotations/"))
39
+ dicts = []
40
+ for fileid in fileids:
41
+ anno_file = os.path.join(annotation_dirname, fileid + ".xml")
42
+ jpeg_file = os.path.join(dirname, "JPEGImages", fileid + ".jpg")
43
+
44
+ with PathManager.open(anno_file) as f:
45
+ tree = ET.parse(f)
46
+
47
+ r = {
48
+ "file_name": jpeg_file,
49
+ "image_id": fileid,
50
+ "height": int(tree.findall("./size/height")[0].text),
51
+ "width": int(tree.findall("./size/width")[0].text),
52
+ }
53
+ instances = []
54
+
55
+ for obj in tree.findall("object"):
56
+ cls = obj.find("name").text
57
+ # We include "difficult" samples in training.
58
+ # Based on limited experiments, they don't hurt accuracy.
59
+ # difficult = int(obj.find("difficult").text)
60
+ # if difficult == 1:
61
+ # continue
62
+ bbox = obj.find("bndbox")
63
+ bbox = [float(bbox.find(x).text) for x in ["xmin", "ymin", "xmax", "ymax"]]
64
+ # Original annotations are integers in the range [1, W or H]
65
+ # Assuming they mean 1-based pixel indices (inclusive),
66
+ # a box with annotation (xmin=1, xmax=W) covers the whole image.
67
+ # In coordinate space this is represented by (xmin=0, xmax=W)
68
+ bbox[0] -= 1.0
69
+ bbox[1] -= 1.0
70
+ instances.append(
71
+ {"category_id": class_names.index(cls), "bbox": bbox, "bbox_mode": BoxMode.XYXY_ABS}
72
+ )
73
+ r["annotations"] = instances
74
+ dicts.append(r)
75
+ return dicts
76
+
77
+
78
+ def register_pascal_voc(name, dirname, split, year, class_names=CLASS_NAMES):
79
+ DatasetCatalog.register(name, lambda: load_voc_instances(dirname, split, class_names))
80
+ MetadataCatalog.get(name).set(
81
+ thing_classes=list(class_names), dirname=dirname, year=year, split=split
82
+ )
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/datasets/register_coco.py ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from .coco import register_coco_instances # noqa
3
+ from .coco_panoptic import register_coco_panoptic_separated # noqa
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/detection_utils.py ADDED
@@ -0,0 +1,659 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ """
5
+ Common data processing utilities that are used in a
6
+ typical object detection data pipeline.
7
+ """
8
+ import logging
9
+ import numpy as np
10
+ from typing import List, Union
11
+ import pycocotools.mask as mask_util
12
+ import torch
13
+ from PIL import Image
14
+
15
+ from annotator.oneformer.detectron2.structures import (
16
+ BitMasks,
17
+ Boxes,
18
+ BoxMode,
19
+ Instances,
20
+ Keypoints,
21
+ PolygonMasks,
22
+ RotatedBoxes,
23
+ polygons_to_bitmask,
24
+ )
25
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
26
+
27
+ from . import transforms as T
28
+ from .catalog import MetadataCatalog
29
+
30
+ __all__ = [
31
+ "SizeMismatchError",
32
+ "convert_image_to_rgb",
33
+ "check_image_size",
34
+ "transform_proposals",
35
+ "transform_instance_annotations",
36
+ "annotations_to_instances",
37
+ "annotations_to_instances_rotated",
38
+ "build_augmentation",
39
+ "build_transform_gen",
40
+ "create_keypoint_hflip_indices",
41
+ "filter_empty_instances",
42
+ "read_image",
43
+ ]
44
+
45
+
46
+ class SizeMismatchError(ValueError):
47
+ """
48
+ When loaded image has difference width/height compared with annotation.
49
+ """
50
+
51
+
52
+ # https://en.wikipedia.org/wiki/YUV#SDTV_with_BT.601
53
+ _M_RGB2YUV = [[0.299, 0.587, 0.114], [-0.14713, -0.28886, 0.436], [0.615, -0.51499, -0.10001]]
54
+ _M_YUV2RGB = [[1.0, 0.0, 1.13983], [1.0, -0.39465, -0.58060], [1.0, 2.03211, 0.0]]
55
+
56
+ # https://www.exiv2.org/tags.html
57
+ _EXIF_ORIENT = 274 # exif 'Orientation' tag
58
+
59
+
60
+ def convert_PIL_to_numpy(image, format):
61
+ """
62
+ Convert PIL image to numpy array of target format.
63
+
64
+ Args:
65
+ image (PIL.Image): a PIL image
66
+ format (str): the format of output image
67
+
68
+ Returns:
69
+ (np.ndarray): also see `read_image`
70
+ """
71
+ if format is not None:
72
+ # PIL only supports RGB, so convert to RGB and flip channels over below
73
+ conversion_format = format
74
+ if format in ["BGR", "YUV-BT.601"]:
75
+ conversion_format = "RGB"
76
+ image = image.convert(conversion_format)
77
+ image = np.asarray(image)
78
+ # PIL squeezes out the channel dimension for "L", so make it HWC
79
+ if format == "L":
80
+ image = np.expand_dims(image, -1)
81
+
82
+ # handle formats not supported by PIL
83
+ elif format == "BGR":
84
+ # flip channels if needed
85
+ image = image[:, :, ::-1]
86
+ elif format == "YUV-BT.601":
87
+ image = image / 255.0
88
+ image = np.dot(image, np.array(_M_RGB2YUV).T)
89
+
90
+ return image
91
+
92
+
93
+ def convert_image_to_rgb(image, format):
94
+ """
95
+ Convert an image from given format to RGB.
96
+
97
+ Args:
98
+ image (np.ndarray or Tensor): an HWC image
99
+ format (str): the format of input image, also see `read_image`
100
+
101
+ Returns:
102
+ (np.ndarray): (H,W,3) RGB image in 0-255 range, can be either float or uint8
103
+ """
104
+ if isinstance(image, torch.Tensor):
105
+ image = image.cpu().numpy()
106
+ if format == "BGR":
107
+ image = image[:, :, [2, 1, 0]]
108
+ elif format == "YUV-BT.601":
109
+ image = np.dot(image, np.array(_M_YUV2RGB).T)
110
+ image = image * 255.0
111
+ else:
112
+ if format == "L":
113
+ image = image[:, :, 0]
114
+ image = image.astype(np.uint8)
115
+ image = np.asarray(Image.fromarray(image, mode=format).convert("RGB"))
116
+ return image
117
+
118
+
119
+ def _apply_exif_orientation(image):
120
+ """
121
+ Applies the exif orientation correctly.
122
+
123
+ This code exists per the bug:
124
+ https://github.com/python-pillow/Pillow/issues/3973
125
+ with the function `ImageOps.exif_transpose`. The Pillow source raises errors with
126
+ various methods, especially `tobytes`
127
+
128
+ Function based on:
129
+ https://github.com/wkentaro/labelme/blob/v4.5.4/labelme/utils/image.py#L59
130
+ https://github.com/python-pillow/Pillow/blob/7.1.2/src/PIL/ImageOps.py#L527
131
+
132
+ Args:
133
+ image (PIL.Image): a PIL image
134
+
135
+ Returns:
136
+ (PIL.Image): the PIL image with exif orientation applied, if applicable
137
+ """
138
+ if not hasattr(image, "getexif"):
139
+ return image
140
+
141
+ try:
142
+ exif = image.getexif()
143
+ except Exception: # https://github.com/facebookresearch/detectron2/issues/1885
144
+ exif = None
145
+
146
+ if exif is None:
147
+ return image
148
+
149
+ orientation = exif.get(_EXIF_ORIENT)
150
+
151
+ method = {
152
+ 2: Image.FLIP_LEFT_RIGHT,
153
+ 3: Image.ROTATE_180,
154
+ 4: Image.FLIP_TOP_BOTTOM,
155
+ 5: Image.TRANSPOSE,
156
+ 6: Image.ROTATE_270,
157
+ 7: Image.TRANSVERSE,
158
+ 8: Image.ROTATE_90,
159
+ }.get(orientation)
160
+
161
+ if method is not None:
162
+ return image.transpose(method)
163
+ return image
164
+
165
+
166
+ def read_image(file_name, format=None):
167
+ """
168
+ Read an image into the given format.
169
+ Will apply rotation and flipping if the image has such exif information.
170
+
171
+ Args:
172
+ file_name (str): image file path
173
+ format (str): one of the supported image modes in PIL, or "BGR" or "YUV-BT.601".
174
+
175
+ Returns:
176
+ image (np.ndarray):
177
+ an HWC image in the given format, which is 0-255, uint8 for
178
+ supported image modes in PIL or "BGR"; float (0-1 for Y) for YUV-BT.601.
179
+ """
180
+ with PathManager.open(file_name, "rb") as f:
181
+ image = Image.open(f)
182
+
183
+ # work around this bug: https://github.com/python-pillow/Pillow/issues/3973
184
+ image = _apply_exif_orientation(image)
185
+ return convert_PIL_to_numpy(image, format)
186
+
187
+
188
+ def check_image_size(dataset_dict, image):
189
+ """
190
+ Raise an error if the image does not match the size specified in the dict.
191
+ """
192
+ if "width" in dataset_dict or "height" in dataset_dict:
193
+ image_wh = (image.shape[1], image.shape[0])
194
+ expected_wh = (dataset_dict["width"], dataset_dict["height"])
195
+ if not image_wh == expected_wh:
196
+ raise SizeMismatchError(
197
+ "Mismatched image shape{}, got {}, expect {}.".format(
198
+ " for image " + dataset_dict["file_name"]
199
+ if "file_name" in dataset_dict
200
+ else "",
201
+ image_wh,
202
+ expected_wh,
203
+ )
204
+ + " Please check the width/height in your annotation."
205
+ )
206
+
207
+ # To ensure bbox always remap to original image size
208
+ if "width" not in dataset_dict:
209
+ dataset_dict["width"] = image.shape[1]
210
+ if "height" not in dataset_dict:
211
+ dataset_dict["height"] = image.shape[0]
212
+
213
+
214
+ def transform_proposals(dataset_dict, image_shape, transforms, *, proposal_topk, min_box_size=0):
215
+ """
216
+ Apply transformations to the proposals in dataset_dict, if any.
217
+
218
+ Args:
219
+ dataset_dict (dict): a dict read from the dataset, possibly
220
+ contains fields "proposal_boxes", "proposal_objectness_logits", "proposal_bbox_mode"
221
+ image_shape (tuple): height, width
222
+ transforms (TransformList):
223
+ proposal_topk (int): only keep top-K scoring proposals
224
+ min_box_size (int): proposals with either side smaller than this
225
+ threshold are removed
226
+
227
+ The input dict is modified in-place, with abovementioned keys removed. A new
228
+ key "proposals" will be added. Its value is an `Instances`
229
+ object which contains the transformed proposals in its field
230
+ "proposal_boxes" and "objectness_logits".
231
+ """
232
+ if "proposal_boxes" in dataset_dict:
233
+ # Transform proposal boxes
234
+ boxes = transforms.apply_box(
235
+ BoxMode.convert(
236
+ dataset_dict.pop("proposal_boxes"),
237
+ dataset_dict.pop("proposal_bbox_mode"),
238
+ BoxMode.XYXY_ABS,
239
+ )
240
+ )
241
+ boxes = Boxes(boxes)
242
+ objectness_logits = torch.as_tensor(
243
+ dataset_dict.pop("proposal_objectness_logits").astype("float32")
244
+ )
245
+
246
+ boxes.clip(image_shape)
247
+ keep = boxes.nonempty(threshold=min_box_size)
248
+ boxes = boxes[keep]
249
+ objectness_logits = objectness_logits[keep]
250
+
251
+ proposals = Instances(image_shape)
252
+ proposals.proposal_boxes = boxes[:proposal_topk]
253
+ proposals.objectness_logits = objectness_logits[:proposal_topk]
254
+ dataset_dict["proposals"] = proposals
255
+
256
+
257
+ def get_bbox(annotation):
258
+ """
259
+ Get bbox from data
260
+ Args:
261
+ annotation (dict): dict of instance annotations for a single instance.
262
+ Returns:
263
+ bbox (ndarray): x1, y1, x2, y2 coordinates
264
+ """
265
+ # bbox is 1d (per-instance bounding box)
266
+ bbox = BoxMode.convert(annotation["bbox"], annotation["bbox_mode"], BoxMode.XYXY_ABS)
267
+ return bbox
268
+
269
+
270
+ def transform_instance_annotations(
271
+ annotation, transforms, image_size, *, keypoint_hflip_indices=None
272
+ ):
273
+ """
274
+ Apply transforms to box, segmentation and keypoints annotations of a single instance.
275
+
276
+ It will use `transforms.apply_box` for the box, and
277
+ `transforms.apply_coords` for segmentation polygons & keypoints.
278
+ If you need anything more specially designed for each data structure,
279
+ you'll need to implement your own version of this function or the transforms.
280
+
281
+ Args:
282
+ annotation (dict): dict of instance annotations for a single instance.
283
+ It will be modified in-place.
284
+ transforms (TransformList or list[Transform]):
285
+ image_size (tuple): the height, width of the transformed image
286
+ keypoint_hflip_indices (ndarray[int]): see `create_keypoint_hflip_indices`.
287
+
288
+ Returns:
289
+ dict:
290
+ the same input dict with fields "bbox", "segmentation", "keypoints"
291
+ transformed according to `transforms`.
292
+ The "bbox_mode" field will be set to XYXY_ABS.
293
+ """
294
+ if isinstance(transforms, (tuple, list)):
295
+ transforms = T.TransformList(transforms)
296
+ # bbox is 1d (per-instance bounding box)
297
+ bbox = BoxMode.convert(annotation["bbox"], annotation["bbox_mode"], BoxMode.XYXY_ABS)
298
+ # clip transformed bbox to image size
299
+ bbox = transforms.apply_box(np.array([bbox]))[0].clip(min=0)
300
+ annotation["bbox"] = np.minimum(bbox, list(image_size + image_size)[::-1])
301
+ annotation["bbox_mode"] = BoxMode.XYXY_ABS
302
+
303
+ if "segmentation" in annotation:
304
+ # each instance contains 1 or more polygons
305
+ segm = annotation["segmentation"]
306
+ if isinstance(segm, list):
307
+ # polygons
308
+ polygons = [np.asarray(p).reshape(-1, 2) for p in segm]
309
+ annotation["segmentation"] = [
310
+ p.reshape(-1) for p in transforms.apply_polygons(polygons)
311
+ ]
312
+ elif isinstance(segm, dict):
313
+ # RLE
314
+ mask = mask_util.decode(segm)
315
+ mask = transforms.apply_segmentation(mask)
316
+ assert tuple(mask.shape[:2]) == image_size
317
+ annotation["segmentation"] = mask
318
+ else:
319
+ raise ValueError(
320
+ "Cannot transform segmentation of type '{}'!"
321
+ "Supported types are: polygons as list[list[float] or ndarray],"
322
+ " COCO-style RLE as a dict.".format(type(segm))
323
+ )
324
+
325
+ if "keypoints" in annotation:
326
+ keypoints = transform_keypoint_annotations(
327
+ annotation["keypoints"], transforms, image_size, keypoint_hflip_indices
328
+ )
329
+ annotation["keypoints"] = keypoints
330
+
331
+ return annotation
332
+
333
+
334
+ def transform_keypoint_annotations(keypoints, transforms, image_size, keypoint_hflip_indices=None):
335
+ """
336
+ Transform keypoint annotations of an image.
337
+ If a keypoint is transformed out of image boundary, it will be marked "unlabeled" (visibility=0)
338
+
339
+ Args:
340
+ keypoints (list[float]): Nx3 float in Detectron2's Dataset format.
341
+ Each point is represented by (x, y, visibility).
342
+ transforms (TransformList):
343
+ image_size (tuple): the height, width of the transformed image
344
+ keypoint_hflip_indices (ndarray[int]): see `create_keypoint_hflip_indices`.
345
+ When `transforms` includes horizontal flip, will use the index
346
+ mapping to flip keypoints.
347
+ """
348
+ # (N*3,) -> (N, 3)
349
+ keypoints = np.asarray(keypoints, dtype="float64").reshape(-1, 3)
350
+ keypoints_xy = transforms.apply_coords(keypoints[:, :2])
351
+
352
+ # Set all out-of-boundary points to "unlabeled"
353
+ inside = (keypoints_xy >= np.array([0, 0])) & (keypoints_xy <= np.array(image_size[::-1]))
354
+ inside = inside.all(axis=1)
355
+ keypoints[:, :2] = keypoints_xy
356
+ keypoints[:, 2][~inside] = 0
357
+
358
+ # This assumes that HorizFlipTransform is the only one that does flip
359
+ do_hflip = sum(isinstance(t, T.HFlipTransform) for t in transforms.transforms) % 2 == 1
360
+
361
+ # Alternative way: check if probe points was horizontally flipped.
362
+ # probe = np.asarray([[0.0, 0.0], [image_width, 0.0]])
363
+ # probe_aug = transforms.apply_coords(probe.copy())
364
+ # do_hflip = np.sign(probe[1][0] - probe[0][0]) != np.sign(probe_aug[1][0] - probe_aug[0][0]) # noqa
365
+
366
+ # If flipped, swap each keypoint with its opposite-handed equivalent
367
+ if do_hflip:
368
+ if keypoint_hflip_indices is None:
369
+ raise ValueError("Cannot flip keypoints without providing flip indices!")
370
+ if len(keypoints) != len(keypoint_hflip_indices):
371
+ raise ValueError(
372
+ "Keypoint data has {} points, but metadata "
373
+ "contains {} points!".format(len(keypoints), len(keypoint_hflip_indices))
374
+ )
375
+ keypoints = keypoints[np.asarray(keypoint_hflip_indices, dtype=np.int32), :]
376
+
377
+ # Maintain COCO convention that if visibility == 0 (unlabeled), then x, y = 0
378
+ keypoints[keypoints[:, 2] == 0] = 0
379
+ return keypoints
380
+
381
+
382
+ def annotations_to_instances(annos, image_size, mask_format="polygon"):
383
+ """
384
+ Create an :class:`Instances` object used by the models,
385
+ from instance annotations in the dataset dict.
386
+
387
+ Args:
388
+ annos (list[dict]): a list of instance annotations in one image, each
389
+ element for one instance.
390
+ image_size (tuple): height, width
391
+
392
+ Returns:
393
+ Instances:
394
+ It will contain fields "gt_boxes", "gt_classes",
395
+ "gt_masks", "gt_keypoints", if they can be obtained from `annos`.
396
+ This is the format that builtin models expect.
397
+ """
398
+ boxes = (
399
+ np.stack(
400
+ [BoxMode.convert(obj["bbox"], obj["bbox_mode"], BoxMode.XYXY_ABS) for obj in annos]
401
+ )
402
+ if len(annos)
403
+ else np.zeros((0, 4))
404
+ )
405
+ target = Instances(image_size)
406
+ target.gt_boxes = Boxes(boxes)
407
+
408
+ classes = [int(obj["category_id"]) for obj in annos]
409
+ classes = torch.tensor(classes, dtype=torch.int64)
410
+ target.gt_classes = classes
411
+
412
+ if len(annos) and "segmentation" in annos[0]:
413
+ segms = [obj["segmentation"] for obj in annos]
414
+ if mask_format == "polygon":
415
+ try:
416
+ masks = PolygonMasks(segms)
417
+ except ValueError as e:
418
+ raise ValueError(
419
+ "Failed to use mask_format=='polygon' from the given annotations!"
420
+ ) from e
421
+ else:
422
+ assert mask_format == "bitmask", mask_format
423
+ masks = []
424
+ for segm in segms:
425
+ if isinstance(segm, list):
426
+ # polygon
427
+ masks.append(polygons_to_bitmask(segm, *image_size))
428
+ elif isinstance(segm, dict):
429
+ # COCO RLE
430
+ masks.append(mask_util.decode(segm))
431
+ elif isinstance(segm, np.ndarray):
432
+ assert segm.ndim == 2, "Expect segmentation of 2 dimensions, got {}.".format(
433
+ segm.ndim
434
+ )
435
+ # mask array
436
+ masks.append(segm)
437
+ else:
438
+ raise ValueError(
439
+ "Cannot convert segmentation of type '{}' to BitMasks!"
440
+ "Supported types are: polygons as list[list[float] or ndarray],"
441
+ " COCO-style RLE as a dict, or a binary segmentation mask "
442
+ " in a 2D numpy array of shape HxW.".format(type(segm))
443
+ )
444
+ # torch.from_numpy does not support array with negative stride.
445
+ masks = BitMasks(
446
+ torch.stack([torch.from_numpy(np.ascontiguousarray(x)) for x in masks])
447
+ )
448
+ target.gt_masks = masks
449
+
450
+ if len(annos) and "keypoints" in annos[0]:
451
+ kpts = [obj.get("keypoints", []) for obj in annos]
452
+ target.gt_keypoints = Keypoints(kpts)
453
+
454
+ return target
455
+
456
+
457
+ def annotations_to_instances_rotated(annos, image_size):
458
+ """
459
+ Create an :class:`Instances` object used by the models,
460
+ from instance annotations in the dataset dict.
461
+ Compared to `annotations_to_instances`, this function is for rotated boxes only
462
+
463
+ Args:
464
+ annos (list[dict]): a list of instance annotations in one image, each
465
+ element for one instance.
466
+ image_size (tuple): height, width
467
+
468
+ Returns:
469
+ Instances:
470
+ Containing fields "gt_boxes", "gt_classes",
471
+ if they can be obtained from `annos`.
472
+ This is the format that builtin models expect.
473
+ """
474
+ boxes = [obj["bbox"] for obj in annos]
475
+ target = Instances(image_size)
476
+ boxes = target.gt_boxes = RotatedBoxes(boxes)
477
+ boxes.clip(image_size)
478
+
479
+ classes = [obj["category_id"] for obj in annos]
480
+ classes = torch.tensor(classes, dtype=torch.int64)
481
+ target.gt_classes = classes
482
+
483
+ return target
484
+
485
+
486
+ def filter_empty_instances(
487
+ instances, by_box=True, by_mask=True, box_threshold=1e-5, return_mask=False
488
+ ):
489
+ """
490
+ Filter out empty instances in an `Instances` object.
491
+
492
+ Args:
493
+ instances (Instances):
494
+ by_box (bool): whether to filter out instances with empty boxes
495
+ by_mask (bool): whether to filter out instances with empty masks
496
+ box_threshold (float): minimum width and height to be considered non-empty
497
+ return_mask (bool): whether to return boolean mask of filtered instances
498
+
499
+ Returns:
500
+ Instances: the filtered instances.
501
+ tensor[bool], optional: boolean mask of filtered instances
502
+ """
503
+ assert by_box or by_mask
504
+ r = []
505
+ if by_box:
506
+ r.append(instances.gt_boxes.nonempty(threshold=box_threshold))
507
+ if instances.has("gt_masks") and by_mask:
508
+ r.append(instances.gt_masks.nonempty())
509
+
510
+ # TODO: can also filter visible keypoints
511
+
512
+ if not r:
513
+ return instances
514
+ m = r[0]
515
+ for x in r[1:]:
516
+ m = m & x
517
+ if return_mask:
518
+ return instances[m], m
519
+ return instances[m]
520
+
521
+
522
+ def create_keypoint_hflip_indices(dataset_names: Union[str, List[str]]) -> List[int]:
523
+ """
524
+ Args:
525
+ dataset_names: list of dataset names
526
+
527
+ Returns:
528
+ list[int]: a list of size=#keypoints, storing the
529
+ horizontally-flipped keypoint indices.
530
+ """
531
+ if isinstance(dataset_names, str):
532
+ dataset_names = [dataset_names]
533
+
534
+ check_metadata_consistency("keypoint_names", dataset_names)
535
+ check_metadata_consistency("keypoint_flip_map", dataset_names)
536
+
537
+ meta = MetadataCatalog.get(dataset_names[0])
538
+ names = meta.keypoint_names
539
+ # TODO flip -> hflip
540
+ flip_map = dict(meta.keypoint_flip_map)
541
+ flip_map.update({v: k for k, v in flip_map.items()})
542
+ flipped_names = [i if i not in flip_map else flip_map[i] for i in names]
543
+ flip_indices = [names.index(i) for i in flipped_names]
544
+ return flip_indices
545
+
546
+
547
+ def get_fed_loss_cls_weights(dataset_names: Union[str, List[str]], freq_weight_power=1.0):
548
+ """
549
+ Get frequency weight for each class sorted by class id.
550
+ We now calcualte freqency weight using image_count to the power freq_weight_power.
551
+
552
+ Args:
553
+ dataset_names: list of dataset names
554
+ freq_weight_power: power value
555
+ """
556
+ if isinstance(dataset_names, str):
557
+ dataset_names = [dataset_names]
558
+
559
+ check_metadata_consistency("class_image_count", dataset_names)
560
+
561
+ meta = MetadataCatalog.get(dataset_names[0])
562
+ class_freq_meta = meta.class_image_count
563
+ class_freq = torch.tensor(
564
+ [c["image_count"] for c in sorted(class_freq_meta, key=lambda x: x["id"])]
565
+ )
566
+ class_freq_weight = class_freq.float() ** freq_weight_power
567
+ return class_freq_weight
568
+
569
+
570
+ def gen_crop_transform_with_instance(crop_size, image_size, instance):
571
+ """
572
+ Generate a CropTransform so that the cropping region contains
573
+ the center of the given instance.
574
+
575
+ Args:
576
+ crop_size (tuple): h, w in pixels
577
+ image_size (tuple): h, w
578
+ instance (dict): an annotation dict of one instance, in Detectron2's
579
+ dataset format.
580
+ """
581
+ crop_size = np.asarray(crop_size, dtype=np.int32)
582
+ bbox = BoxMode.convert(instance["bbox"], instance["bbox_mode"], BoxMode.XYXY_ABS)
583
+ center_yx = (bbox[1] + bbox[3]) * 0.5, (bbox[0] + bbox[2]) * 0.5
584
+ assert (
585
+ image_size[0] >= center_yx[0] and image_size[1] >= center_yx[1]
586
+ ), "The annotation bounding box is outside of the image!"
587
+ assert (
588
+ image_size[0] >= crop_size[0] and image_size[1] >= crop_size[1]
589
+ ), "Crop size is larger than image size!"
590
+
591
+ min_yx = np.maximum(np.floor(center_yx).astype(np.int32) - crop_size, 0)
592
+ max_yx = np.maximum(np.asarray(image_size, dtype=np.int32) - crop_size, 0)
593
+ max_yx = np.minimum(max_yx, np.ceil(center_yx).astype(np.int32))
594
+
595
+ y0 = np.random.randint(min_yx[0], max_yx[0] + 1)
596
+ x0 = np.random.randint(min_yx[1], max_yx[1] + 1)
597
+ return T.CropTransform(x0, y0, crop_size[1], crop_size[0])
598
+
599
+
600
+ def check_metadata_consistency(key, dataset_names):
601
+ """
602
+ Check that the datasets have consistent metadata.
603
+
604
+ Args:
605
+ key (str): a metadata key
606
+ dataset_names (list[str]): a list of dataset names
607
+
608
+ Raises:
609
+ AttributeError: if the key does not exist in the metadata
610
+ ValueError: if the given datasets do not have the same metadata values defined by key
611
+ """
612
+ if len(dataset_names) == 0:
613
+ return
614
+ logger = logging.getLogger(__name__)
615
+ entries_per_dataset = [getattr(MetadataCatalog.get(d), key) for d in dataset_names]
616
+ for idx, entry in enumerate(entries_per_dataset):
617
+ if entry != entries_per_dataset[0]:
618
+ logger.error(
619
+ "Metadata '{}' for dataset '{}' is '{}'".format(key, dataset_names[idx], str(entry))
620
+ )
621
+ logger.error(
622
+ "Metadata '{}' for dataset '{}' is '{}'".format(
623
+ key, dataset_names[0], str(entries_per_dataset[0])
624
+ )
625
+ )
626
+ raise ValueError("Datasets have different metadata '{}'!".format(key))
627
+
628
+
629
+ def build_augmentation(cfg, is_train):
630
+ """
631
+ Create a list of default :class:`Augmentation` from config.
632
+ Now it includes resizing and flipping.
633
+
634
+ Returns:
635
+ list[Augmentation]
636
+ """
637
+ if is_train:
638
+ min_size = cfg.INPUT.MIN_SIZE_TRAIN
639
+ max_size = cfg.INPUT.MAX_SIZE_TRAIN
640
+ sample_style = cfg.INPUT.MIN_SIZE_TRAIN_SAMPLING
641
+ else:
642
+ min_size = cfg.INPUT.MIN_SIZE_TEST
643
+ max_size = cfg.INPUT.MAX_SIZE_TEST
644
+ sample_style = "choice"
645
+ augmentation = [T.ResizeShortestEdge(min_size, max_size, sample_style)]
646
+ if is_train and cfg.INPUT.RANDOM_FLIP != "none":
647
+ augmentation.append(
648
+ T.RandomFlip(
649
+ horizontal=cfg.INPUT.RANDOM_FLIP == "horizontal",
650
+ vertical=cfg.INPUT.RANDOM_FLIP == "vertical",
651
+ )
652
+ )
653
+ return augmentation
654
+
655
+
656
+ build_transform_gen = build_augmentation
657
+ """
658
+ Alias for backward-compatibility.
659
+ """
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/samplers/grouped_batch_sampler.py ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import numpy as np
3
+ from torch.utils.data.sampler import BatchSampler, Sampler
4
+
5
+
6
+ class GroupedBatchSampler(BatchSampler):
7
+ """
8
+ Wraps another sampler to yield a mini-batch of indices.
9
+ It enforces that the batch only contain elements from the same group.
10
+ It also tries to provide mini-batches which follows an ordering which is
11
+ as close as possible to the ordering from the original sampler.
12
+ """
13
+
14
+ def __init__(self, sampler, group_ids, batch_size):
15
+ """
16
+ Args:
17
+ sampler (Sampler): Base sampler.
18
+ group_ids (list[int]): If the sampler produces indices in range [0, N),
19
+ `group_ids` must be a list of `N` ints which contains the group id of each sample.
20
+ The group ids must be a set of integers in the range [0, num_groups).
21
+ batch_size (int): Size of mini-batch.
22
+ """
23
+ if not isinstance(sampler, Sampler):
24
+ raise ValueError(
25
+ "sampler should be an instance of "
26
+ "torch.utils.data.Sampler, but got sampler={}".format(sampler)
27
+ )
28
+ self.sampler = sampler
29
+ self.group_ids = np.asarray(group_ids)
30
+ assert self.group_ids.ndim == 1
31
+ self.batch_size = batch_size
32
+ groups = np.unique(self.group_ids).tolist()
33
+
34
+ # buffer the indices of each group until batch size is reached
35
+ self.buffer_per_group = {k: [] for k in groups}
36
+
37
+ def __iter__(self):
38
+ for idx in self.sampler:
39
+ group_id = self.group_ids[idx]
40
+ group_buffer = self.buffer_per_group[group_id]
41
+ group_buffer.append(idx)
42
+ if len(group_buffer) == self.batch_size:
43
+ yield group_buffer[:] # yield a copy of the list
44
+ del group_buffer[:]
45
+
46
+ def __len__(self):
47
+ raise NotImplementedError("len() of GroupedBatchSampler is not well-defined.")
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/__init__.py ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from fvcore.transforms.transform import Transform, TransformList # order them first
3
+ from fvcore.transforms.transform import *
4
+ from .transform import *
5
+ from .augmentation import *
6
+ from .augmentation_impl import *
7
+
8
+ __all__ = [k for k in globals().keys() if not k.startswith("_")]
9
+
10
+
11
+ from annotator.oneformer.detectron2.utils.env import fixup_module_metadata
12
+
13
+ fixup_module_metadata(__name__, globals(), __all__)
14
+ del fixup_module_metadata
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/augmentation.py ADDED
@@ -0,0 +1,380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ import inspect
5
+ import numpy as np
6
+ import pprint
7
+ from typing import Any, List, Optional, Tuple, Union
8
+ from fvcore.transforms.transform import Transform, TransformList
9
+
10
+ """
11
+ See "Data Augmentation" tutorial for an overview of the system:
12
+ https://detectron2.readthedocs.io/tutorials/augmentation.html
13
+ """
14
+
15
+
16
+ __all__ = [
17
+ "Augmentation",
18
+ "AugmentationList",
19
+ "AugInput",
20
+ "TransformGen",
21
+ "apply_transform_gens",
22
+ "StandardAugInput",
23
+ "apply_augmentations",
24
+ ]
25
+
26
+
27
+ def _check_img_dtype(img):
28
+ assert isinstance(img, np.ndarray), "[Augmentation] Needs an numpy array, but got a {}!".format(
29
+ type(img)
30
+ )
31
+ assert not isinstance(img.dtype, np.integer) or (
32
+ img.dtype == np.uint8
33
+ ), "[Augmentation] Got image of type {}, use uint8 or floating points instead!".format(
34
+ img.dtype
35
+ )
36
+ assert img.ndim in [2, 3], img.ndim
37
+
38
+
39
+ def _get_aug_input_args(aug, aug_input) -> List[Any]:
40
+ """
41
+ Get the arguments to be passed to ``aug.get_transform`` from the input ``aug_input``.
42
+ """
43
+ if aug.input_args is None:
44
+ # Decide what attributes are needed automatically
45
+ prms = list(inspect.signature(aug.get_transform).parameters.items())
46
+ # The default behavior is: if there is one parameter, then its "image"
47
+ # (work automatically for majority of use cases, and also avoid BC breaking),
48
+ # Otherwise, use the argument names.
49
+ if len(prms) == 1:
50
+ names = ("image",)
51
+ else:
52
+ names = []
53
+ for name, prm in prms:
54
+ if prm.kind in (
55
+ inspect.Parameter.VAR_POSITIONAL,
56
+ inspect.Parameter.VAR_KEYWORD,
57
+ ):
58
+ raise TypeError(
59
+ f""" \
60
+ The default implementation of `{type(aug)}.__call__` does not allow \
61
+ `{type(aug)}.get_transform` to use variable-length arguments (*args, **kwargs)! \
62
+ If arguments are unknown, reimplement `__call__` instead. \
63
+ """
64
+ )
65
+ names.append(name)
66
+ aug.input_args = tuple(names)
67
+
68
+ args = []
69
+ for f in aug.input_args:
70
+ try:
71
+ args.append(getattr(aug_input, f))
72
+ except AttributeError as e:
73
+ raise AttributeError(
74
+ f"{type(aug)}.get_transform needs input attribute '{f}', "
75
+ f"but it is not an attribute of {type(aug_input)}!"
76
+ ) from e
77
+ return args
78
+
79
+
80
+ class Augmentation:
81
+ """
82
+ Augmentation defines (often random) policies/strategies to generate :class:`Transform`
83
+ from data. It is often used for pre-processing of input data.
84
+
85
+ A "policy" that generates a :class:`Transform` may, in the most general case,
86
+ need arbitrary information from input data in order to determine what transforms
87
+ to apply. Therefore, each :class:`Augmentation` instance defines the arguments
88
+ needed by its :meth:`get_transform` method. When called with the positional arguments,
89
+ the :meth:`get_transform` method executes the policy.
90
+
91
+ Note that :class:`Augmentation` defines the policies to create a :class:`Transform`,
92
+ but not how to execute the actual transform operations to those data.
93
+ Its :meth:`__call__` method will use :meth:`AugInput.transform` to execute the transform.
94
+
95
+ The returned `Transform` object is meant to describe deterministic transformation, which means
96
+ it can be re-applied on associated data, e.g. the geometry of an image and its segmentation
97
+ masks need to be transformed together.
98
+ (If such re-application is not needed, then determinism is not a crucial requirement.)
99
+ """
100
+
101
+ input_args: Optional[Tuple[str]] = None
102
+ """
103
+ Stores the attribute names needed by :meth:`get_transform`, e.g. ``("image", "sem_seg")``.
104
+ By default, it is just a tuple of argument names in :meth:`self.get_transform`, which often only
105
+ contain "image". As long as the argument name convention is followed, there is no need for
106
+ users to touch this attribute.
107
+ """
108
+
109
+ def _init(self, params=None):
110
+ if params:
111
+ for k, v in params.items():
112
+ if k != "self" and not k.startswith("_"):
113
+ setattr(self, k, v)
114
+
115
+ def get_transform(self, *args) -> Transform:
116
+ """
117
+ Execute the policy based on input data, and decide what transform to apply to inputs.
118
+
119
+ Args:
120
+ args: Any fixed-length positional arguments. By default, the name of the arguments
121
+ should exist in the :class:`AugInput` to be used.
122
+
123
+ Returns:
124
+ Transform: Returns the deterministic transform to apply to the input.
125
+
126
+ Examples:
127
+ ::
128
+ class MyAug:
129
+ # if a policy needs to know both image and semantic segmentation
130
+ def get_transform(image, sem_seg) -> T.Transform:
131
+ pass
132
+ tfm: Transform = MyAug().get_transform(image, sem_seg)
133
+ new_image = tfm.apply_image(image)
134
+
135
+ Notes:
136
+ Users can freely use arbitrary new argument names in custom
137
+ :meth:`get_transform` method, as long as they are available in the
138
+ input data. In detectron2 we use the following convention:
139
+
140
+ * image: (H,W) or (H,W,C) ndarray of type uint8 in range [0, 255], or
141
+ floating point in range [0, 1] or [0, 255].
142
+ * boxes: (N,4) ndarray of float32. It represents the instance bounding boxes
143
+ of N instances. Each is in XYXY format in unit of absolute coordinates.
144
+ * sem_seg: (H,W) ndarray of type uint8. Each element is an integer label of pixel.
145
+
146
+ We do not specify convention for other types and do not include builtin
147
+ :class:`Augmentation` that uses other types in detectron2.
148
+ """
149
+ raise NotImplementedError
150
+
151
+ def __call__(self, aug_input) -> Transform:
152
+ """
153
+ Augment the given `aug_input` **in-place**, and return the transform that's used.
154
+
155
+ This method will be called to apply the augmentation. In most augmentation, it
156
+ is enough to use the default implementation, which calls :meth:`get_transform`
157
+ using the inputs. But a subclass can overwrite it to have more complicated logic.
158
+
159
+ Args:
160
+ aug_input (AugInput): an object that has attributes needed by this augmentation
161
+ (defined by ``self.get_transform``). Its ``transform`` method will be called
162
+ to in-place transform it.
163
+
164
+ Returns:
165
+ Transform: the transform that is applied on the input.
166
+ """
167
+ args = _get_aug_input_args(self, aug_input)
168
+ tfm = self.get_transform(*args)
169
+ assert isinstance(tfm, (Transform, TransformList)), (
170
+ f"{type(self)}.get_transform must return an instance of Transform! "
171
+ f"Got {type(tfm)} instead."
172
+ )
173
+ aug_input.transform(tfm)
174
+ return tfm
175
+
176
+ def _rand_range(self, low=1.0, high=None, size=None):
177
+ """
178
+ Uniform float random number between low and high.
179
+ """
180
+ if high is None:
181
+ low, high = 0, low
182
+ if size is None:
183
+ size = []
184
+ return np.random.uniform(low, high, size)
185
+
186
+ def __repr__(self):
187
+ """
188
+ Produce something like:
189
+ "MyAugmentation(field1={self.field1}, field2={self.field2})"
190
+ """
191
+ try:
192
+ sig = inspect.signature(self.__init__)
193
+ classname = type(self).__name__
194
+ argstr = []
195
+ for name, param in sig.parameters.items():
196
+ assert (
197
+ param.kind != param.VAR_POSITIONAL and param.kind != param.VAR_KEYWORD
198
+ ), "The default __repr__ doesn't support *args or **kwargs"
199
+ assert hasattr(self, name), (
200
+ "Attribute {} not found! "
201
+ "Default __repr__ only works if attributes match the constructor.".format(name)
202
+ )
203
+ attr = getattr(self, name)
204
+ default = param.default
205
+ if default is attr:
206
+ continue
207
+ attr_str = pprint.pformat(attr)
208
+ if "\n" in attr_str:
209
+ # don't show it if pformat decides to use >1 lines
210
+ attr_str = "..."
211
+ argstr.append("{}={}".format(name, attr_str))
212
+ return "{}({})".format(classname, ", ".join(argstr))
213
+ except AssertionError:
214
+ return super().__repr__()
215
+
216
+ __str__ = __repr__
217
+
218
+
219
+ class _TransformToAug(Augmentation):
220
+ def __init__(self, tfm: Transform):
221
+ self.tfm = tfm
222
+
223
+ def get_transform(self, *args):
224
+ return self.tfm
225
+
226
+ def __repr__(self):
227
+ return repr(self.tfm)
228
+
229
+ __str__ = __repr__
230
+
231
+
232
+ def _transform_to_aug(tfm_or_aug):
233
+ """
234
+ Wrap Transform into Augmentation.
235
+ Private, used internally to implement augmentations.
236
+ """
237
+ assert isinstance(tfm_or_aug, (Transform, Augmentation)), tfm_or_aug
238
+ if isinstance(tfm_or_aug, Augmentation):
239
+ return tfm_or_aug
240
+ else:
241
+ return _TransformToAug(tfm_or_aug)
242
+
243
+
244
+ class AugmentationList(Augmentation):
245
+ """
246
+ Apply a sequence of augmentations.
247
+
248
+ It has ``__call__`` method to apply the augmentations.
249
+
250
+ Note that :meth:`get_transform` method is impossible (will throw error if called)
251
+ for :class:`AugmentationList`, because in order to apply a sequence of augmentations,
252
+ the kth augmentation must be applied first, to provide inputs needed by the (k+1)th
253
+ augmentation.
254
+ """
255
+
256
+ def __init__(self, augs):
257
+ """
258
+ Args:
259
+ augs (list[Augmentation or Transform]):
260
+ """
261
+ super().__init__()
262
+ self.augs = [_transform_to_aug(x) for x in augs]
263
+
264
+ def __call__(self, aug_input) -> TransformList:
265
+ tfms = []
266
+ for x in self.augs:
267
+ tfm = x(aug_input)
268
+ tfms.append(tfm)
269
+ return TransformList(tfms)
270
+
271
+ def __repr__(self):
272
+ msgs = [str(x) for x in self.augs]
273
+ return "AugmentationList[{}]".format(", ".join(msgs))
274
+
275
+ __str__ = __repr__
276
+
277
+
278
+ class AugInput:
279
+ """
280
+ Input that can be used with :meth:`Augmentation.__call__`.
281
+ This is a standard implementation for the majority of use cases.
282
+ This class provides the standard attributes **"image", "boxes", "sem_seg"**
283
+ defined in :meth:`__init__` and they may be needed by different augmentations.
284
+ Most augmentation policies do not need attributes beyond these three.
285
+
286
+ After applying augmentations to these attributes (using :meth:`AugInput.transform`),
287
+ the returned transforms can then be used to transform other data structures that users have.
288
+
289
+ Examples:
290
+ ::
291
+ input = AugInput(image, boxes=boxes)
292
+ tfms = augmentation(input)
293
+ transformed_image = input.image
294
+ transformed_boxes = input.boxes
295
+ transformed_other_data = tfms.apply_other(other_data)
296
+
297
+ An extended project that works with new data types may implement augmentation policies
298
+ that need other inputs. An algorithm may need to transform inputs in a way different
299
+ from the standard approach defined in this class. In those rare situations, users can
300
+ implement a class similar to this class, that satify the following condition:
301
+
302
+ * The input must provide access to these data in the form of attribute access
303
+ (``getattr``). For example, if an :class:`Augmentation` to be applied needs "image"
304
+ and "sem_seg" arguments, its input must have the attribute "image" and "sem_seg".
305
+ * The input must have a ``transform(tfm: Transform) -> None`` method which
306
+ in-place transforms all its attributes.
307
+ """
308
+
309
+ # TODO maybe should support more builtin data types here
310
+ def __init__(
311
+ self,
312
+ image: np.ndarray,
313
+ *,
314
+ boxes: Optional[np.ndarray] = None,
315
+ sem_seg: Optional[np.ndarray] = None,
316
+ ):
317
+ """
318
+ Args:
319
+ image (ndarray): (H,W) or (H,W,C) ndarray of type uint8 in range [0, 255], or
320
+ floating point in range [0, 1] or [0, 255]. The meaning of C is up
321
+ to users.
322
+ boxes (ndarray or None): Nx4 float32 boxes in XYXY_ABS mode
323
+ sem_seg (ndarray or None): HxW uint8 semantic segmentation mask. Each element
324
+ is an integer label of pixel.
325
+ """
326
+ _check_img_dtype(image)
327
+ self.image = image
328
+ self.boxes = boxes
329
+ self.sem_seg = sem_seg
330
+
331
+ def transform(self, tfm: Transform) -> None:
332
+ """
333
+ In-place transform all attributes of this class.
334
+
335
+ By "in-place", it means after calling this method, accessing an attribute such
336
+ as ``self.image`` will return transformed data.
337
+ """
338
+ self.image = tfm.apply_image(self.image)
339
+ if self.boxes is not None:
340
+ self.boxes = tfm.apply_box(self.boxes)
341
+ if self.sem_seg is not None:
342
+ self.sem_seg = tfm.apply_segmentation(self.sem_seg)
343
+
344
+ def apply_augmentations(
345
+ self, augmentations: List[Union[Augmentation, Transform]]
346
+ ) -> TransformList:
347
+ """
348
+ Equivalent of ``AugmentationList(augmentations)(self)``
349
+ """
350
+ return AugmentationList(augmentations)(self)
351
+
352
+
353
+ def apply_augmentations(augmentations: List[Union[Transform, Augmentation]], inputs):
354
+ """
355
+ Use ``T.AugmentationList(augmentations)(inputs)`` instead.
356
+ """
357
+ if isinstance(inputs, np.ndarray):
358
+ # handle the common case of image-only Augmentation, also for backward compatibility
359
+ image_only = True
360
+ inputs = AugInput(inputs)
361
+ else:
362
+ image_only = False
363
+ tfms = inputs.apply_augmentations(augmentations)
364
+ return inputs.image if image_only else inputs, tfms
365
+
366
+
367
+ apply_transform_gens = apply_augmentations
368
+ """
369
+ Alias for backward-compatibility.
370
+ """
371
+
372
+ TransformGen = Augmentation
373
+ """
374
+ Alias for Augmentation, since it is something that generates :class:`Transform`s
375
+ """
376
+
377
+ StandardAugInput = AugInput
378
+ """
379
+ Alias for compatibility. It's not worth the complexity to have two classes.
380
+ """
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/augmentation_impl.py ADDED
@@ -0,0 +1,736 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+ """
4
+ Implement many useful :class:`Augmentation`.
5
+ """
6
+ import numpy as np
7
+ import sys
8
+ from numpy import random
9
+ from typing import Tuple
10
+ import torch
11
+ from fvcore.transforms.transform import (
12
+ BlendTransform,
13
+ CropTransform,
14
+ HFlipTransform,
15
+ NoOpTransform,
16
+ PadTransform,
17
+ Transform,
18
+ TransformList,
19
+ VFlipTransform,
20
+ )
21
+ from PIL import Image
22
+
23
+ from annotator.oneformer.detectron2.structures import Boxes, pairwise_iou
24
+
25
+ from .augmentation import Augmentation, _transform_to_aug
26
+ from .transform import ExtentTransform, ResizeTransform, RotationTransform
27
+
28
+ __all__ = [
29
+ "FixedSizeCrop",
30
+ "RandomApply",
31
+ "RandomBrightness",
32
+ "RandomContrast",
33
+ "RandomCrop",
34
+ "RandomExtent",
35
+ "RandomFlip",
36
+ "RandomSaturation",
37
+ "RandomLighting",
38
+ "RandomRotation",
39
+ "Resize",
40
+ "ResizeScale",
41
+ "ResizeShortestEdge",
42
+ "RandomCrop_CategoryAreaConstraint",
43
+ "RandomResize",
44
+ "MinIoURandomCrop",
45
+ ]
46
+
47
+
48
+ class RandomApply(Augmentation):
49
+ """
50
+ Randomly apply an augmentation with a given probability.
51
+ """
52
+
53
+ def __init__(self, tfm_or_aug, prob=0.5):
54
+ """
55
+ Args:
56
+ tfm_or_aug (Transform, Augmentation): the transform or augmentation
57
+ to be applied. It can either be a `Transform` or `Augmentation`
58
+ instance.
59
+ prob (float): probability between 0.0 and 1.0 that
60
+ the wrapper transformation is applied
61
+ """
62
+ super().__init__()
63
+ self.aug = _transform_to_aug(tfm_or_aug)
64
+ assert 0.0 <= prob <= 1.0, f"Probablity must be between 0.0 and 1.0 (given: {prob})"
65
+ self.prob = prob
66
+
67
+ def get_transform(self, *args):
68
+ do = self._rand_range() < self.prob
69
+ if do:
70
+ return self.aug.get_transform(*args)
71
+ else:
72
+ return NoOpTransform()
73
+
74
+ def __call__(self, aug_input):
75
+ do = self._rand_range() < self.prob
76
+ if do:
77
+ return self.aug(aug_input)
78
+ else:
79
+ return NoOpTransform()
80
+
81
+
82
+ class RandomFlip(Augmentation):
83
+ """
84
+ Flip the image horizontally or vertically with the given probability.
85
+ """
86
+
87
+ def __init__(self, prob=0.5, *, horizontal=True, vertical=False):
88
+ """
89
+ Args:
90
+ prob (float): probability of flip.
91
+ horizontal (boolean): whether to apply horizontal flipping
92
+ vertical (boolean): whether to apply vertical flipping
93
+ """
94
+ super().__init__()
95
+
96
+ if horizontal and vertical:
97
+ raise ValueError("Cannot do both horiz and vert. Please use two Flip instead.")
98
+ if not horizontal and not vertical:
99
+ raise ValueError("At least one of horiz or vert has to be True!")
100
+ self._init(locals())
101
+
102
+ def get_transform(self, image):
103
+ h, w = image.shape[:2]
104
+ do = self._rand_range() < self.prob
105
+ if do:
106
+ if self.horizontal:
107
+ return HFlipTransform(w)
108
+ elif self.vertical:
109
+ return VFlipTransform(h)
110
+ else:
111
+ return NoOpTransform()
112
+
113
+
114
+ class Resize(Augmentation):
115
+ """Resize image to a fixed target size"""
116
+
117
+ def __init__(self, shape, interp=Image.BILINEAR):
118
+ """
119
+ Args:
120
+ shape: (h, w) tuple or a int
121
+ interp: PIL interpolation method
122
+ """
123
+ if isinstance(shape, int):
124
+ shape = (shape, shape)
125
+ shape = tuple(shape)
126
+ self._init(locals())
127
+
128
+ def get_transform(self, image):
129
+ return ResizeTransform(
130
+ image.shape[0], image.shape[1], self.shape[0], self.shape[1], self.interp
131
+ )
132
+
133
+
134
+ class ResizeShortestEdge(Augmentation):
135
+ """
136
+ Resize the image while keeping the aspect ratio unchanged.
137
+ It attempts to scale the shorter edge to the given `short_edge_length`,
138
+ as long as the longer edge does not exceed `max_size`.
139
+ If `max_size` is reached, then downscale so that the longer edge does not exceed max_size.
140
+ """
141
+
142
+ @torch.jit.unused
143
+ def __init__(
144
+ self, short_edge_length, max_size=sys.maxsize, sample_style="range", interp=Image.BILINEAR
145
+ ):
146
+ """
147
+ Args:
148
+ short_edge_length (list[int]): If ``sample_style=="range"``,
149
+ a [min, max] interval from which to sample the shortest edge length.
150
+ If ``sample_style=="choice"``, a list of shortest edge lengths to sample from.
151
+ max_size (int): maximum allowed longest edge length.
152
+ sample_style (str): either "range" or "choice".
153
+ """
154
+ super().__init__()
155
+ assert sample_style in ["range", "choice"], sample_style
156
+
157
+ self.is_range = sample_style == "range"
158
+ if isinstance(short_edge_length, int):
159
+ short_edge_length = (short_edge_length, short_edge_length)
160
+ if self.is_range:
161
+ assert len(short_edge_length) == 2, (
162
+ "short_edge_length must be two values using 'range' sample style."
163
+ f" Got {short_edge_length}!"
164
+ )
165
+ self._init(locals())
166
+
167
+ @torch.jit.unused
168
+ def get_transform(self, image):
169
+ h, w = image.shape[:2]
170
+ if self.is_range:
171
+ size = np.random.randint(self.short_edge_length[0], self.short_edge_length[1] + 1)
172
+ else:
173
+ size = np.random.choice(self.short_edge_length)
174
+ if size == 0:
175
+ return NoOpTransform()
176
+
177
+ newh, neww = ResizeShortestEdge.get_output_shape(h, w, size, self.max_size)
178
+ return ResizeTransform(h, w, newh, neww, self.interp)
179
+
180
+ @staticmethod
181
+ def get_output_shape(
182
+ oldh: int, oldw: int, short_edge_length: int, max_size: int
183
+ ) -> Tuple[int, int]:
184
+ """
185
+ Compute the output size given input size and target short edge length.
186
+ """
187
+ h, w = oldh, oldw
188
+ size = short_edge_length * 1.0
189
+ scale = size / min(h, w)
190
+ if h < w:
191
+ newh, neww = size, scale * w
192
+ else:
193
+ newh, neww = scale * h, size
194
+ if max(newh, neww) > max_size:
195
+ scale = max_size * 1.0 / max(newh, neww)
196
+ newh = newh * scale
197
+ neww = neww * scale
198
+ neww = int(neww + 0.5)
199
+ newh = int(newh + 0.5)
200
+ return (newh, neww)
201
+
202
+
203
+ class ResizeScale(Augmentation):
204
+ """
205
+ Takes target size as input and randomly scales the given target size between `min_scale`
206
+ and `max_scale`. It then scales the input image such that it fits inside the scaled target
207
+ box, keeping the aspect ratio constant.
208
+ This implements the resize part of the Google's 'resize_and_crop' data augmentation:
209
+ https://github.com/tensorflow/tpu/blob/master/models/official/detection/utils/input_utils.py#L127
210
+ """
211
+
212
+ def __init__(
213
+ self,
214
+ min_scale: float,
215
+ max_scale: float,
216
+ target_height: int,
217
+ target_width: int,
218
+ interp: int = Image.BILINEAR,
219
+ ):
220
+ """
221
+ Args:
222
+ min_scale: minimum image scale range.
223
+ max_scale: maximum image scale range.
224
+ target_height: target image height.
225
+ target_width: target image width.
226
+ interp: image interpolation method.
227
+ """
228
+ super().__init__()
229
+ self._init(locals())
230
+
231
+ def _get_resize(self, image: np.ndarray, scale: float) -> Transform:
232
+ input_size = image.shape[:2]
233
+
234
+ # Compute new target size given a scale.
235
+ target_size = (self.target_height, self.target_width)
236
+ target_scale_size = np.multiply(target_size, scale)
237
+
238
+ # Compute actual rescaling applied to input image and output size.
239
+ output_scale = np.minimum(
240
+ target_scale_size[0] / input_size[0], target_scale_size[1] / input_size[1]
241
+ )
242
+ output_size = np.round(np.multiply(input_size, output_scale)).astype(int)
243
+
244
+ return ResizeTransform(
245
+ input_size[0], input_size[1], output_size[0], output_size[1], self.interp
246
+ )
247
+
248
+ def get_transform(self, image: np.ndarray) -> Transform:
249
+ random_scale = np.random.uniform(self.min_scale, self.max_scale)
250
+ return self._get_resize(image, random_scale)
251
+
252
+
253
+ class RandomRotation(Augmentation):
254
+ """
255
+ This method returns a copy of this image, rotated the given
256
+ number of degrees counter clockwise around the given center.
257
+ """
258
+
259
+ def __init__(self, angle, expand=True, center=None, sample_style="range", interp=None):
260
+ """
261
+ Args:
262
+ angle (list[float]): If ``sample_style=="range"``,
263
+ a [min, max] interval from which to sample the angle (in degrees).
264
+ If ``sample_style=="choice"``, a list of angles to sample from
265
+ expand (bool): choose if the image should be resized to fit the whole
266
+ rotated image (default), or simply cropped
267
+ center (list[[float, float]]): If ``sample_style=="range"``,
268
+ a [[minx, miny], [maxx, maxy]] relative interval from which to sample the center,
269
+ [0, 0] being the top left of the image and [1, 1] the bottom right.
270
+ If ``sample_style=="choice"``, a list of centers to sample from
271
+ Default: None, which means that the center of rotation is the center of the image
272
+ center has no effect if expand=True because it only affects shifting
273
+ """
274
+ super().__init__()
275
+ assert sample_style in ["range", "choice"], sample_style
276
+ self.is_range = sample_style == "range"
277
+ if isinstance(angle, (float, int)):
278
+ angle = (angle, angle)
279
+ if center is not None and isinstance(center[0], (float, int)):
280
+ center = (center, center)
281
+ self._init(locals())
282
+
283
+ def get_transform(self, image):
284
+ h, w = image.shape[:2]
285
+ center = None
286
+ if self.is_range:
287
+ angle = np.random.uniform(self.angle[0], self.angle[1])
288
+ if self.center is not None:
289
+ center = (
290
+ np.random.uniform(self.center[0][0], self.center[1][0]),
291
+ np.random.uniform(self.center[0][1], self.center[1][1]),
292
+ )
293
+ else:
294
+ angle = np.random.choice(self.angle)
295
+ if self.center is not None:
296
+ center = np.random.choice(self.center)
297
+
298
+ if center is not None:
299
+ center = (w * center[0], h * center[1]) # Convert to absolute coordinates
300
+
301
+ if angle % 360 == 0:
302
+ return NoOpTransform()
303
+
304
+ return RotationTransform(h, w, angle, expand=self.expand, center=center, interp=self.interp)
305
+
306
+
307
+ class FixedSizeCrop(Augmentation):
308
+ """
309
+ If `crop_size` is smaller than the input image size, then it uses a random crop of
310
+ the crop size. If `crop_size` is larger than the input image size, then it pads
311
+ the right and the bottom of the image to the crop size if `pad` is True, otherwise
312
+ it returns the smaller image.
313
+ """
314
+
315
+ def __init__(
316
+ self,
317
+ crop_size: Tuple[int],
318
+ pad: bool = True,
319
+ pad_value: float = 128.0,
320
+ seg_pad_value: int = 255,
321
+ ):
322
+ """
323
+ Args:
324
+ crop_size: target image (height, width).
325
+ pad: if True, will pad images smaller than `crop_size` up to `crop_size`
326
+ pad_value: the padding value to the image.
327
+ seg_pad_value: the padding value to the segmentation mask.
328
+ """
329
+ super().__init__()
330
+ self._init(locals())
331
+
332
+ def _get_crop(self, image: np.ndarray) -> Transform:
333
+ # Compute the image scale and scaled size.
334
+ input_size = image.shape[:2]
335
+ output_size = self.crop_size
336
+
337
+ # Add random crop if the image is scaled up.
338
+ max_offset = np.subtract(input_size, output_size)
339
+ max_offset = np.maximum(max_offset, 0)
340
+ offset = np.multiply(max_offset, np.random.uniform(0.0, 1.0))
341
+ offset = np.round(offset).astype(int)
342
+ return CropTransform(
343
+ offset[1], offset[0], output_size[1], output_size[0], input_size[1], input_size[0]
344
+ )
345
+
346
+ def _get_pad(self, image: np.ndarray) -> Transform:
347
+ # Compute the image scale and scaled size.
348
+ input_size = image.shape[:2]
349
+ output_size = self.crop_size
350
+
351
+ # Add padding if the image is scaled down.
352
+ pad_size = np.subtract(output_size, input_size)
353
+ pad_size = np.maximum(pad_size, 0)
354
+ original_size = np.minimum(input_size, output_size)
355
+ return PadTransform(
356
+ 0,
357
+ 0,
358
+ pad_size[1],
359
+ pad_size[0],
360
+ original_size[1],
361
+ original_size[0],
362
+ self.pad_value,
363
+ self.seg_pad_value,
364
+ )
365
+
366
+ def get_transform(self, image: np.ndarray) -> TransformList:
367
+ transforms = [self._get_crop(image)]
368
+ if self.pad:
369
+ transforms.append(self._get_pad(image))
370
+ return TransformList(transforms)
371
+
372
+
373
+ class RandomCrop(Augmentation):
374
+ """
375
+ Randomly crop a rectangle region out of an image.
376
+ """
377
+
378
+ def __init__(self, crop_type: str, crop_size):
379
+ """
380
+ Args:
381
+ crop_type (str): one of "relative_range", "relative", "absolute", "absolute_range".
382
+ crop_size (tuple[float, float]): two floats, explained below.
383
+
384
+ - "relative": crop a (H * crop_size[0], W * crop_size[1]) region from an input image of
385
+ size (H, W). crop size should be in (0, 1]
386
+ - "relative_range": uniformly sample two values from [crop_size[0], 1]
387
+ and [crop_size[1]], 1], and use them as in "relative" crop type.
388
+ - "absolute" crop a (crop_size[0], crop_size[1]) region from input image.
389
+ crop_size must be smaller than the input image size.
390
+ - "absolute_range", for an input of size (H, W), uniformly sample H_crop in
391
+ [crop_size[0], min(H, crop_size[1])] and W_crop in [crop_size[0], min(W, crop_size[1])].
392
+ Then crop a region (H_crop, W_crop).
393
+ """
394
+ # TODO style of relative_range and absolute_range are not consistent:
395
+ # one takes (h, w) but another takes (min, max)
396
+ super().__init__()
397
+ assert crop_type in ["relative_range", "relative", "absolute", "absolute_range"]
398
+ self._init(locals())
399
+
400
+ def get_transform(self, image):
401
+ h, w = image.shape[:2]
402
+ croph, cropw = self.get_crop_size((h, w))
403
+ assert h >= croph and w >= cropw, "Shape computation in {} has bugs.".format(self)
404
+ h0 = np.random.randint(h - croph + 1)
405
+ w0 = np.random.randint(w - cropw + 1)
406
+ return CropTransform(w0, h0, cropw, croph)
407
+
408
+ def get_crop_size(self, image_size):
409
+ """
410
+ Args:
411
+ image_size (tuple): height, width
412
+
413
+ Returns:
414
+ crop_size (tuple): height, width in absolute pixels
415
+ """
416
+ h, w = image_size
417
+ if self.crop_type == "relative":
418
+ ch, cw = self.crop_size
419
+ return int(h * ch + 0.5), int(w * cw + 0.5)
420
+ elif self.crop_type == "relative_range":
421
+ crop_size = np.asarray(self.crop_size, dtype=np.float32)
422
+ ch, cw = crop_size + np.random.rand(2) * (1 - crop_size)
423
+ return int(h * ch + 0.5), int(w * cw + 0.5)
424
+ elif self.crop_type == "absolute":
425
+ return (min(self.crop_size[0], h), min(self.crop_size[1], w))
426
+ elif self.crop_type == "absolute_range":
427
+ assert self.crop_size[0] <= self.crop_size[1]
428
+ ch = np.random.randint(min(h, self.crop_size[0]), min(h, self.crop_size[1]) + 1)
429
+ cw = np.random.randint(min(w, self.crop_size[0]), min(w, self.crop_size[1]) + 1)
430
+ return ch, cw
431
+ else:
432
+ raise NotImplementedError("Unknown crop type {}".format(self.crop_type))
433
+
434
+
435
+ class RandomCrop_CategoryAreaConstraint(Augmentation):
436
+ """
437
+ Similar to :class:`RandomCrop`, but find a cropping window such that no single category
438
+ occupies a ratio of more than `single_category_max_area` in semantic segmentation ground
439
+ truth, which can cause unstability in training. The function attempts to find such a valid
440
+ cropping window for at most 10 times.
441
+ """
442
+
443
+ def __init__(
444
+ self,
445
+ crop_type: str,
446
+ crop_size,
447
+ single_category_max_area: float = 1.0,
448
+ ignored_category: int = None,
449
+ ):
450
+ """
451
+ Args:
452
+ crop_type, crop_size: same as in :class:`RandomCrop`
453
+ single_category_max_area: the maximum allowed area ratio of a
454
+ category. Set to 1.0 to disable
455
+ ignored_category: allow this category in the semantic segmentation
456
+ ground truth to exceed the area ratio. Usually set to the category
457
+ that's ignored in training.
458
+ """
459
+ self.crop_aug = RandomCrop(crop_type, crop_size)
460
+ self._init(locals())
461
+
462
+ def get_transform(self, image, sem_seg):
463
+ if self.single_category_max_area >= 1.0:
464
+ return self.crop_aug.get_transform(image)
465
+ else:
466
+ h, w = sem_seg.shape
467
+ for _ in range(10):
468
+ crop_size = self.crop_aug.get_crop_size((h, w))
469
+ y0 = np.random.randint(h - crop_size[0] + 1)
470
+ x0 = np.random.randint(w - crop_size[1] + 1)
471
+ sem_seg_temp = sem_seg[y0 : y0 + crop_size[0], x0 : x0 + crop_size[1]]
472
+ labels, cnt = np.unique(sem_seg_temp, return_counts=True)
473
+ if self.ignored_category is not None:
474
+ cnt = cnt[labels != self.ignored_category]
475
+ if len(cnt) > 1 and np.max(cnt) < np.sum(cnt) * self.single_category_max_area:
476
+ break
477
+ crop_tfm = CropTransform(x0, y0, crop_size[1], crop_size[0])
478
+ return crop_tfm
479
+
480
+
481
+ class RandomExtent(Augmentation):
482
+ """
483
+ Outputs an image by cropping a random "subrect" of the source image.
484
+
485
+ The subrect can be parameterized to include pixels outside the source image,
486
+ in which case they will be set to zeros (i.e. black). The size of the output
487
+ image will vary with the size of the random subrect.
488
+ """
489
+
490
+ def __init__(self, scale_range, shift_range):
491
+ """
492
+ Args:
493
+ output_size (h, w): Dimensions of output image
494
+ scale_range (l, h): Range of input-to-output size scaling factor
495
+ shift_range (x, y): Range of shifts of the cropped subrect. The rect
496
+ is shifted by [w / 2 * Uniform(-x, x), h / 2 * Uniform(-y, y)],
497
+ where (w, h) is the (width, height) of the input image. Set each
498
+ component to zero to crop at the image's center.
499
+ """
500
+ super().__init__()
501
+ self._init(locals())
502
+
503
+ def get_transform(self, image):
504
+ img_h, img_w = image.shape[:2]
505
+
506
+ # Initialize src_rect to fit the input image.
507
+ src_rect = np.array([-0.5 * img_w, -0.5 * img_h, 0.5 * img_w, 0.5 * img_h])
508
+
509
+ # Apply a random scaling to the src_rect.
510
+ src_rect *= np.random.uniform(self.scale_range[0], self.scale_range[1])
511
+
512
+ # Apply a random shift to the coordinates origin.
513
+ src_rect[0::2] += self.shift_range[0] * img_w * (np.random.rand() - 0.5)
514
+ src_rect[1::2] += self.shift_range[1] * img_h * (np.random.rand() - 0.5)
515
+
516
+ # Map src_rect coordinates into image coordinates (center at corner).
517
+ src_rect[0::2] += 0.5 * img_w
518
+ src_rect[1::2] += 0.5 * img_h
519
+
520
+ return ExtentTransform(
521
+ src_rect=(src_rect[0], src_rect[1], src_rect[2], src_rect[3]),
522
+ output_size=(int(src_rect[3] - src_rect[1]), int(src_rect[2] - src_rect[0])),
523
+ )
524
+
525
+
526
+ class RandomContrast(Augmentation):
527
+ """
528
+ Randomly transforms image contrast.
529
+
530
+ Contrast intensity is uniformly sampled in (intensity_min, intensity_max).
531
+ - intensity < 1 will reduce contrast
532
+ - intensity = 1 will preserve the input image
533
+ - intensity > 1 will increase contrast
534
+
535
+ See: https://pillow.readthedocs.io/en/3.0.x/reference/ImageEnhance.html
536
+ """
537
+
538
+ def __init__(self, intensity_min, intensity_max):
539
+ """
540
+ Args:
541
+ intensity_min (float): Minimum augmentation
542
+ intensity_max (float): Maximum augmentation
543
+ """
544
+ super().__init__()
545
+ self._init(locals())
546
+
547
+ def get_transform(self, image):
548
+ w = np.random.uniform(self.intensity_min, self.intensity_max)
549
+ return BlendTransform(src_image=image.mean(), src_weight=1 - w, dst_weight=w)
550
+
551
+
552
+ class RandomBrightness(Augmentation):
553
+ """
554
+ Randomly transforms image brightness.
555
+
556
+ Brightness intensity is uniformly sampled in (intensity_min, intensity_max).
557
+ - intensity < 1 will reduce brightness
558
+ - intensity = 1 will preserve the input image
559
+ - intensity > 1 will increase brightness
560
+
561
+ See: https://pillow.readthedocs.io/en/3.0.x/reference/ImageEnhance.html
562
+ """
563
+
564
+ def __init__(self, intensity_min, intensity_max):
565
+ """
566
+ Args:
567
+ intensity_min (float): Minimum augmentation
568
+ intensity_max (float): Maximum augmentation
569
+ """
570
+ super().__init__()
571
+ self._init(locals())
572
+
573
+ def get_transform(self, image):
574
+ w = np.random.uniform(self.intensity_min, self.intensity_max)
575
+ return BlendTransform(src_image=0, src_weight=1 - w, dst_weight=w)
576
+
577
+
578
+ class RandomSaturation(Augmentation):
579
+ """
580
+ Randomly transforms saturation of an RGB image.
581
+ Input images are assumed to have 'RGB' channel order.
582
+
583
+ Saturation intensity is uniformly sampled in (intensity_min, intensity_max).
584
+ - intensity < 1 will reduce saturation (make the image more grayscale)
585
+ - intensity = 1 will preserve the input image
586
+ - intensity > 1 will increase saturation
587
+
588
+ See: https://pillow.readthedocs.io/en/3.0.x/reference/ImageEnhance.html
589
+ """
590
+
591
+ def __init__(self, intensity_min, intensity_max):
592
+ """
593
+ Args:
594
+ intensity_min (float): Minimum augmentation (1 preserves input).
595
+ intensity_max (float): Maximum augmentation (1 preserves input).
596
+ """
597
+ super().__init__()
598
+ self._init(locals())
599
+
600
+ def get_transform(self, image):
601
+ assert image.shape[-1] == 3, "RandomSaturation only works on RGB images"
602
+ w = np.random.uniform(self.intensity_min, self.intensity_max)
603
+ grayscale = image.dot([0.299, 0.587, 0.114])[:, :, np.newaxis]
604
+ return BlendTransform(src_image=grayscale, src_weight=1 - w, dst_weight=w)
605
+
606
+
607
+ class RandomLighting(Augmentation):
608
+ """
609
+ The "lighting" augmentation described in AlexNet, using fixed PCA over ImageNet.
610
+ Input images are assumed to have 'RGB' channel order.
611
+
612
+ The degree of color jittering is randomly sampled via a normal distribution,
613
+ with standard deviation given by the scale parameter.
614
+ """
615
+
616
+ def __init__(self, scale):
617
+ """
618
+ Args:
619
+ scale (float): Standard deviation of principal component weighting.
620
+ """
621
+ super().__init__()
622
+ self._init(locals())
623
+ self.eigen_vecs = np.array(
624
+ [[-0.5675, 0.7192, 0.4009], [-0.5808, -0.0045, -0.8140], [-0.5836, -0.6948, 0.4203]]
625
+ )
626
+ self.eigen_vals = np.array([0.2175, 0.0188, 0.0045])
627
+
628
+ def get_transform(self, image):
629
+ assert image.shape[-1] == 3, "RandomLighting only works on RGB images"
630
+ weights = np.random.normal(scale=self.scale, size=3)
631
+ return BlendTransform(
632
+ src_image=self.eigen_vecs.dot(weights * self.eigen_vals), src_weight=1.0, dst_weight=1.0
633
+ )
634
+
635
+
636
+ class RandomResize(Augmentation):
637
+ """Randomly resize image to a target size in shape_list"""
638
+
639
+ def __init__(self, shape_list, interp=Image.BILINEAR):
640
+ """
641
+ Args:
642
+ shape_list: a list of shapes in (h, w)
643
+ interp: PIL interpolation method
644
+ """
645
+ self.shape_list = shape_list
646
+ self._init(locals())
647
+
648
+ def get_transform(self, image):
649
+ shape_idx = np.random.randint(low=0, high=len(self.shape_list))
650
+ h, w = self.shape_list[shape_idx]
651
+ return ResizeTransform(image.shape[0], image.shape[1], h, w, self.interp)
652
+
653
+
654
+ class MinIoURandomCrop(Augmentation):
655
+ """Random crop the image & bboxes, the cropped patches have minimum IoU
656
+ requirement with original image & bboxes, the IoU threshold is randomly
657
+ selected from min_ious.
658
+
659
+ Args:
660
+ min_ious (tuple): minimum IoU threshold for all intersections with
661
+ bounding boxes
662
+ min_crop_size (float): minimum crop's size (i.e. h,w := a*h, a*w,
663
+ where a >= min_crop_size)
664
+ mode_trials: number of trials for sampling min_ious threshold
665
+ crop_trials: number of trials for sampling crop_size after cropping
666
+ """
667
+
668
+ def __init__(
669
+ self,
670
+ min_ious=(0.1, 0.3, 0.5, 0.7, 0.9),
671
+ min_crop_size=0.3,
672
+ mode_trials=1000,
673
+ crop_trials=50,
674
+ ):
675
+ self.min_ious = min_ious
676
+ self.sample_mode = (1, *min_ious, 0)
677
+ self.min_crop_size = min_crop_size
678
+ self.mode_trials = mode_trials
679
+ self.crop_trials = crop_trials
680
+
681
+ def get_transform(self, image, boxes):
682
+ """Call function to crop images and bounding boxes with minimum IoU
683
+ constraint.
684
+
685
+ Args:
686
+ boxes: ground truth boxes in (x1, y1, x2, y2) format
687
+ """
688
+ if boxes is None:
689
+ return NoOpTransform()
690
+ h, w, c = image.shape
691
+ for _ in range(self.mode_trials):
692
+ mode = random.choice(self.sample_mode)
693
+ self.mode = mode
694
+ if mode == 1:
695
+ return NoOpTransform()
696
+
697
+ min_iou = mode
698
+ for _ in range(self.crop_trials):
699
+ new_w = random.uniform(self.min_crop_size * w, w)
700
+ new_h = random.uniform(self.min_crop_size * h, h)
701
+
702
+ # h / w in [0.5, 2]
703
+ if new_h / new_w < 0.5 or new_h / new_w > 2:
704
+ continue
705
+
706
+ left = random.uniform(w - new_w)
707
+ top = random.uniform(h - new_h)
708
+
709
+ patch = np.array((int(left), int(top), int(left + new_w), int(top + new_h)))
710
+ # Line or point crop is not allowed
711
+ if patch[2] == patch[0] or patch[3] == patch[1]:
712
+ continue
713
+ overlaps = pairwise_iou(
714
+ Boxes(patch.reshape(-1, 4)), Boxes(boxes.reshape(-1, 4))
715
+ ).reshape(-1)
716
+ if len(overlaps) > 0 and overlaps.min() < min_iou:
717
+ continue
718
+
719
+ # center of boxes should inside the crop img
720
+ # only adjust boxes and instance masks when the gt is not empty
721
+ if len(overlaps) > 0:
722
+ # adjust boxes
723
+ def is_center_of_bboxes_in_patch(boxes, patch):
724
+ center = (boxes[:, :2] + boxes[:, 2:]) / 2
725
+ mask = (
726
+ (center[:, 0] > patch[0])
727
+ * (center[:, 1] > patch[1])
728
+ * (center[:, 0] < patch[2])
729
+ * (center[:, 1] < patch[3])
730
+ )
731
+ return mask
732
+
733
+ mask = is_center_of_bboxes_in_patch(boxes, patch)
734
+ if not mask.any():
735
+ continue
736
+ return CropTransform(int(left), int(top), int(new_w), int(new_h))
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/data/transforms/transform.py ADDED
@@ -0,0 +1,351 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ """
5
+ See "Data Augmentation" tutorial for an overview of the system:
6
+ https://detectron2.readthedocs.io/tutorials/augmentation.html
7
+ """
8
+
9
+ import numpy as np
10
+ import torch
11
+ import torch.nn.functional as F
12
+ from fvcore.transforms.transform import (
13
+ CropTransform,
14
+ HFlipTransform,
15
+ NoOpTransform,
16
+ Transform,
17
+ TransformList,
18
+ )
19
+ from PIL import Image
20
+
21
+ try:
22
+ import cv2 # noqa
23
+ except ImportError:
24
+ # OpenCV is an optional dependency at the moment
25
+ pass
26
+
27
+ __all__ = [
28
+ "ExtentTransform",
29
+ "ResizeTransform",
30
+ "RotationTransform",
31
+ "ColorTransform",
32
+ "PILColorTransform",
33
+ ]
34
+
35
+
36
+ class ExtentTransform(Transform):
37
+ """
38
+ Extracts a subregion from the source image and scales it to the output size.
39
+
40
+ The fill color is used to map pixels from the source rect that fall outside
41
+ the source image.
42
+
43
+ See: https://pillow.readthedocs.io/en/latest/PIL.html#PIL.ImageTransform.ExtentTransform
44
+ """
45
+
46
+ def __init__(self, src_rect, output_size, interp=Image.LINEAR, fill=0):
47
+ """
48
+ Args:
49
+ src_rect (x0, y0, x1, y1): src coordinates
50
+ output_size (h, w): dst image size
51
+ interp: PIL interpolation methods
52
+ fill: Fill color used when src_rect extends outside image
53
+ """
54
+ super().__init__()
55
+ self._set_attributes(locals())
56
+
57
+ def apply_image(self, img, interp=None):
58
+ h, w = self.output_size
59
+ if len(img.shape) > 2 and img.shape[2] == 1:
60
+ pil_image = Image.fromarray(img[:, :, 0], mode="L")
61
+ else:
62
+ pil_image = Image.fromarray(img)
63
+ pil_image = pil_image.transform(
64
+ size=(w, h),
65
+ method=Image.EXTENT,
66
+ data=self.src_rect,
67
+ resample=interp if interp else self.interp,
68
+ fill=self.fill,
69
+ )
70
+ ret = np.asarray(pil_image)
71
+ if len(img.shape) > 2 and img.shape[2] == 1:
72
+ ret = np.expand_dims(ret, -1)
73
+ return ret
74
+
75
+ def apply_coords(self, coords):
76
+ # Transform image center from source coordinates into output coordinates
77
+ # and then map the new origin to the corner of the output image.
78
+ h, w = self.output_size
79
+ x0, y0, x1, y1 = self.src_rect
80
+ new_coords = coords.astype(np.float32)
81
+ new_coords[:, 0] -= 0.5 * (x0 + x1)
82
+ new_coords[:, 1] -= 0.5 * (y0 + y1)
83
+ new_coords[:, 0] *= w / (x1 - x0)
84
+ new_coords[:, 1] *= h / (y1 - y0)
85
+ new_coords[:, 0] += 0.5 * w
86
+ new_coords[:, 1] += 0.5 * h
87
+ return new_coords
88
+
89
+ def apply_segmentation(self, segmentation):
90
+ segmentation = self.apply_image(segmentation, interp=Image.NEAREST)
91
+ return segmentation
92
+
93
+
94
+ class ResizeTransform(Transform):
95
+ """
96
+ Resize the image to a target size.
97
+ """
98
+
99
+ def __init__(self, h, w, new_h, new_w, interp=None):
100
+ """
101
+ Args:
102
+ h, w (int): original image size
103
+ new_h, new_w (int): new image size
104
+ interp: PIL interpolation methods, defaults to bilinear.
105
+ """
106
+ # TODO decide on PIL vs opencv
107
+ super().__init__()
108
+ if interp is None:
109
+ interp = Image.BILINEAR
110
+ self._set_attributes(locals())
111
+
112
+ def apply_image(self, img, interp=None):
113
+ assert img.shape[:2] == (self.h, self.w)
114
+ assert len(img.shape) <= 4
115
+ interp_method = interp if interp is not None else self.interp
116
+
117
+ if img.dtype == np.uint8:
118
+ if len(img.shape) > 2 and img.shape[2] == 1:
119
+ pil_image = Image.fromarray(img[:, :, 0], mode="L")
120
+ else:
121
+ pil_image = Image.fromarray(img)
122
+ pil_image = pil_image.resize((self.new_w, self.new_h), interp_method)
123
+ ret = np.asarray(pil_image)
124
+ if len(img.shape) > 2 and img.shape[2] == 1:
125
+ ret = np.expand_dims(ret, -1)
126
+ else:
127
+ # PIL only supports uint8
128
+ if any(x < 0 for x in img.strides):
129
+ img = np.ascontiguousarray(img)
130
+ img = torch.from_numpy(img)
131
+ shape = list(img.shape)
132
+ shape_4d = shape[:2] + [1] * (4 - len(shape)) + shape[2:]
133
+ img = img.view(shape_4d).permute(2, 3, 0, 1) # hw(c) -> nchw
134
+ _PIL_RESIZE_TO_INTERPOLATE_MODE = {
135
+ Image.NEAREST: "nearest",
136
+ Image.BILINEAR: "bilinear",
137
+ Image.BICUBIC: "bicubic",
138
+ }
139
+ mode = _PIL_RESIZE_TO_INTERPOLATE_MODE[interp_method]
140
+ align_corners = None if mode == "nearest" else False
141
+ img = F.interpolate(
142
+ img, (self.new_h, self.new_w), mode=mode, align_corners=align_corners
143
+ )
144
+ shape[:2] = (self.new_h, self.new_w)
145
+ ret = img.permute(2, 3, 0, 1).view(shape).numpy() # nchw -> hw(c)
146
+
147
+ return ret
148
+
149
+ def apply_coords(self, coords):
150
+ coords[:, 0] = coords[:, 0] * (self.new_w * 1.0 / self.w)
151
+ coords[:, 1] = coords[:, 1] * (self.new_h * 1.0 / self.h)
152
+ return coords
153
+
154
+ def apply_segmentation(self, segmentation):
155
+ segmentation = self.apply_image(segmentation, interp=Image.NEAREST)
156
+ return segmentation
157
+
158
+ def inverse(self):
159
+ return ResizeTransform(self.new_h, self.new_w, self.h, self.w, self.interp)
160
+
161
+
162
+ class RotationTransform(Transform):
163
+ """
164
+ This method returns a copy of this image, rotated the given
165
+ number of degrees counter clockwise around its center.
166
+ """
167
+
168
+ def __init__(self, h, w, angle, expand=True, center=None, interp=None):
169
+ """
170
+ Args:
171
+ h, w (int): original image size
172
+ angle (float): degrees for rotation
173
+ expand (bool): choose if the image should be resized to fit the whole
174
+ rotated image (default), or simply cropped
175
+ center (tuple (width, height)): coordinates of the rotation center
176
+ if left to None, the center will be fit to the center of each image
177
+ center has no effect if expand=True because it only affects shifting
178
+ interp: cv2 interpolation method, default cv2.INTER_LINEAR
179
+ """
180
+ super().__init__()
181
+ image_center = np.array((w / 2, h / 2))
182
+ if center is None:
183
+ center = image_center
184
+ if interp is None:
185
+ interp = cv2.INTER_LINEAR
186
+ abs_cos, abs_sin = (abs(np.cos(np.deg2rad(angle))), abs(np.sin(np.deg2rad(angle))))
187
+ if expand:
188
+ # find the new width and height bounds
189
+ bound_w, bound_h = np.rint(
190
+ [h * abs_sin + w * abs_cos, h * abs_cos + w * abs_sin]
191
+ ).astype(int)
192
+ else:
193
+ bound_w, bound_h = w, h
194
+
195
+ self._set_attributes(locals())
196
+ self.rm_coords = self.create_rotation_matrix()
197
+ # Needed because of this problem https://github.com/opencv/opencv/issues/11784
198
+ self.rm_image = self.create_rotation_matrix(offset=-0.5)
199
+
200
+ def apply_image(self, img, interp=None):
201
+ """
202
+ img should be a numpy array, formatted as Height * Width * Nchannels
203
+ """
204
+ if len(img) == 0 or self.angle % 360 == 0:
205
+ return img
206
+ assert img.shape[:2] == (self.h, self.w)
207
+ interp = interp if interp is not None else self.interp
208
+ return cv2.warpAffine(img, self.rm_image, (self.bound_w, self.bound_h), flags=interp)
209
+
210
+ def apply_coords(self, coords):
211
+ """
212
+ coords should be a N * 2 array-like, containing N couples of (x, y) points
213
+ """
214
+ coords = np.asarray(coords, dtype=float)
215
+ if len(coords) == 0 or self.angle % 360 == 0:
216
+ return coords
217
+ return cv2.transform(coords[:, np.newaxis, :], self.rm_coords)[:, 0, :]
218
+
219
+ def apply_segmentation(self, segmentation):
220
+ segmentation = self.apply_image(segmentation, interp=cv2.INTER_NEAREST)
221
+ return segmentation
222
+
223
+ def create_rotation_matrix(self, offset=0):
224
+ center = (self.center[0] + offset, self.center[1] + offset)
225
+ rm = cv2.getRotationMatrix2D(tuple(center), self.angle, 1)
226
+ if self.expand:
227
+ # Find the coordinates of the center of rotation in the new image
228
+ # The only point for which we know the future coordinates is the center of the image
229
+ rot_im_center = cv2.transform(self.image_center[None, None, :] + offset, rm)[0, 0, :]
230
+ new_center = np.array([self.bound_w / 2, self.bound_h / 2]) + offset - rot_im_center
231
+ # shift the rotation center to the new coordinates
232
+ rm[:, 2] += new_center
233
+ return rm
234
+
235
+ def inverse(self):
236
+ """
237
+ The inverse is to rotate it back with expand, and crop to get the original shape.
238
+ """
239
+ if not self.expand: # Not possible to inverse if a part of the image is lost
240
+ raise NotImplementedError()
241
+ rotation = RotationTransform(
242
+ self.bound_h, self.bound_w, -self.angle, True, None, self.interp
243
+ )
244
+ crop = CropTransform(
245
+ (rotation.bound_w - self.w) // 2, (rotation.bound_h - self.h) // 2, self.w, self.h
246
+ )
247
+ return TransformList([rotation, crop])
248
+
249
+
250
+ class ColorTransform(Transform):
251
+ """
252
+ Generic wrapper for any photometric transforms.
253
+ These transformations should only affect the color space and
254
+ not the coordinate space of the image (e.g. annotation
255
+ coordinates such as bounding boxes should not be changed)
256
+ """
257
+
258
+ def __init__(self, op):
259
+ """
260
+ Args:
261
+ op (Callable): operation to be applied to the image,
262
+ which takes in an ndarray and returns an ndarray.
263
+ """
264
+ if not callable(op):
265
+ raise ValueError("op parameter should be callable")
266
+ super().__init__()
267
+ self._set_attributes(locals())
268
+
269
+ def apply_image(self, img):
270
+ return self.op(img)
271
+
272
+ def apply_coords(self, coords):
273
+ return coords
274
+
275
+ def inverse(self):
276
+ return NoOpTransform()
277
+
278
+ def apply_segmentation(self, segmentation):
279
+ return segmentation
280
+
281
+
282
+ class PILColorTransform(ColorTransform):
283
+ """
284
+ Generic wrapper for PIL Photometric image transforms,
285
+ which affect the color space and not the coordinate
286
+ space of the image
287
+ """
288
+
289
+ def __init__(self, op):
290
+ """
291
+ Args:
292
+ op (Callable): operation to be applied to the image,
293
+ which takes in a PIL Image and returns a transformed
294
+ PIL Image.
295
+ For reference on possible operations see:
296
+ - https://pillow.readthedocs.io/en/stable/
297
+ """
298
+ if not callable(op):
299
+ raise ValueError("op parameter should be callable")
300
+ super().__init__(op)
301
+
302
+ def apply_image(self, img):
303
+ img = Image.fromarray(img)
304
+ return np.asarray(super().apply_image(img))
305
+
306
+
307
+ def HFlip_rotated_box(transform, rotated_boxes):
308
+ """
309
+ Apply the horizontal flip transform on rotated boxes.
310
+
311
+ Args:
312
+ rotated_boxes (ndarray): Nx5 floating point array of
313
+ (x_center, y_center, width, height, angle_degrees) format
314
+ in absolute coordinates.
315
+ """
316
+ # Transform x_center
317
+ rotated_boxes[:, 0] = transform.width - rotated_boxes[:, 0]
318
+ # Transform angle
319
+ rotated_boxes[:, 4] = -rotated_boxes[:, 4]
320
+ return rotated_boxes
321
+
322
+
323
+ def Resize_rotated_box(transform, rotated_boxes):
324
+ """
325
+ Apply the resizing transform on rotated boxes. For details of how these (approximation)
326
+ formulas are derived, please refer to :meth:`RotatedBoxes.scale`.
327
+
328
+ Args:
329
+ rotated_boxes (ndarray): Nx5 floating point array of
330
+ (x_center, y_center, width, height, angle_degrees) format
331
+ in absolute coordinates.
332
+ """
333
+ scale_factor_x = transform.new_w * 1.0 / transform.w
334
+ scale_factor_y = transform.new_h * 1.0 / transform.h
335
+ rotated_boxes[:, 0] *= scale_factor_x
336
+ rotated_boxes[:, 1] *= scale_factor_y
337
+ theta = rotated_boxes[:, 4] * np.pi / 180.0
338
+ c = np.cos(theta)
339
+ s = np.sin(theta)
340
+ rotated_boxes[:, 2] *= np.sqrt(np.square(scale_factor_x * c) + np.square(scale_factor_y * s))
341
+ rotated_boxes[:, 3] *= np.sqrt(np.square(scale_factor_x * s) + np.square(scale_factor_y * c))
342
+ rotated_boxes[:, 4] = np.arctan2(scale_factor_x * s, scale_factor_y * c) * 180 / np.pi
343
+
344
+ return rotated_boxes
345
+
346
+
347
+ HFlipTransform.register_type("rotated_box", HFlip_rotated_box)
348
+ ResizeTransform.register_type("rotated_box", Resize_rotated_box)
349
+
350
+ # not necessary any more with latest fvcore
351
+ NoOpTransform.register_type("rotated_box", lambda t, x: x)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/__init__.py ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+
3
+ from .launch import *
4
+ from .train_loop import *
5
+
6
+ __all__ = [k for k in globals().keys() if not k.startswith("_")]
7
+
8
+
9
+ # prefer to let hooks and defaults live in separate namespaces (therefore not in __all__)
10
+ # but still make them available here
11
+ from .hooks import *
12
+ from .defaults import *
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/defaults.py ADDED
@@ -0,0 +1,715 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ """
5
+ This file contains components with some default boilerplate logic user may need
6
+ in training / testing. They will not work for everyone, but many users may find them useful.
7
+
8
+ The behavior of functions/classes in this file is subject to change,
9
+ since they are meant to represent the "common default behavior" people need in their projects.
10
+ """
11
+
12
+ import argparse
13
+ import logging
14
+ import os
15
+ import sys
16
+ import weakref
17
+ from collections import OrderedDict
18
+ from typing import Optional
19
+ import torch
20
+ from fvcore.nn.precise_bn import get_bn_modules
21
+ from omegaconf import OmegaConf
22
+ from torch.nn.parallel import DistributedDataParallel
23
+
24
+ import annotator.oneformer.detectron2.data.transforms as T
25
+ from annotator.oneformer.detectron2.checkpoint import DetectionCheckpointer
26
+ from annotator.oneformer.detectron2.config import CfgNode, LazyConfig
27
+ from annotator.oneformer.detectron2.data import (
28
+ MetadataCatalog,
29
+ build_detection_test_loader,
30
+ build_detection_train_loader,
31
+ )
32
+ from annotator.oneformer.detectron2.evaluation import (
33
+ DatasetEvaluator,
34
+ inference_on_dataset,
35
+ print_csv_format,
36
+ verify_results,
37
+ )
38
+ from annotator.oneformer.detectron2.modeling import build_model
39
+ from annotator.oneformer.detectron2.solver import build_lr_scheduler, build_optimizer
40
+ from annotator.oneformer.detectron2.utils import comm
41
+ from annotator.oneformer.detectron2.utils.collect_env import collect_env_info
42
+ from annotator.oneformer.detectron2.utils.env import seed_all_rng
43
+ from annotator.oneformer.detectron2.utils.events import CommonMetricPrinter, JSONWriter, TensorboardXWriter
44
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
45
+ from annotator.oneformer.detectron2.utils.logger import setup_logger
46
+
47
+ from . import hooks
48
+ from .train_loop import AMPTrainer, SimpleTrainer, TrainerBase
49
+
50
+ __all__ = [
51
+ "create_ddp_model",
52
+ "default_argument_parser",
53
+ "default_setup",
54
+ "default_writers",
55
+ "DefaultPredictor",
56
+ "DefaultTrainer",
57
+ ]
58
+
59
+
60
+ def create_ddp_model(model, *, fp16_compression=False, **kwargs):
61
+ """
62
+ Create a DistributedDataParallel model if there are >1 processes.
63
+
64
+ Args:
65
+ model: a torch.nn.Module
66
+ fp16_compression: add fp16 compression hooks to the ddp object.
67
+ See more at https://pytorch.org/docs/stable/ddp_comm_hooks.html#torch.distributed.algorithms.ddp_comm_hooks.default_hooks.fp16_compress_hook
68
+ kwargs: other arguments of :module:`torch.nn.parallel.DistributedDataParallel`.
69
+ """ # noqa
70
+ if comm.get_world_size() == 1:
71
+ return model
72
+ if "device_ids" not in kwargs:
73
+ kwargs["device_ids"] = [comm.get_local_rank()]
74
+ ddp = DistributedDataParallel(model, **kwargs)
75
+ if fp16_compression:
76
+ from torch.distributed.algorithms.ddp_comm_hooks import default as comm_hooks
77
+
78
+ ddp.register_comm_hook(state=None, hook=comm_hooks.fp16_compress_hook)
79
+ return ddp
80
+
81
+
82
+ def default_argument_parser(epilog=None):
83
+ """
84
+ Create a parser with some common arguments used by detectron2 users.
85
+
86
+ Args:
87
+ epilog (str): epilog passed to ArgumentParser describing the usage.
88
+
89
+ Returns:
90
+ argparse.ArgumentParser:
91
+ """
92
+ parser = argparse.ArgumentParser(
93
+ epilog=epilog
94
+ or f"""
95
+ Examples:
96
+
97
+ Run on single machine:
98
+ $ {sys.argv[0]} --num-gpus 8 --config-file cfg.yaml
99
+
100
+ Change some config options:
101
+ $ {sys.argv[0]} --config-file cfg.yaml MODEL.WEIGHTS /path/to/weight.pth SOLVER.BASE_LR 0.001
102
+
103
+ Run on multiple machines:
104
+ (machine0)$ {sys.argv[0]} --machine-rank 0 --num-machines 2 --dist-url <URL> [--other-flags]
105
+ (machine1)$ {sys.argv[0]} --machine-rank 1 --num-machines 2 --dist-url <URL> [--other-flags]
106
+ """,
107
+ formatter_class=argparse.RawDescriptionHelpFormatter,
108
+ )
109
+ parser.add_argument("--config-file", default="", metavar="FILE", help="path to config file")
110
+ parser.add_argument(
111
+ "--resume",
112
+ action="store_true",
113
+ help="Whether to attempt to resume from the checkpoint directory. "
114
+ "See documentation of `DefaultTrainer.resume_or_load()` for what it means.",
115
+ )
116
+ parser.add_argument("--eval-only", action="store_true", help="perform evaluation only")
117
+ parser.add_argument("--num-gpus", type=int, default=1, help="number of gpus *per machine*")
118
+ parser.add_argument("--num-machines", type=int, default=1, help="total number of machines")
119
+ parser.add_argument(
120
+ "--machine-rank", type=int, default=0, help="the rank of this machine (unique per machine)"
121
+ )
122
+
123
+ # PyTorch still may leave orphan processes in multi-gpu training.
124
+ # Therefore we use a deterministic way to obtain port,
125
+ # so that users are aware of orphan processes by seeing the port occupied.
126
+ port = 2**15 + 2**14 + hash(os.getuid() if sys.platform != "win32" else 1) % 2**14
127
+ parser.add_argument(
128
+ "--dist-url",
129
+ default="tcp://127.0.0.1:{}".format(port),
130
+ help="initialization URL for pytorch distributed backend. See "
131
+ "https://pytorch.org/docs/stable/distributed.html for details.",
132
+ )
133
+ parser.add_argument(
134
+ "opts",
135
+ help="""
136
+ Modify config options at the end of the command. For Yacs configs, use
137
+ space-separated "PATH.KEY VALUE" pairs.
138
+ For python-based LazyConfig, use "path.key=value".
139
+ """.strip(),
140
+ default=None,
141
+ nargs=argparse.REMAINDER,
142
+ )
143
+ return parser
144
+
145
+
146
+ def _try_get_key(cfg, *keys, default=None):
147
+ """
148
+ Try select keys from cfg until the first key that exists. Otherwise return default.
149
+ """
150
+ if isinstance(cfg, CfgNode):
151
+ cfg = OmegaConf.create(cfg.dump())
152
+ for k in keys:
153
+ none = object()
154
+ p = OmegaConf.select(cfg, k, default=none)
155
+ if p is not none:
156
+ return p
157
+ return default
158
+
159
+
160
+ def _highlight(code, filename):
161
+ try:
162
+ import pygments
163
+ except ImportError:
164
+ return code
165
+
166
+ from pygments.lexers import Python3Lexer, YamlLexer
167
+ from pygments.formatters import Terminal256Formatter
168
+
169
+ lexer = Python3Lexer() if filename.endswith(".py") else YamlLexer()
170
+ code = pygments.highlight(code, lexer, Terminal256Formatter(style="monokai"))
171
+ return code
172
+
173
+
174
+ def default_setup(cfg, args):
175
+ """
176
+ Perform some basic common setups at the beginning of a job, including:
177
+
178
+ 1. Set up the detectron2 logger
179
+ 2. Log basic information about environment, cmdline arguments, and config
180
+ 3. Backup the config to the output directory
181
+
182
+ Args:
183
+ cfg (CfgNode or omegaconf.DictConfig): the full config to be used
184
+ args (argparse.NameSpace): the command line arguments to be logged
185
+ """
186
+ output_dir = _try_get_key(cfg, "OUTPUT_DIR", "output_dir", "train.output_dir")
187
+ if comm.is_main_process() and output_dir:
188
+ PathManager.mkdirs(output_dir)
189
+
190
+ rank = comm.get_rank()
191
+ setup_logger(output_dir, distributed_rank=rank, name="fvcore")
192
+ logger = setup_logger(output_dir, distributed_rank=rank)
193
+
194
+ logger.info("Rank of current process: {}. World size: {}".format(rank, comm.get_world_size()))
195
+ logger.info("Environment info:\n" + collect_env_info())
196
+
197
+ logger.info("Command line arguments: " + str(args))
198
+ if hasattr(args, "config_file") and args.config_file != "":
199
+ logger.info(
200
+ "Contents of args.config_file={}:\n{}".format(
201
+ args.config_file,
202
+ _highlight(PathManager.open(args.config_file, "r").read(), args.config_file),
203
+ )
204
+ )
205
+
206
+ if comm.is_main_process() and output_dir:
207
+ # Note: some of our scripts may expect the existence of
208
+ # config.yaml in output directory
209
+ path = os.path.join(output_dir, "config.yaml")
210
+ if isinstance(cfg, CfgNode):
211
+ logger.info("Running with full config:\n{}".format(_highlight(cfg.dump(), ".yaml")))
212
+ with PathManager.open(path, "w") as f:
213
+ f.write(cfg.dump())
214
+ else:
215
+ LazyConfig.save(cfg, path)
216
+ logger.info("Full config saved to {}".format(path))
217
+
218
+ # make sure each worker has a different, yet deterministic seed if specified
219
+ seed = _try_get_key(cfg, "SEED", "train.seed", default=-1)
220
+ seed_all_rng(None if seed < 0 else seed + rank)
221
+
222
+ # cudnn benchmark has large overhead. It shouldn't be used considering the small size of
223
+ # typical validation set.
224
+ if not (hasattr(args, "eval_only") and args.eval_only):
225
+ torch.backends.cudnn.benchmark = _try_get_key(
226
+ cfg, "CUDNN_BENCHMARK", "train.cudnn_benchmark", default=False
227
+ )
228
+
229
+
230
+ def default_writers(output_dir: str, max_iter: Optional[int] = None):
231
+ """
232
+ Build a list of :class:`EventWriter` to be used.
233
+ It now consists of a :class:`CommonMetricPrinter`,
234
+ :class:`TensorboardXWriter` and :class:`JSONWriter`.
235
+
236
+ Args:
237
+ output_dir: directory to store JSON metrics and tensorboard events
238
+ max_iter: the total number of iterations
239
+
240
+ Returns:
241
+ list[EventWriter]: a list of :class:`EventWriter` objects.
242
+ """
243
+ PathManager.mkdirs(output_dir)
244
+ return [
245
+ # It may not always print what you want to see, since it prints "common" metrics only.
246
+ CommonMetricPrinter(max_iter),
247
+ JSONWriter(os.path.join(output_dir, "metrics.json")),
248
+ TensorboardXWriter(output_dir),
249
+ ]
250
+
251
+
252
+ class DefaultPredictor:
253
+ """
254
+ Create a simple end-to-end predictor with the given config that runs on
255
+ single device for a single input image.
256
+
257
+ Compared to using the model directly, this class does the following additions:
258
+
259
+ 1. Load checkpoint from `cfg.MODEL.WEIGHTS`.
260
+ 2. Always take BGR image as the input and apply conversion defined by `cfg.INPUT.FORMAT`.
261
+ 3. Apply resizing defined by `cfg.INPUT.{MIN,MAX}_SIZE_TEST`.
262
+ 4. Take one input image and produce a single output, instead of a batch.
263
+
264
+ This is meant for simple demo purposes, so it does the above steps automatically.
265
+ This is not meant for benchmarks or running complicated inference logic.
266
+ If you'd like to do anything more complicated, please refer to its source code as
267
+ examples to build and use the model manually.
268
+
269
+ Attributes:
270
+ metadata (Metadata): the metadata of the underlying dataset, obtained from
271
+ cfg.DATASETS.TEST.
272
+
273
+ Examples:
274
+ ::
275
+ pred = DefaultPredictor(cfg)
276
+ inputs = cv2.imread("input.jpg")
277
+ outputs = pred(inputs)
278
+ """
279
+
280
+ def __init__(self, cfg):
281
+ self.cfg = cfg.clone() # cfg can be modified by model
282
+ self.model = build_model(self.cfg)
283
+ self.model.eval()
284
+ if len(cfg.DATASETS.TEST):
285
+ self.metadata = MetadataCatalog.get(cfg.DATASETS.TEST[0])
286
+
287
+ checkpointer = DetectionCheckpointer(self.model)
288
+ checkpointer.load(cfg.MODEL.WEIGHTS)
289
+
290
+ self.aug = T.ResizeShortestEdge(
291
+ [cfg.INPUT.MIN_SIZE_TEST, cfg.INPUT.MIN_SIZE_TEST], cfg.INPUT.MAX_SIZE_TEST
292
+ )
293
+
294
+ self.input_format = cfg.INPUT.FORMAT
295
+ assert self.input_format in ["RGB", "BGR"], self.input_format
296
+
297
+ def __call__(self, original_image):
298
+ """
299
+ Args:
300
+ original_image (np.ndarray): an image of shape (H, W, C) (in BGR order).
301
+
302
+ Returns:
303
+ predictions (dict):
304
+ the output of the model for one image only.
305
+ See :doc:`/tutorials/models` for details about the format.
306
+ """
307
+ with torch.no_grad(): # https://github.com/sphinx-doc/sphinx/issues/4258
308
+ # Apply pre-processing to image.
309
+ if self.input_format == "RGB":
310
+ # whether the model expects BGR inputs or RGB
311
+ original_image = original_image[:, :, ::-1]
312
+ height, width = original_image.shape[:2]
313
+ image = self.aug.get_transform(original_image).apply_image(original_image)
314
+ image = torch.as_tensor(image.astype("float32").transpose(2, 0, 1))
315
+
316
+ inputs = {"image": image, "height": height, "width": width}
317
+ predictions = self.model([inputs])[0]
318
+ return predictions
319
+
320
+
321
+ class DefaultTrainer(TrainerBase):
322
+ """
323
+ A trainer with default training logic. It does the following:
324
+
325
+ 1. Create a :class:`SimpleTrainer` using model, optimizer, dataloader
326
+ defined by the given config. Create a LR scheduler defined by the config.
327
+ 2. Load the last checkpoint or `cfg.MODEL.WEIGHTS`, if exists, when
328
+ `resume_or_load` is called.
329
+ 3. Register a few common hooks defined by the config.
330
+
331
+ It is created to simplify the **standard model training workflow** and reduce code boilerplate
332
+ for users who only need the standard training workflow, with standard features.
333
+ It means this class makes *many assumptions* about your training logic that
334
+ may easily become invalid in a new research. In fact, any assumptions beyond those made in the
335
+ :class:`SimpleTrainer` are too much for research.
336
+
337
+ The code of this class has been annotated about restrictive assumptions it makes.
338
+ When they do not work for you, you're encouraged to:
339
+
340
+ 1. Overwrite methods of this class, OR:
341
+ 2. Use :class:`SimpleTrainer`, which only does minimal SGD training and
342
+ nothing else. You can then add your own hooks if needed. OR:
343
+ 3. Write your own training loop similar to `tools/plain_train_net.py`.
344
+
345
+ See the :doc:`/tutorials/training` tutorials for more details.
346
+
347
+ Note that the behavior of this class, like other functions/classes in
348
+ this file, is not stable, since it is meant to represent the "common default behavior".
349
+ It is only guaranteed to work well with the standard models and training workflow in detectron2.
350
+ To obtain more stable behavior, write your own training logic with other public APIs.
351
+
352
+ Examples:
353
+ ::
354
+ trainer = DefaultTrainer(cfg)
355
+ trainer.resume_or_load() # load last checkpoint or MODEL.WEIGHTS
356
+ trainer.train()
357
+
358
+ Attributes:
359
+ scheduler:
360
+ checkpointer (DetectionCheckpointer):
361
+ cfg (CfgNode):
362
+ """
363
+
364
+ def __init__(self, cfg):
365
+ """
366
+ Args:
367
+ cfg (CfgNode):
368
+ """
369
+ super().__init__()
370
+ logger = logging.getLogger("detectron2")
371
+ if not logger.isEnabledFor(logging.INFO): # setup_logger is not called for d2
372
+ setup_logger()
373
+ cfg = DefaultTrainer.auto_scale_workers(cfg, comm.get_world_size())
374
+
375
+ # Assume these objects must be constructed in this order.
376
+ model = self.build_model(cfg)
377
+ optimizer = self.build_optimizer(cfg, model)
378
+ data_loader = self.build_train_loader(cfg)
379
+
380
+ model = create_ddp_model(model, broadcast_buffers=False)
381
+ self._trainer = (AMPTrainer if cfg.SOLVER.AMP.ENABLED else SimpleTrainer)(
382
+ model, data_loader, optimizer
383
+ )
384
+
385
+ self.scheduler = self.build_lr_scheduler(cfg, optimizer)
386
+ self.checkpointer = DetectionCheckpointer(
387
+ # Assume you want to save checkpoints together with logs/statistics
388
+ model,
389
+ cfg.OUTPUT_DIR,
390
+ trainer=weakref.proxy(self),
391
+ )
392
+ self.start_iter = 0
393
+ self.max_iter = cfg.SOLVER.MAX_ITER
394
+ self.cfg = cfg
395
+
396
+ self.register_hooks(self.build_hooks())
397
+
398
+ def resume_or_load(self, resume=True):
399
+ """
400
+ If `resume==True` and `cfg.OUTPUT_DIR` contains the last checkpoint (defined by
401
+ a `last_checkpoint` file), resume from the file. Resuming means loading all
402
+ available states (eg. optimizer and scheduler) and update iteration counter
403
+ from the checkpoint. ``cfg.MODEL.WEIGHTS`` will not be used.
404
+
405
+ Otherwise, this is considered as an independent training. The method will load model
406
+ weights from the file `cfg.MODEL.WEIGHTS` (but will not load other states) and start
407
+ from iteration 0.
408
+
409
+ Args:
410
+ resume (bool): whether to do resume or not
411
+ """
412
+ self.checkpointer.resume_or_load(self.cfg.MODEL.WEIGHTS, resume=resume)
413
+ if resume and self.checkpointer.has_checkpoint():
414
+ # The checkpoint stores the training iteration that just finished, thus we start
415
+ # at the next iteration
416
+ self.start_iter = self.iter + 1
417
+
418
+ def build_hooks(self):
419
+ """
420
+ Build a list of default hooks, including timing, evaluation,
421
+ checkpointing, lr scheduling, precise BN, writing events.
422
+
423
+ Returns:
424
+ list[HookBase]:
425
+ """
426
+ cfg = self.cfg.clone()
427
+ cfg.defrost()
428
+ cfg.DATALOADER.NUM_WORKERS = 0 # save some memory and time for PreciseBN
429
+
430
+ ret = [
431
+ hooks.IterationTimer(),
432
+ hooks.LRScheduler(),
433
+ hooks.PreciseBN(
434
+ # Run at the same freq as (but before) evaluation.
435
+ cfg.TEST.EVAL_PERIOD,
436
+ self.model,
437
+ # Build a new data loader to not affect training
438
+ self.build_train_loader(cfg),
439
+ cfg.TEST.PRECISE_BN.NUM_ITER,
440
+ )
441
+ if cfg.TEST.PRECISE_BN.ENABLED and get_bn_modules(self.model)
442
+ else None,
443
+ ]
444
+
445
+ # Do PreciseBN before checkpointer, because it updates the model and need to
446
+ # be saved by checkpointer.
447
+ # This is not always the best: if checkpointing has a different frequency,
448
+ # some checkpoints may have more precise statistics than others.
449
+ if comm.is_main_process():
450
+ ret.append(hooks.PeriodicCheckpointer(self.checkpointer, cfg.SOLVER.CHECKPOINT_PERIOD))
451
+
452
+ def test_and_save_results():
453
+ self._last_eval_results = self.test(self.cfg, self.model)
454
+ return self._last_eval_results
455
+
456
+ # Do evaluation after checkpointer, because then if it fails,
457
+ # we can use the saved checkpoint to debug.
458
+ ret.append(hooks.EvalHook(cfg.TEST.EVAL_PERIOD, test_and_save_results))
459
+
460
+ if comm.is_main_process():
461
+ # Here the default print/log frequency of each writer is used.
462
+ # run writers in the end, so that evaluation metrics are written
463
+ ret.append(hooks.PeriodicWriter(self.build_writers(), period=20))
464
+ return ret
465
+
466
+ def build_writers(self):
467
+ """
468
+ Build a list of writers to be used using :func:`default_writers()`.
469
+ If you'd like a different list of writers, you can overwrite it in
470
+ your trainer.
471
+
472
+ Returns:
473
+ list[EventWriter]: a list of :class:`EventWriter` objects.
474
+ """
475
+ return default_writers(self.cfg.OUTPUT_DIR, self.max_iter)
476
+
477
+ def train(self):
478
+ """
479
+ Run training.
480
+
481
+ Returns:
482
+ OrderedDict of results, if evaluation is enabled. Otherwise None.
483
+ """
484
+ super().train(self.start_iter, self.max_iter)
485
+ if len(self.cfg.TEST.EXPECTED_RESULTS) and comm.is_main_process():
486
+ assert hasattr(
487
+ self, "_last_eval_results"
488
+ ), "No evaluation results obtained during training!"
489
+ verify_results(self.cfg, self._last_eval_results)
490
+ return self._last_eval_results
491
+
492
+ def run_step(self):
493
+ self._trainer.iter = self.iter
494
+ self._trainer.run_step()
495
+
496
+ def state_dict(self):
497
+ ret = super().state_dict()
498
+ ret["_trainer"] = self._trainer.state_dict()
499
+ return ret
500
+
501
+ def load_state_dict(self, state_dict):
502
+ super().load_state_dict(state_dict)
503
+ self._trainer.load_state_dict(state_dict["_trainer"])
504
+
505
+ @classmethod
506
+ def build_model(cls, cfg):
507
+ """
508
+ Returns:
509
+ torch.nn.Module:
510
+
511
+ It now calls :func:`detectron2.modeling.build_model`.
512
+ Overwrite it if you'd like a different model.
513
+ """
514
+ model = build_model(cfg)
515
+ logger = logging.getLogger(__name__)
516
+ logger.info("Model:\n{}".format(model))
517
+ return model
518
+
519
+ @classmethod
520
+ def build_optimizer(cls, cfg, model):
521
+ """
522
+ Returns:
523
+ torch.optim.Optimizer:
524
+
525
+ It now calls :func:`detectron2.solver.build_optimizer`.
526
+ Overwrite it if you'd like a different optimizer.
527
+ """
528
+ return build_optimizer(cfg, model)
529
+
530
+ @classmethod
531
+ def build_lr_scheduler(cls, cfg, optimizer):
532
+ """
533
+ It now calls :func:`detectron2.solver.build_lr_scheduler`.
534
+ Overwrite it if you'd like a different scheduler.
535
+ """
536
+ return build_lr_scheduler(cfg, optimizer)
537
+
538
+ @classmethod
539
+ def build_train_loader(cls, cfg):
540
+ """
541
+ Returns:
542
+ iterable
543
+
544
+ It now calls :func:`detectron2.data.build_detection_train_loader`.
545
+ Overwrite it if you'd like a different data loader.
546
+ """
547
+ return build_detection_train_loader(cfg)
548
+
549
+ @classmethod
550
+ def build_test_loader(cls, cfg, dataset_name):
551
+ """
552
+ Returns:
553
+ iterable
554
+
555
+ It now calls :func:`detectron2.data.build_detection_test_loader`.
556
+ Overwrite it if you'd like a different data loader.
557
+ """
558
+ return build_detection_test_loader(cfg, dataset_name)
559
+
560
+ @classmethod
561
+ def build_evaluator(cls, cfg, dataset_name):
562
+ """
563
+ Returns:
564
+ DatasetEvaluator or None
565
+
566
+ It is not implemented by default.
567
+ """
568
+ raise NotImplementedError(
569
+ """
570
+ If you want DefaultTrainer to automatically run evaluation,
571
+ please implement `build_evaluator()` in subclasses (see train_net.py for example).
572
+ Alternatively, you can call evaluation functions yourself (see Colab balloon tutorial for example).
573
+ """
574
+ )
575
+
576
+ @classmethod
577
+ def test(cls, cfg, model, evaluators=None):
578
+ """
579
+ Evaluate the given model. The given model is expected to already contain
580
+ weights to evaluate.
581
+
582
+ Args:
583
+ cfg (CfgNode):
584
+ model (nn.Module):
585
+ evaluators (list[DatasetEvaluator] or None): if None, will call
586
+ :meth:`build_evaluator`. Otherwise, must have the same length as
587
+ ``cfg.DATASETS.TEST``.
588
+
589
+ Returns:
590
+ dict: a dict of result metrics
591
+ """
592
+ logger = logging.getLogger(__name__)
593
+ if isinstance(evaluators, DatasetEvaluator):
594
+ evaluators = [evaluators]
595
+ if evaluators is not None:
596
+ assert len(cfg.DATASETS.TEST) == len(evaluators), "{} != {}".format(
597
+ len(cfg.DATASETS.TEST), len(evaluators)
598
+ )
599
+
600
+ results = OrderedDict()
601
+ for idx, dataset_name in enumerate(cfg.DATASETS.TEST):
602
+ data_loader = cls.build_test_loader(cfg, dataset_name)
603
+ # When evaluators are passed in as arguments,
604
+ # implicitly assume that evaluators can be created before data_loader.
605
+ if evaluators is not None:
606
+ evaluator = evaluators[idx]
607
+ else:
608
+ try:
609
+ evaluator = cls.build_evaluator(cfg, dataset_name)
610
+ except NotImplementedError:
611
+ logger.warn(
612
+ "No evaluator found. Use `DefaultTrainer.test(evaluators=)`, "
613
+ "or implement its `build_evaluator` method."
614
+ )
615
+ results[dataset_name] = {}
616
+ continue
617
+ results_i = inference_on_dataset(model, data_loader, evaluator)
618
+ results[dataset_name] = results_i
619
+ if comm.is_main_process():
620
+ assert isinstance(
621
+ results_i, dict
622
+ ), "Evaluator must return a dict on the main process. Got {} instead.".format(
623
+ results_i
624
+ )
625
+ logger.info("Evaluation results for {} in csv format:".format(dataset_name))
626
+ print_csv_format(results_i)
627
+
628
+ if len(results) == 1:
629
+ results = list(results.values())[0]
630
+ return results
631
+
632
+ @staticmethod
633
+ def auto_scale_workers(cfg, num_workers: int):
634
+ """
635
+ When the config is defined for certain number of workers (according to
636
+ ``cfg.SOLVER.REFERENCE_WORLD_SIZE``) that's different from the number of
637
+ workers currently in use, returns a new cfg where the total batch size
638
+ is scaled so that the per-GPU batch size stays the same as the
639
+ original ``IMS_PER_BATCH // REFERENCE_WORLD_SIZE``.
640
+
641
+ Other config options are also scaled accordingly:
642
+ * training steps and warmup steps are scaled inverse proportionally.
643
+ * learning rate are scaled proportionally, following :paper:`ImageNet in 1h`.
644
+
645
+ For example, with the original config like the following:
646
+
647
+ .. code-block:: yaml
648
+
649
+ IMS_PER_BATCH: 16
650
+ BASE_LR: 0.1
651
+ REFERENCE_WORLD_SIZE: 8
652
+ MAX_ITER: 5000
653
+ STEPS: (4000,)
654
+ CHECKPOINT_PERIOD: 1000
655
+
656
+ When this config is used on 16 GPUs instead of the reference number 8,
657
+ calling this method will return a new config with:
658
+
659
+ .. code-block:: yaml
660
+
661
+ IMS_PER_BATCH: 32
662
+ BASE_LR: 0.2
663
+ REFERENCE_WORLD_SIZE: 16
664
+ MAX_ITER: 2500
665
+ STEPS: (2000,)
666
+ CHECKPOINT_PERIOD: 500
667
+
668
+ Note that both the original config and this new config can be trained on 16 GPUs.
669
+ It's up to user whether to enable this feature (by setting ``REFERENCE_WORLD_SIZE``).
670
+
671
+ Returns:
672
+ CfgNode: a new config. Same as original if ``cfg.SOLVER.REFERENCE_WORLD_SIZE==0``.
673
+ """
674
+ old_world_size = cfg.SOLVER.REFERENCE_WORLD_SIZE
675
+ if old_world_size == 0 or old_world_size == num_workers:
676
+ return cfg
677
+ cfg = cfg.clone()
678
+ frozen = cfg.is_frozen()
679
+ cfg.defrost()
680
+
681
+ assert (
682
+ cfg.SOLVER.IMS_PER_BATCH % old_world_size == 0
683
+ ), "Invalid REFERENCE_WORLD_SIZE in config!"
684
+ scale = num_workers / old_world_size
685
+ bs = cfg.SOLVER.IMS_PER_BATCH = int(round(cfg.SOLVER.IMS_PER_BATCH * scale))
686
+ lr = cfg.SOLVER.BASE_LR = cfg.SOLVER.BASE_LR * scale
687
+ max_iter = cfg.SOLVER.MAX_ITER = int(round(cfg.SOLVER.MAX_ITER / scale))
688
+ warmup_iter = cfg.SOLVER.WARMUP_ITERS = int(round(cfg.SOLVER.WARMUP_ITERS / scale))
689
+ cfg.SOLVER.STEPS = tuple(int(round(s / scale)) for s in cfg.SOLVER.STEPS)
690
+ cfg.TEST.EVAL_PERIOD = int(round(cfg.TEST.EVAL_PERIOD / scale))
691
+ cfg.SOLVER.CHECKPOINT_PERIOD = int(round(cfg.SOLVER.CHECKPOINT_PERIOD / scale))
692
+ cfg.SOLVER.REFERENCE_WORLD_SIZE = num_workers # maintain invariant
693
+ logger = logging.getLogger(__name__)
694
+ logger.info(
695
+ f"Auto-scaling the config to batch_size={bs}, learning_rate={lr}, "
696
+ f"max_iter={max_iter}, warmup={warmup_iter}."
697
+ )
698
+
699
+ if frozen:
700
+ cfg.freeze()
701
+ return cfg
702
+
703
+
704
+ # Access basic attributes from the underlying trainer
705
+ for _attr in ["model", "data_loader", "optimizer"]:
706
+ setattr(
707
+ DefaultTrainer,
708
+ _attr,
709
+ property(
710
+ # getter
711
+ lambda self, x=_attr: getattr(self._trainer, x),
712
+ # setter
713
+ lambda self, value, x=_attr: setattr(self._trainer, x, value),
714
+ ),
715
+ )
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/hooks.py ADDED
@@ -0,0 +1,690 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ import datetime
5
+ import itertools
6
+ import logging
7
+ import math
8
+ import operator
9
+ import os
10
+ import tempfile
11
+ import time
12
+ import warnings
13
+ from collections import Counter
14
+ import torch
15
+ from fvcore.common.checkpoint import Checkpointer
16
+ from fvcore.common.checkpoint import PeriodicCheckpointer as _PeriodicCheckpointer
17
+ from fvcore.common.param_scheduler import ParamScheduler
18
+ from fvcore.common.timer import Timer
19
+ from fvcore.nn.precise_bn import get_bn_modules, update_bn_stats
20
+
21
+ import annotator.oneformer.detectron2.utils.comm as comm
22
+ from annotator.oneformer.detectron2.evaluation.testing import flatten_results_dict
23
+ from annotator.oneformer.detectron2.solver import LRMultiplier
24
+ from annotator.oneformer.detectron2.solver import LRScheduler as _LRScheduler
25
+ from annotator.oneformer.detectron2.utils.events import EventStorage, EventWriter
26
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
27
+
28
+ from .train_loop import HookBase
29
+
30
+ __all__ = [
31
+ "CallbackHook",
32
+ "IterationTimer",
33
+ "PeriodicWriter",
34
+ "PeriodicCheckpointer",
35
+ "BestCheckpointer",
36
+ "LRScheduler",
37
+ "AutogradProfiler",
38
+ "EvalHook",
39
+ "PreciseBN",
40
+ "TorchProfiler",
41
+ "TorchMemoryStats",
42
+ ]
43
+
44
+
45
+ """
46
+ Implement some common hooks.
47
+ """
48
+
49
+
50
+ class CallbackHook(HookBase):
51
+ """
52
+ Create a hook using callback functions provided by the user.
53
+ """
54
+
55
+ def __init__(self, *, before_train=None, after_train=None, before_step=None, after_step=None):
56
+ """
57
+ Each argument is a function that takes one argument: the trainer.
58
+ """
59
+ self._before_train = before_train
60
+ self._before_step = before_step
61
+ self._after_step = after_step
62
+ self._after_train = after_train
63
+
64
+ def before_train(self):
65
+ if self._before_train:
66
+ self._before_train(self.trainer)
67
+
68
+ def after_train(self):
69
+ if self._after_train:
70
+ self._after_train(self.trainer)
71
+ # The functions may be closures that hold reference to the trainer
72
+ # Therefore, delete them to avoid circular reference.
73
+ del self._before_train, self._after_train
74
+ del self._before_step, self._after_step
75
+
76
+ def before_step(self):
77
+ if self._before_step:
78
+ self._before_step(self.trainer)
79
+
80
+ def after_step(self):
81
+ if self._after_step:
82
+ self._after_step(self.trainer)
83
+
84
+
85
+ class IterationTimer(HookBase):
86
+ """
87
+ Track the time spent for each iteration (each run_step call in the trainer).
88
+ Print a summary in the end of training.
89
+
90
+ This hook uses the time between the call to its :meth:`before_step`
91
+ and :meth:`after_step` methods.
92
+ Under the convention that :meth:`before_step` of all hooks should only
93
+ take negligible amount of time, the :class:`IterationTimer` hook should be
94
+ placed at the beginning of the list of hooks to obtain accurate timing.
95
+ """
96
+
97
+ def __init__(self, warmup_iter=3):
98
+ """
99
+ Args:
100
+ warmup_iter (int): the number of iterations at the beginning to exclude
101
+ from timing.
102
+ """
103
+ self._warmup_iter = warmup_iter
104
+ self._step_timer = Timer()
105
+ self._start_time = time.perf_counter()
106
+ self._total_timer = Timer()
107
+
108
+ def before_train(self):
109
+ self._start_time = time.perf_counter()
110
+ self._total_timer.reset()
111
+ self._total_timer.pause()
112
+
113
+ def after_train(self):
114
+ logger = logging.getLogger(__name__)
115
+ total_time = time.perf_counter() - self._start_time
116
+ total_time_minus_hooks = self._total_timer.seconds()
117
+ hook_time = total_time - total_time_minus_hooks
118
+
119
+ num_iter = self.trainer.storage.iter + 1 - self.trainer.start_iter - self._warmup_iter
120
+
121
+ if num_iter > 0 and total_time_minus_hooks > 0:
122
+ # Speed is meaningful only after warmup
123
+ # NOTE this format is parsed by grep in some scripts
124
+ logger.info(
125
+ "Overall training speed: {} iterations in {} ({:.4f} s / it)".format(
126
+ num_iter,
127
+ str(datetime.timedelta(seconds=int(total_time_minus_hooks))),
128
+ total_time_minus_hooks / num_iter,
129
+ )
130
+ )
131
+
132
+ logger.info(
133
+ "Total training time: {} ({} on hooks)".format(
134
+ str(datetime.timedelta(seconds=int(total_time))),
135
+ str(datetime.timedelta(seconds=int(hook_time))),
136
+ )
137
+ )
138
+
139
+ def before_step(self):
140
+ self._step_timer.reset()
141
+ self._total_timer.resume()
142
+
143
+ def after_step(self):
144
+ # +1 because we're in after_step, the current step is done
145
+ # but not yet counted
146
+ iter_done = self.trainer.storage.iter - self.trainer.start_iter + 1
147
+ if iter_done >= self._warmup_iter:
148
+ sec = self._step_timer.seconds()
149
+ self.trainer.storage.put_scalars(time=sec)
150
+ else:
151
+ self._start_time = time.perf_counter()
152
+ self._total_timer.reset()
153
+
154
+ self._total_timer.pause()
155
+
156
+
157
+ class PeriodicWriter(HookBase):
158
+ """
159
+ Write events to EventStorage (by calling ``writer.write()``) periodically.
160
+
161
+ It is executed every ``period`` iterations and after the last iteration.
162
+ Note that ``period`` does not affect how data is smoothed by each writer.
163
+ """
164
+
165
+ def __init__(self, writers, period=20):
166
+ """
167
+ Args:
168
+ writers (list[EventWriter]): a list of EventWriter objects
169
+ period (int):
170
+ """
171
+ self._writers = writers
172
+ for w in writers:
173
+ assert isinstance(w, EventWriter), w
174
+ self._period = period
175
+
176
+ def after_step(self):
177
+ if (self.trainer.iter + 1) % self._period == 0 or (
178
+ self.trainer.iter == self.trainer.max_iter - 1
179
+ ):
180
+ for writer in self._writers:
181
+ writer.write()
182
+
183
+ def after_train(self):
184
+ for writer in self._writers:
185
+ # If any new data is found (e.g. produced by other after_train),
186
+ # write them before closing
187
+ writer.write()
188
+ writer.close()
189
+
190
+
191
+ class PeriodicCheckpointer(_PeriodicCheckpointer, HookBase):
192
+ """
193
+ Same as :class:`detectron2.checkpoint.PeriodicCheckpointer`, but as a hook.
194
+
195
+ Note that when used as a hook,
196
+ it is unable to save additional data other than what's defined
197
+ by the given `checkpointer`.
198
+
199
+ It is executed every ``period`` iterations and after the last iteration.
200
+ """
201
+
202
+ def before_train(self):
203
+ self.max_iter = self.trainer.max_iter
204
+
205
+ def after_step(self):
206
+ # No way to use **kwargs
207
+ self.step(self.trainer.iter)
208
+
209
+
210
+ class BestCheckpointer(HookBase):
211
+ """
212
+ Checkpoints best weights based off given metric.
213
+
214
+ This hook should be used in conjunction to and executed after the hook
215
+ that produces the metric, e.g. `EvalHook`.
216
+ """
217
+
218
+ def __init__(
219
+ self,
220
+ eval_period: int,
221
+ checkpointer: Checkpointer,
222
+ val_metric: str,
223
+ mode: str = "max",
224
+ file_prefix: str = "model_best",
225
+ ) -> None:
226
+ """
227
+ Args:
228
+ eval_period (int): the period `EvalHook` is set to run.
229
+ checkpointer: the checkpointer object used to save checkpoints.
230
+ val_metric (str): validation metric to track for best checkpoint, e.g. "bbox/AP50"
231
+ mode (str): one of {'max', 'min'}. controls whether the chosen val metric should be
232
+ maximized or minimized, e.g. for "bbox/AP50" it should be "max"
233
+ file_prefix (str): the prefix of checkpoint's filename, defaults to "model_best"
234
+ """
235
+ self._logger = logging.getLogger(__name__)
236
+ self._period = eval_period
237
+ self._val_metric = val_metric
238
+ assert mode in [
239
+ "max",
240
+ "min",
241
+ ], f'Mode "{mode}" to `BestCheckpointer` is unknown. It should be one of {"max", "min"}.'
242
+ if mode == "max":
243
+ self._compare = operator.gt
244
+ else:
245
+ self._compare = operator.lt
246
+ self._checkpointer = checkpointer
247
+ self._file_prefix = file_prefix
248
+ self.best_metric = None
249
+ self.best_iter = None
250
+
251
+ def _update_best(self, val, iteration):
252
+ if math.isnan(val) or math.isinf(val):
253
+ return False
254
+ self.best_metric = val
255
+ self.best_iter = iteration
256
+ return True
257
+
258
+ def _best_checking(self):
259
+ metric_tuple = self.trainer.storage.latest().get(self._val_metric)
260
+ if metric_tuple is None:
261
+ self._logger.warning(
262
+ f"Given val metric {self._val_metric} does not seem to be computed/stored."
263
+ "Will not be checkpointing based on it."
264
+ )
265
+ return
266
+ else:
267
+ latest_metric, metric_iter = metric_tuple
268
+
269
+ if self.best_metric is None:
270
+ if self._update_best(latest_metric, metric_iter):
271
+ additional_state = {"iteration": metric_iter}
272
+ self._checkpointer.save(f"{self._file_prefix}", **additional_state)
273
+ self._logger.info(
274
+ f"Saved first model at {self.best_metric:0.5f} @ {self.best_iter} steps"
275
+ )
276
+ elif self._compare(latest_metric, self.best_metric):
277
+ additional_state = {"iteration": metric_iter}
278
+ self._checkpointer.save(f"{self._file_prefix}", **additional_state)
279
+ self._logger.info(
280
+ f"Saved best model as latest eval score for {self._val_metric} is "
281
+ f"{latest_metric:0.5f}, better than last best score "
282
+ f"{self.best_metric:0.5f} @ iteration {self.best_iter}."
283
+ )
284
+ self._update_best(latest_metric, metric_iter)
285
+ else:
286
+ self._logger.info(
287
+ f"Not saving as latest eval score for {self._val_metric} is {latest_metric:0.5f}, "
288
+ f"not better than best score {self.best_metric:0.5f} @ iteration {self.best_iter}."
289
+ )
290
+
291
+ def after_step(self):
292
+ # same conditions as `EvalHook`
293
+ next_iter = self.trainer.iter + 1
294
+ if (
295
+ self._period > 0
296
+ and next_iter % self._period == 0
297
+ and next_iter != self.trainer.max_iter
298
+ ):
299
+ self._best_checking()
300
+
301
+ def after_train(self):
302
+ # same conditions as `EvalHook`
303
+ if self.trainer.iter + 1 >= self.trainer.max_iter:
304
+ self._best_checking()
305
+
306
+
307
+ class LRScheduler(HookBase):
308
+ """
309
+ A hook which executes a torch builtin LR scheduler and summarizes the LR.
310
+ It is executed after every iteration.
311
+ """
312
+
313
+ def __init__(self, optimizer=None, scheduler=None):
314
+ """
315
+ Args:
316
+ optimizer (torch.optim.Optimizer):
317
+ scheduler (torch.optim.LRScheduler or fvcore.common.param_scheduler.ParamScheduler):
318
+ if a :class:`ParamScheduler` object, it defines the multiplier over the base LR
319
+ in the optimizer.
320
+
321
+ If any argument is not given, will try to obtain it from the trainer.
322
+ """
323
+ self._optimizer = optimizer
324
+ self._scheduler = scheduler
325
+
326
+ def before_train(self):
327
+ self._optimizer = self._optimizer or self.trainer.optimizer
328
+ if isinstance(self.scheduler, ParamScheduler):
329
+ self._scheduler = LRMultiplier(
330
+ self._optimizer,
331
+ self.scheduler,
332
+ self.trainer.max_iter,
333
+ last_iter=self.trainer.iter - 1,
334
+ )
335
+ self._best_param_group_id = LRScheduler.get_best_param_group_id(self._optimizer)
336
+
337
+ @staticmethod
338
+ def get_best_param_group_id(optimizer):
339
+ # NOTE: some heuristics on what LR to summarize
340
+ # summarize the param group with most parameters
341
+ largest_group = max(len(g["params"]) for g in optimizer.param_groups)
342
+
343
+ if largest_group == 1:
344
+ # If all groups have one parameter,
345
+ # then find the most common initial LR, and use it for summary
346
+ lr_count = Counter([g["lr"] for g in optimizer.param_groups])
347
+ lr = lr_count.most_common()[0][0]
348
+ for i, g in enumerate(optimizer.param_groups):
349
+ if g["lr"] == lr:
350
+ return i
351
+ else:
352
+ for i, g in enumerate(optimizer.param_groups):
353
+ if len(g["params"]) == largest_group:
354
+ return i
355
+
356
+ def after_step(self):
357
+ lr = self._optimizer.param_groups[self._best_param_group_id]["lr"]
358
+ self.trainer.storage.put_scalar("lr", lr, smoothing_hint=False)
359
+ self.scheduler.step()
360
+
361
+ @property
362
+ def scheduler(self):
363
+ return self._scheduler or self.trainer.scheduler
364
+
365
+ def state_dict(self):
366
+ if isinstance(self.scheduler, _LRScheduler):
367
+ return self.scheduler.state_dict()
368
+ return {}
369
+
370
+ def load_state_dict(self, state_dict):
371
+ if isinstance(self.scheduler, _LRScheduler):
372
+ logger = logging.getLogger(__name__)
373
+ logger.info("Loading scheduler from state_dict ...")
374
+ self.scheduler.load_state_dict(state_dict)
375
+
376
+
377
+ class TorchProfiler(HookBase):
378
+ """
379
+ A hook which runs `torch.profiler.profile`.
380
+
381
+ Examples:
382
+ ::
383
+ hooks.TorchProfiler(
384
+ lambda trainer: 10 < trainer.iter < 20, self.cfg.OUTPUT_DIR
385
+ )
386
+
387
+ The above example will run the profiler for iteration 10~20 and dump
388
+ results to ``OUTPUT_DIR``. We did not profile the first few iterations
389
+ because they are typically slower than the rest.
390
+ The result files can be loaded in the ``chrome://tracing`` page in chrome browser,
391
+ and the tensorboard visualizations can be visualized using
392
+ ``tensorboard --logdir OUTPUT_DIR/log``
393
+ """
394
+
395
+ def __init__(self, enable_predicate, output_dir, *, activities=None, save_tensorboard=True):
396
+ """
397
+ Args:
398
+ enable_predicate (callable[trainer -> bool]): a function which takes a trainer,
399
+ and returns whether to enable the profiler.
400
+ It will be called once every step, and can be used to select which steps to profile.
401
+ output_dir (str): the output directory to dump tracing files.
402
+ activities (iterable): same as in `torch.profiler.profile`.
403
+ save_tensorboard (bool): whether to save tensorboard visualizations at (output_dir)/log/
404
+ """
405
+ self._enable_predicate = enable_predicate
406
+ self._activities = activities
407
+ self._output_dir = output_dir
408
+ self._save_tensorboard = save_tensorboard
409
+
410
+ def before_step(self):
411
+ if self._enable_predicate(self.trainer):
412
+ if self._save_tensorboard:
413
+ on_trace_ready = torch.profiler.tensorboard_trace_handler(
414
+ os.path.join(
415
+ self._output_dir,
416
+ "log",
417
+ "profiler-tensorboard-iter{}".format(self.trainer.iter),
418
+ ),
419
+ f"worker{comm.get_rank()}",
420
+ )
421
+ else:
422
+ on_trace_ready = None
423
+ self._profiler = torch.profiler.profile(
424
+ activities=self._activities,
425
+ on_trace_ready=on_trace_ready,
426
+ record_shapes=True,
427
+ profile_memory=True,
428
+ with_stack=True,
429
+ with_flops=True,
430
+ )
431
+ self._profiler.__enter__()
432
+ else:
433
+ self._profiler = None
434
+
435
+ def after_step(self):
436
+ if self._profiler is None:
437
+ return
438
+ self._profiler.__exit__(None, None, None)
439
+ if not self._save_tensorboard:
440
+ PathManager.mkdirs(self._output_dir)
441
+ out_file = os.path.join(
442
+ self._output_dir, "profiler-trace-iter{}.json".format(self.trainer.iter)
443
+ )
444
+ if "://" not in out_file:
445
+ self._profiler.export_chrome_trace(out_file)
446
+ else:
447
+ # Support non-posix filesystems
448
+ with tempfile.TemporaryDirectory(prefix="detectron2_profiler") as d:
449
+ tmp_file = os.path.join(d, "tmp.json")
450
+ self._profiler.export_chrome_trace(tmp_file)
451
+ with open(tmp_file) as f:
452
+ content = f.read()
453
+ with PathManager.open(out_file, "w") as f:
454
+ f.write(content)
455
+
456
+
457
+ class AutogradProfiler(TorchProfiler):
458
+ """
459
+ A hook which runs `torch.autograd.profiler.profile`.
460
+
461
+ Examples:
462
+ ::
463
+ hooks.AutogradProfiler(
464
+ lambda trainer: 10 < trainer.iter < 20, self.cfg.OUTPUT_DIR
465
+ )
466
+
467
+ The above example will run the profiler for iteration 10~20 and dump
468
+ results to ``OUTPUT_DIR``. We did not profile the first few iterations
469
+ because they are typically slower than the rest.
470
+ The result files can be loaded in the ``chrome://tracing`` page in chrome browser.
471
+
472
+ Note:
473
+ When used together with NCCL on older version of GPUs,
474
+ autograd profiler may cause deadlock because it unnecessarily allocates
475
+ memory on every device it sees. The memory management calls, if
476
+ interleaved with NCCL calls, lead to deadlock on GPUs that do not
477
+ support ``cudaLaunchCooperativeKernelMultiDevice``.
478
+ """
479
+
480
+ def __init__(self, enable_predicate, output_dir, *, use_cuda=True):
481
+ """
482
+ Args:
483
+ enable_predicate (callable[trainer -> bool]): a function which takes a trainer,
484
+ and returns whether to enable the profiler.
485
+ It will be called once every step, and can be used to select which steps to profile.
486
+ output_dir (str): the output directory to dump tracing files.
487
+ use_cuda (bool): same as in `torch.autograd.profiler.profile`.
488
+ """
489
+ warnings.warn("AutogradProfiler has been deprecated in favor of TorchProfiler.")
490
+ self._enable_predicate = enable_predicate
491
+ self._use_cuda = use_cuda
492
+ self._output_dir = output_dir
493
+
494
+ def before_step(self):
495
+ if self._enable_predicate(self.trainer):
496
+ self._profiler = torch.autograd.profiler.profile(use_cuda=self._use_cuda)
497
+ self._profiler.__enter__()
498
+ else:
499
+ self._profiler = None
500
+
501
+
502
+ class EvalHook(HookBase):
503
+ """
504
+ Run an evaluation function periodically, and at the end of training.
505
+
506
+ It is executed every ``eval_period`` iterations and after the last iteration.
507
+ """
508
+
509
+ def __init__(self, eval_period, eval_function, eval_after_train=True):
510
+ """
511
+ Args:
512
+ eval_period (int): the period to run `eval_function`. Set to 0 to
513
+ not evaluate periodically (but still evaluate after the last iteration
514
+ if `eval_after_train` is True).
515
+ eval_function (callable): a function which takes no arguments, and
516
+ returns a nested dict of evaluation metrics.
517
+ eval_after_train (bool): whether to evaluate after the last iteration
518
+
519
+ Note:
520
+ This hook must be enabled in all or none workers.
521
+ If you would like only certain workers to perform evaluation,
522
+ give other workers a no-op function (`eval_function=lambda: None`).
523
+ """
524
+ self._period = eval_period
525
+ self._func = eval_function
526
+ self._eval_after_train = eval_after_train
527
+
528
+ def _do_eval(self):
529
+ results = self._func()
530
+
531
+ if results:
532
+ assert isinstance(
533
+ results, dict
534
+ ), "Eval function must return a dict. Got {} instead.".format(results)
535
+
536
+ flattened_results = flatten_results_dict(results)
537
+ for k, v in flattened_results.items():
538
+ try:
539
+ v = float(v)
540
+ except Exception as e:
541
+ raise ValueError(
542
+ "[EvalHook] eval_function should return a nested dict of float. "
543
+ "Got '{}: {}' instead.".format(k, v)
544
+ ) from e
545
+ self.trainer.storage.put_scalars(**flattened_results, smoothing_hint=False)
546
+
547
+ # Evaluation may take different time among workers.
548
+ # A barrier make them start the next iteration together.
549
+ comm.synchronize()
550
+
551
+ def after_step(self):
552
+ next_iter = self.trainer.iter + 1
553
+ if self._period > 0 and next_iter % self._period == 0:
554
+ # do the last eval in after_train
555
+ if next_iter != self.trainer.max_iter:
556
+ self._do_eval()
557
+
558
+ def after_train(self):
559
+ # This condition is to prevent the eval from running after a failed training
560
+ if self._eval_after_train and self.trainer.iter + 1 >= self.trainer.max_iter:
561
+ self._do_eval()
562
+ # func is likely a closure that holds reference to the trainer
563
+ # therefore we clean it to avoid circular reference in the end
564
+ del self._func
565
+
566
+
567
+ class PreciseBN(HookBase):
568
+ """
569
+ The standard implementation of BatchNorm uses EMA in inference, which is
570
+ sometimes suboptimal.
571
+ This class computes the true average of statistics rather than the moving average,
572
+ and put true averages to every BN layer in the given model.
573
+
574
+ It is executed every ``period`` iterations and after the last iteration.
575
+ """
576
+
577
+ def __init__(self, period, model, data_loader, num_iter):
578
+ """
579
+ Args:
580
+ period (int): the period this hook is run, or 0 to not run during training.
581
+ The hook will always run in the end of training.
582
+ model (nn.Module): a module whose all BN layers in training mode will be
583
+ updated by precise BN.
584
+ Note that user is responsible for ensuring the BN layers to be
585
+ updated are in training mode when this hook is triggered.
586
+ data_loader (iterable): it will produce data to be run by `model(data)`.
587
+ num_iter (int): number of iterations used to compute the precise
588
+ statistics.
589
+ """
590
+ self._logger = logging.getLogger(__name__)
591
+ if len(get_bn_modules(model)) == 0:
592
+ self._logger.info(
593
+ "PreciseBN is disabled because model does not contain BN layers in training mode."
594
+ )
595
+ self._disabled = True
596
+ return
597
+
598
+ self._model = model
599
+ self._data_loader = data_loader
600
+ self._num_iter = num_iter
601
+ self._period = period
602
+ self._disabled = False
603
+
604
+ self._data_iter = None
605
+
606
+ def after_step(self):
607
+ next_iter = self.trainer.iter + 1
608
+ is_final = next_iter == self.trainer.max_iter
609
+ if is_final or (self._period > 0 and next_iter % self._period == 0):
610
+ self.update_stats()
611
+
612
+ def update_stats(self):
613
+ """
614
+ Update the model with precise statistics. Users can manually call this method.
615
+ """
616
+ if self._disabled:
617
+ return
618
+
619
+ if self._data_iter is None:
620
+ self._data_iter = iter(self._data_loader)
621
+
622
+ def data_loader():
623
+ for num_iter in itertools.count(1):
624
+ if num_iter % 100 == 0:
625
+ self._logger.info(
626
+ "Running precise-BN ... {}/{} iterations.".format(num_iter, self._num_iter)
627
+ )
628
+ # This way we can reuse the same iterator
629
+ yield next(self._data_iter)
630
+
631
+ with EventStorage(): # capture events in a new storage to discard them
632
+ self._logger.info(
633
+ "Running precise-BN for {} iterations... ".format(self._num_iter)
634
+ + "Note that this could produce different statistics every time."
635
+ )
636
+ update_bn_stats(self._model, data_loader(), self._num_iter)
637
+
638
+
639
+ class TorchMemoryStats(HookBase):
640
+ """
641
+ Writes pytorch's cuda memory statistics periodically.
642
+ """
643
+
644
+ def __init__(self, period=20, max_runs=10):
645
+ """
646
+ Args:
647
+ period (int): Output stats each 'period' iterations
648
+ max_runs (int): Stop the logging after 'max_runs'
649
+ """
650
+
651
+ self._logger = logging.getLogger(__name__)
652
+ self._period = period
653
+ self._max_runs = max_runs
654
+ self._runs = 0
655
+
656
+ def after_step(self):
657
+ if self._runs > self._max_runs:
658
+ return
659
+
660
+ if (self.trainer.iter + 1) % self._period == 0 or (
661
+ self.trainer.iter == self.trainer.max_iter - 1
662
+ ):
663
+ if torch.cuda.is_available():
664
+ max_reserved_mb = torch.cuda.max_memory_reserved() / 1024.0 / 1024.0
665
+ reserved_mb = torch.cuda.memory_reserved() / 1024.0 / 1024.0
666
+ max_allocated_mb = torch.cuda.max_memory_allocated() / 1024.0 / 1024.0
667
+ allocated_mb = torch.cuda.memory_allocated() / 1024.0 / 1024.0
668
+
669
+ self._logger.info(
670
+ (
671
+ " iter: {} "
672
+ " max_reserved_mem: {:.0f}MB "
673
+ " reserved_mem: {:.0f}MB "
674
+ " max_allocated_mem: {:.0f}MB "
675
+ " allocated_mem: {:.0f}MB "
676
+ ).format(
677
+ self.trainer.iter,
678
+ max_reserved_mb,
679
+ reserved_mb,
680
+ max_allocated_mb,
681
+ allocated_mb,
682
+ )
683
+ )
684
+
685
+ self._runs += 1
686
+ if self._runs == self._max_runs:
687
+ mem_summary = torch.cuda.memory_summary()
688
+ self._logger.info("\n" + mem_summary)
689
+
690
+ torch.cuda.reset_peak_memory_stats()
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/launch.py ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import logging
3
+ from datetime import timedelta
4
+ import torch
5
+ import torch.distributed as dist
6
+ import torch.multiprocessing as mp
7
+
8
+ from annotator.oneformer.detectron2.utils import comm
9
+
10
+ __all__ = ["DEFAULT_TIMEOUT", "launch"]
11
+
12
+ DEFAULT_TIMEOUT = timedelta(minutes=30)
13
+
14
+
15
+ def _find_free_port():
16
+ import socket
17
+
18
+ sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
19
+ # Binding to port 0 will cause the OS to find an available port for us
20
+ sock.bind(("", 0))
21
+ port = sock.getsockname()[1]
22
+ sock.close()
23
+ # NOTE: there is still a chance the port could be taken by other processes.
24
+ return port
25
+
26
+
27
+ def launch(
28
+ main_func,
29
+ # Should be num_processes_per_machine, but kept for compatibility.
30
+ num_gpus_per_machine,
31
+ num_machines=1,
32
+ machine_rank=0,
33
+ dist_url=None,
34
+ args=(),
35
+ timeout=DEFAULT_TIMEOUT,
36
+ ):
37
+ """
38
+ Launch multi-process or distributed training.
39
+ This function must be called on all machines involved in the training.
40
+ It will spawn child processes (defined by ``num_gpus_per_machine``) on each machine.
41
+
42
+ Args:
43
+ main_func: a function that will be called by `main_func(*args)`
44
+ num_gpus_per_machine (int): number of processes per machine. When
45
+ using GPUs, this should be the number of GPUs.
46
+ num_machines (int): the total number of machines
47
+ machine_rank (int): the rank of this machine
48
+ dist_url (str): url to connect to for distributed jobs, including protocol
49
+ e.g. "tcp://127.0.0.1:8686".
50
+ Can be set to "auto" to automatically select a free port on localhost
51
+ timeout (timedelta): timeout of the distributed workers
52
+ args (tuple): arguments passed to main_func
53
+ """
54
+ world_size = num_machines * num_gpus_per_machine
55
+ if world_size > 1:
56
+ # https://github.com/pytorch/pytorch/pull/14391
57
+ # TODO prctl in spawned processes
58
+
59
+ if dist_url == "auto":
60
+ assert num_machines == 1, "dist_url=auto not supported in multi-machine jobs."
61
+ port = _find_free_port()
62
+ dist_url = f"tcp://127.0.0.1:{port}"
63
+ if num_machines > 1 and dist_url.startswith("file://"):
64
+ logger = logging.getLogger(__name__)
65
+ logger.warning(
66
+ "file:// is not a reliable init_method in multi-machine jobs. Prefer tcp://"
67
+ )
68
+
69
+ mp.start_processes(
70
+ _distributed_worker,
71
+ nprocs=num_gpus_per_machine,
72
+ args=(
73
+ main_func,
74
+ world_size,
75
+ num_gpus_per_machine,
76
+ machine_rank,
77
+ dist_url,
78
+ args,
79
+ timeout,
80
+ ),
81
+ daemon=False,
82
+ )
83
+ else:
84
+ main_func(*args)
85
+
86
+
87
+ def _distributed_worker(
88
+ local_rank,
89
+ main_func,
90
+ world_size,
91
+ num_gpus_per_machine,
92
+ machine_rank,
93
+ dist_url,
94
+ args,
95
+ timeout=DEFAULT_TIMEOUT,
96
+ ):
97
+ has_gpu = torch.cuda.is_available()
98
+ if has_gpu:
99
+ assert num_gpus_per_machine <= torch.cuda.device_count()
100
+ global_rank = machine_rank * num_gpus_per_machine + local_rank
101
+ try:
102
+ dist.init_process_group(
103
+ backend="NCCL" if has_gpu else "GLOO",
104
+ init_method=dist_url,
105
+ world_size=world_size,
106
+ rank=global_rank,
107
+ timeout=timeout,
108
+ )
109
+ except Exception as e:
110
+ logger = logging.getLogger(__name__)
111
+ logger.error("Process group URL: {}".format(dist_url))
112
+ raise e
113
+
114
+ # Setup the local process group.
115
+ comm.create_local_process_group(num_gpus_per_machine)
116
+ if has_gpu:
117
+ torch.cuda.set_device(local_rank)
118
+
119
+ # synchronize is needed here to prevent a possible timeout after calling init_process_group
120
+ # See: https://github.com/facebookresearch/maskrcnn-benchmark/issues/172
121
+ comm.synchronize()
122
+
123
+ main_func(*args)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/engine/train_loop.py ADDED
@@ -0,0 +1,469 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ import logging
5
+ import numpy as np
6
+ import time
7
+ import weakref
8
+ from typing import List, Mapping, Optional
9
+ import torch
10
+ from torch.nn.parallel import DataParallel, DistributedDataParallel
11
+
12
+ import annotator.oneformer.detectron2.utils.comm as comm
13
+ from annotator.oneformer.detectron2.utils.events import EventStorage, get_event_storage
14
+ from annotator.oneformer.detectron2.utils.logger import _log_api_usage
15
+
16
+ __all__ = ["HookBase", "TrainerBase", "SimpleTrainer", "AMPTrainer"]
17
+
18
+
19
+ class HookBase:
20
+ """
21
+ Base class for hooks that can be registered with :class:`TrainerBase`.
22
+
23
+ Each hook can implement 4 methods. The way they are called is demonstrated
24
+ in the following snippet:
25
+ ::
26
+ hook.before_train()
27
+ for iter in range(start_iter, max_iter):
28
+ hook.before_step()
29
+ trainer.run_step()
30
+ hook.after_step()
31
+ iter += 1
32
+ hook.after_train()
33
+
34
+ Notes:
35
+ 1. In the hook method, users can access ``self.trainer`` to access more
36
+ properties about the context (e.g., model, current iteration, or config
37
+ if using :class:`DefaultTrainer`).
38
+
39
+ 2. A hook that does something in :meth:`before_step` can often be
40
+ implemented equivalently in :meth:`after_step`.
41
+ If the hook takes non-trivial time, it is strongly recommended to
42
+ implement the hook in :meth:`after_step` instead of :meth:`before_step`.
43
+ The convention is that :meth:`before_step` should only take negligible time.
44
+
45
+ Following this convention will allow hooks that do care about the difference
46
+ between :meth:`before_step` and :meth:`after_step` (e.g., timer) to
47
+ function properly.
48
+
49
+ """
50
+
51
+ trainer: "TrainerBase" = None
52
+ """
53
+ A weak reference to the trainer object. Set by the trainer when the hook is registered.
54
+ """
55
+
56
+ def before_train(self):
57
+ """
58
+ Called before the first iteration.
59
+ """
60
+ pass
61
+
62
+ def after_train(self):
63
+ """
64
+ Called after the last iteration.
65
+ """
66
+ pass
67
+
68
+ def before_step(self):
69
+ """
70
+ Called before each iteration.
71
+ """
72
+ pass
73
+
74
+ def after_backward(self):
75
+ """
76
+ Called after the backward pass of each iteration.
77
+ """
78
+ pass
79
+
80
+ def after_step(self):
81
+ """
82
+ Called after each iteration.
83
+ """
84
+ pass
85
+
86
+ def state_dict(self):
87
+ """
88
+ Hooks are stateless by default, but can be made checkpointable by
89
+ implementing `state_dict` and `load_state_dict`.
90
+ """
91
+ return {}
92
+
93
+
94
+ class TrainerBase:
95
+ """
96
+ Base class for iterative trainer with hooks.
97
+
98
+ The only assumption we made here is: the training runs in a loop.
99
+ A subclass can implement what the loop is.
100
+ We made no assumptions about the existence of dataloader, optimizer, model, etc.
101
+
102
+ Attributes:
103
+ iter(int): the current iteration.
104
+
105
+ start_iter(int): The iteration to start with.
106
+ By convention the minimum possible value is 0.
107
+
108
+ max_iter(int): The iteration to end training.
109
+
110
+ storage(EventStorage): An EventStorage that's opened during the course of training.
111
+ """
112
+
113
+ def __init__(self) -> None:
114
+ self._hooks: List[HookBase] = []
115
+ self.iter: int = 0
116
+ self.start_iter: int = 0
117
+ self.max_iter: int
118
+ self.storage: EventStorage
119
+ _log_api_usage("trainer." + self.__class__.__name__)
120
+
121
+ def register_hooks(self, hooks: List[Optional[HookBase]]) -> None:
122
+ """
123
+ Register hooks to the trainer. The hooks are executed in the order
124
+ they are registered.
125
+
126
+ Args:
127
+ hooks (list[Optional[HookBase]]): list of hooks
128
+ """
129
+ hooks = [h for h in hooks if h is not None]
130
+ for h in hooks:
131
+ assert isinstance(h, HookBase)
132
+ # To avoid circular reference, hooks and trainer cannot own each other.
133
+ # This normally does not matter, but will cause memory leak if the
134
+ # involved objects contain __del__:
135
+ # See http://engineering.hearsaysocial.com/2013/06/16/circular-references-in-python/
136
+ h.trainer = weakref.proxy(self)
137
+ self._hooks.extend(hooks)
138
+
139
+ def train(self, start_iter: int, max_iter: int):
140
+ """
141
+ Args:
142
+ start_iter, max_iter (int): See docs above
143
+ """
144
+ logger = logging.getLogger(__name__)
145
+ logger.info("Starting training from iteration {}".format(start_iter))
146
+
147
+ self.iter = self.start_iter = start_iter
148
+ self.max_iter = max_iter
149
+
150
+ with EventStorage(start_iter) as self.storage:
151
+ try:
152
+ self.before_train()
153
+ for self.iter in range(start_iter, max_iter):
154
+ self.before_step()
155
+ self.run_step()
156
+ self.after_step()
157
+ # self.iter == max_iter can be used by `after_train` to
158
+ # tell whether the training successfully finished or failed
159
+ # due to exceptions.
160
+ self.iter += 1
161
+ except Exception:
162
+ logger.exception("Exception during training:")
163
+ raise
164
+ finally:
165
+ self.after_train()
166
+
167
+ def before_train(self):
168
+ for h in self._hooks:
169
+ h.before_train()
170
+
171
+ def after_train(self):
172
+ self.storage.iter = self.iter
173
+ for h in self._hooks:
174
+ h.after_train()
175
+
176
+ def before_step(self):
177
+ # Maintain the invariant that storage.iter == trainer.iter
178
+ # for the entire execution of each step
179
+ self.storage.iter = self.iter
180
+
181
+ for h in self._hooks:
182
+ h.before_step()
183
+
184
+ def after_backward(self):
185
+ for h in self._hooks:
186
+ h.after_backward()
187
+
188
+ def after_step(self):
189
+ for h in self._hooks:
190
+ h.after_step()
191
+
192
+ def run_step(self):
193
+ raise NotImplementedError
194
+
195
+ def state_dict(self):
196
+ ret = {"iteration": self.iter}
197
+ hooks_state = {}
198
+ for h in self._hooks:
199
+ sd = h.state_dict()
200
+ if sd:
201
+ name = type(h).__qualname__
202
+ if name in hooks_state:
203
+ # TODO handle repetitive stateful hooks
204
+ continue
205
+ hooks_state[name] = sd
206
+ if hooks_state:
207
+ ret["hooks"] = hooks_state
208
+ return ret
209
+
210
+ def load_state_dict(self, state_dict):
211
+ logger = logging.getLogger(__name__)
212
+ self.iter = state_dict["iteration"]
213
+ for key, value in state_dict.get("hooks", {}).items():
214
+ for h in self._hooks:
215
+ try:
216
+ name = type(h).__qualname__
217
+ except AttributeError:
218
+ continue
219
+ if name == key:
220
+ h.load_state_dict(value)
221
+ break
222
+ else:
223
+ logger.warning(f"Cannot find the hook '{key}', its state_dict is ignored.")
224
+
225
+
226
+ class SimpleTrainer(TrainerBase):
227
+ """
228
+ A simple trainer for the most common type of task:
229
+ single-cost single-optimizer single-data-source iterative optimization,
230
+ optionally using data-parallelism.
231
+ It assumes that every step, you:
232
+
233
+ 1. Compute the loss with a data from the data_loader.
234
+ 2. Compute the gradients with the above loss.
235
+ 3. Update the model with the optimizer.
236
+
237
+ All other tasks during training (checkpointing, logging, evaluation, LR schedule)
238
+ are maintained by hooks, which can be registered by :meth:`TrainerBase.register_hooks`.
239
+
240
+ If you want to do anything fancier than this,
241
+ either subclass TrainerBase and implement your own `run_step`,
242
+ or write your own training loop.
243
+ """
244
+
245
+ def __init__(self, model, data_loader, optimizer, gather_metric_period=1):
246
+ """
247
+ Args:
248
+ model: a torch Module. Takes a data from data_loader and returns a
249
+ dict of losses.
250
+ data_loader: an iterable. Contains data to be used to call model.
251
+ optimizer: a torch optimizer.
252
+ gather_metric_period: an int. Every gather_metric_period iterations
253
+ the metrics are gathered from all the ranks to rank 0 and logged.
254
+ """
255
+ super().__init__()
256
+
257
+ """
258
+ We set the model to training mode in the trainer.
259
+ However it's valid to train a model that's in eval mode.
260
+ If you want your model (or a submodule of it) to behave
261
+ like evaluation during training, you can overwrite its train() method.
262
+ """
263
+ model.train()
264
+
265
+ self.model = model
266
+ self.data_loader = data_loader
267
+ # to access the data loader iterator, call `self._data_loader_iter`
268
+ self._data_loader_iter_obj = None
269
+ self.optimizer = optimizer
270
+ self.gather_metric_period = gather_metric_period
271
+
272
+ def run_step(self):
273
+ """
274
+ Implement the standard training logic described above.
275
+ """
276
+ assert self.model.training, "[SimpleTrainer] model was changed to eval mode!"
277
+ start = time.perf_counter()
278
+ """
279
+ If you want to do something with the data, you can wrap the dataloader.
280
+ """
281
+ data = next(self._data_loader_iter)
282
+ data_time = time.perf_counter() - start
283
+
284
+ """
285
+ If you want to do something with the losses, you can wrap the model.
286
+ """
287
+ loss_dict = self.model(data)
288
+ if isinstance(loss_dict, torch.Tensor):
289
+ losses = loss_dict
290
+ loss_dict = {"total_loss": loss_dict}
291
+ else:
292
+ losses = sum(loss_dict.values())
293
+
294
+ """
295
+ If you need to accumulate gradients or do something similar, you can
296
+ wrap the optimizer with your custom `zero_grad()` method.
297
+ """
298
+ self.optimizer.zero_grad()
299
+ losses.backward()
300
+
301
+ self.after_backward()
302
+
303
+ self._write_metrics(loss_dict, data_time)
304
+
305
+ """
306
+ If you need gradient clipping/scaling or other processing, you can
307
+ wrap the optimizer with your custom `step()` method. But it is
308
+ suboptimal as explained in https://arxiv.org/abs/2006.15704 Sec 3.2.4
309
+ """
310
+ self.optimizer.step()
311
+
312
+ @property
313
+ def _data_loader_iter(self):
314
+ # only create the data loader iterator when it is used
315
+ if self._data_loader_iter_obj is None:
316
+ self._data_loader_iter_obj = iter(self.data_loader)
317
+ return self._data_loader_iter_obj
318
+
319
+ def reset_data_loader(self, data_loader_builder):
320
+ """
321
+ Delete and replace the current data loader with a new one, which will be created
322
+ by calling `data_loader_builder` (without argument).
323
+ """
324
+ del self.data_loader
325
+ data_loader = data_loader_builder()
326
+ self.data_loader = data_loader
327
+ self._data_loader_iter_obj = None
328
+
329
+ def _write_metrics(
330
+ self,
331
+ loss_dict: Mapping[str, torch.Tensor],
332
+ data_time: float,
333
+ prefix: str = "",
334
+ ) -> None:
335
+ if (self.iter + 1) % self.gather_metric_period == 0:
336
+ SimpleTrainer.write_metrics(loss_dict, data_time, prefix)
337
+
338
+ @staticmethod
339
+ def write_metrics(
340
+ loss_dict: Mapping[str, torch.Tensor],
341
+ data_time: float,
342
+ prefix: str = "",
343
+ ) -> None:
344
+ """
345
+ Args:
346
+ loss_dict (dict): dict of scalar losses
347
+ data_time (float): time taken by the dataloader iteration
348
+ prefix (str): prefix for logging keys
349
+ """
350
+ metrics_dict = {k: v.detach().cpu().item() for k, v in loss_dict.items()}
351
+ metrics_dict["data_time"] = data_time
352
+
353
+ # Gather metrics among all workers for logging
354
+ # This assumes we do DDP-style training, which is currently the only
355
+ # supported method in detectron2.
356
+ all_metrics_dict = comm.gather(metrics_dict)
357
+
358
+ if comm.is_main_process():
359
+ storage = get_event_storage()
360
+
361
+ # data_time among workers can have high variance. The actual latency
362
+ # caused by data_time is the maximum among workers.
363
+ data_time = np.max([x.pop("data_time") for x in all_metrics_dict])
364
+ storage.put_scalar("data_time", data_time)
365
+
366
+ # average the rest metrics
367
+ metrics_dict = {
368
+ k: np.mean([x[k] for x in all_metrics_dict]) for k in all_metrics_dict[0].keys()
369
+ }
370
+ total_losses_reduced = sum(metrics_dict.values())
371
+ if not np.isfinite(total_losses_reduced):
372
+ raise FloatingPointError(
373
+ f"Loss became infinite or NaN at iteration={storage.iter}!\n"
374
+ f"loss_dict = {metrics_dict}"
375
+ )
376
+
377
+ storage.put_scalar("{}total_loss".format(prefix), total_losses_reduced)
378
+ if len(metrics_dict) > 1:
379
+ storage.put_scalars(**metrics_dict)
380
+
381
+ def state_dict(self):
382
+ ret = super().state_dict()
383
+ ret["optimizer"] = self.optimizer.state_dict()
384
+ return ret
385
+
386
+ def load_state_dict(self, state_dict):
387
+ super().load_state_dict(state_dict)
388
+ self.optimizer.load_state_dict(state_dict["optimizer"])
389
+
390
+
391
+ class AMPTrainer(SimpleTrainer):
392
+ """
393
+ Like :class:`SimpleTrainer`, but uses PyTorch's native automatic mixed precision
394
+ in the training loop.
395
+ """
396
+
397
+ def __init__(
398
+ self,
399
+ model,
400
+ data_loader,
401
+ optimizer,
402
+ gather_metric_period=1,
403
+ grad_scaler=None,
404
+ precision: torch.dtype = torch.float16,
405
+ log_grad_scaler: bool = False,
406
+ ):
407
+ """
408
+ Args:
409
+ model, data_loader, optimizer, gather_metric_period: same as in :class:`SimpleTrainer`.
410
+ grad_scaler: torch GradScaler to automatically scale gradients.
411
+ precision: torch.dtype as the target precision to cast to in computations
412
+ """
413
+ unsupported = "AMPTrainer does not support single-process multi-device training!"
414
+ if isinstance(model, DistributedDataParallel):
415
+ assert not (model.device_ids and len(model.device_ids) > 1), unsupported
416
+ assert not isinstance(model, DataParallel), unsupported
417
+
418
+ super().__init__(model, data_loader, optimizer, gather_metric_period)
419
+
420
+ if grad_scaler is None:
421
+ from torch.cuda.amp import GradScaler
422
+
423
+ grad_scaler = GradScaler()
424
+ self.grad_scaler = grad_scaler
425
+ self.precision = precision
426
+ self.log_grad_scaler = log_grad_scaler
427
+
428
+ def run_step(self):
429
+ """
430
+ Implement the AMP training logic.
431
+ """
432
+ assert self.model.training, "[AMPTrainer] model was changed to eval mode!"
433
+ assert torch.cuda.is_available(), "[AMPTrainer] CUDA is required for AMP training!"
434
+ from torch.cuda.amp import autocast
435
+
436
+ start = time.perf_counter()
437
+ data = next(self._data_loader_iter)
438
+ data_time = time.perf_counter() - start
439
+
440
+ with autocast(dtype=self.precision):
441
+ loss_dict = self.model(data)
442
+ if isinstance(loss_dict, torch.Tensor):
443
+ losses = loss_dict
444
+ loss_dict = {"total_loss": loss_dict}
445
+ else:
446
+ losses = sum(loss_dict.values())
447
+
448
+ self.optimizer.zero_grad()
449
+ self.grad_scaler.scale(losses).backward()
450
+
451
+ if self.log_grad_scaler:
452
+ storage = get_event_storage()
453
+ storage.put_scalar("[metric]grad_scaler", self.grad_scaler.get_scale())
454
+
455
+ self.after_backward()
456
+
457
+ self._write_metrics(loss_dict, data_time)
458
+
459
+ self.grad_scaler.step(self.optimizer)
460
+ self.grad_scaler.update()
461
+
462
+ def state_dict(self):
463
+ ret = super().state_dict()
464
+ ret["grad_scaler"] = self.grad_scaler.state_dict()
465
+ return ret
466
+
467
+ def load_state_dict(self, state_dict):
468
+ super().load_state_dict(state_dict)
469
+ self.grad_scaler.load_state_dict(state_dict["grad_scaler"])
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/__init__.py ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ from .cityscapes_evaluation import CityscapesInstanceEvaluator, CityscapesSemSegEvaluator
3
+ from .coco_evaluation import COCOEvaluator
4
+ from .rotated_coco_evaluation import RotatedCOCOEvaluator
5
+ from .evaluator import DatasetEvaluator, DatasetEvaluators, inference_context, inference_on_dataset
6
+ from .lvis_evaluation import LVISEvaluator
7
+ from .panoptic_evaluation import COCOPanopticEvaluator
8
+ from .pascal_voc_evaluation import PascalVOCDetectionEvaluator
9
+ from .sem_seg_evaluation import SemSegEvaluator
10
+ from .testing import print_csv_format, verify_results
11
+
12
+ __all__ = [k for k in globals().keys() if not k.startswith("_")]
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/cityscapes_evaluation.py ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import glob
3
+ import logging
4
+ import numpy as np
5
+ import os
6
+ import tempfile
7
+ from collections import OrderedDict
8
+ import torch
9
+ from PIL import Image
10
+
11
+ from annotator.oneformer.detectron2.data import MetadataCatalog
12
+ from annotator.oneformer.detectron2.utils import comm
13
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
14
+
15
+ from .evaluator import DatasetEvaluator
16
+
17
+
18
+ class CityscapesEvaluator(DatasetEvaluator):
19
+ """
20
+ Base class for evaluation using cityscapes API.
21
+ """
22
+
23
+ def __init__(self, dataset_name):
24
+ """
25
+ Args:
26
+ dataset_name (str): the name of the dataset.
27
+ It must have the following metadata associated with it:
28
+ "thing_classes", "gt_dir".
29
+ """
30
+ self._metadata = MetadataCatalog.get(dataset_name)
31
+ self._cpu_device = torch.device("cpu")
32
+ self._logger = logging.getLogger(__name__)
33
+
34
+ def reset(self):
35
+ self._working_dir = tempfile.TemporaryDirectory(prefix="cityscapes_eval_")
36
+ self._temp_dir = self._working_dir.name
37
+ # All workers will write to the same results directory
38
+ # TODO this does not work in distributed training
39
+ assert (
40
+ comm.get_local_size() == comm.get_world_size()
41
+ ), "CityscapesEvaluator currently do not work with multiple machines."
42
+ self._temp_dir = comm.all_gather(self._temp_dir)[0]
43
+ if self._temp_dir != self._working_dir.name:
44
+ self._working_dir.cleanup()
45
+ self._logger.info(
46
+ "Writing cityscapes results to temporary directory {} ...".format(self._temp_dir)
47
+ )
48
+
49
+
50
+ class CityscapesInstanceEvaluator(CityscapesEvaluator):
51
+ """
52
+ Evaluate instance segmentation results on cityscapes dataset using cityscapes API.
53
+
54
+ Note:
55
+ * It does not work in multi-machine distributed training.
56
+ * It contains a synchronization, therefore has to be used on all ranks.
57
+ * Only the main process runs evaluation.
58
+ """
59
+
60
+ def process(self, inputs, outputs):
61
+ from cityscapesscripts.helpers.labels import name2label
62
+
63
+ for input, output in zip(inputs, outputs):
64
+ file_name = input["file_name"]
65
+ basename = os.path.splitext(os.path.basename(file_name))[0]
66
+ pred_txt = os.path.join(self._temp_dir, basename + "_pred.txt")
67
+
68
+ if "instances" in output:
69
+ output = output["instances"].to(self._cpu_device)
70
+ num_instances = len(output)
71
+ with open(pred_txt, "w") as fout:
72
+ for i in range(num_instances):
73
+ pred_class = output.pred_classes[i]
74
+ classes = self._metadata.thing_classes[pred_class]
75
+ class_id = name2label[classes].id
76
+ score = output.scores[i]
77
+ mask = output.pred_masks[i].numpy().astype("uint8")
78
+ png_filename = os.path.join(
79
+ self._temp_dir, basename + "_{}_{}.png".format(i, classes)
80
+ )
81
+
82
+ Image.fromarray(mask * 255).save(png_filename)
83
+ fout.write(
84
+ "{} {} {}\n".format(os.path.basename(png_filename), class_id, score)
85
+ )
86
+ else:
87
+ # Cityscapes requires a prediction file for every ground truth image.
88
+ with open(pred_txt, "w") as fout:
89
+ pass
90
+
91
+ def evaluate(self):
92
+ """
93
+ Returns:
94
+ dict: has a key "segm", whose value is a dict of "AP" and "AP50".
95
+ """
96
+ comm.synchronize()
97
+ if comm.get_rank() > 0:
98
+ return
99
+ import cityscapesscripts.evaluation.evalInstanceLevelSemanticLabeling as cityscapes_eval
100
+
101
+ self._logger.info("Evaluating results under {} ...".format(self._temp_dir))
102
+
103
+ # set some global states in cityscapes evaluation API, before evaluating
104
+ cityscapes_eval.args.predictionPath = os.path.abspath(self._temp_dir)
105
+ cityscapes_eval.args.predictionWalk = None
106
+ cityscapes_eval.args.JSONOutput = False
107
+ cityscapes_eval.args.colorized = False
108
+ cityscapes_eval.args.gtInstancesFile = os.path.join(self._temp_dir, "gtInstances.json")
109
+
110
+ # These lines are adopted from
111
+ # https://github.com/mcordts/cityscapesScripts/blob/master/cityscapesscripts/evaluation/evalInstanceLevelSemanticLabeling.py # noqa
112
+ gt_dir = PathManager.get_local_path(self._metadata.gt_dir)
113
+ groundTruthImgList = glob.glob(os.path.join(gt_dir, "*", "*_gtFine_instanceIds.png"))
114
+ assert len(
115
+ groundTruthImgList
116
+ ), "Cannot find any ground truth images to use for evaluation. Searched for: {}".format(
117
+ cityscapes_eval.args.groundTruthSearch
118
+ )
119
+ predictionImgList = []
120
+ for gt in groundTruthImgList:
121
+ predictionImgList.append(cityscapes_eval.getPrediction(gt, cityscapes_eval.args))
122
+ results = cityscapes_eval.evaluateImgLists(
123
+ predictionImgList, groundTruthImgList, cityscapes_eval.args
124
+ )["averages"]
125
+
126
+ ret = OrderedDict()
127
+ ret["segm"] = {"AP": results["allAp"] * 100, "AP50": results["allAp50%"] * 100}
128
+ self._working_dir.cleanup()
129
+ return ret
130
+
131
+
132
+ class CityscapesSemSegEvaluator(CityscapesEvaluator):
133
+ """
134
+ Evaluate semantic segmentation results on cityscapes dataset using cityscapes API.
135
+
136
+ Note:
137
+ * It does not work in multi-machine distributed training.
138
+ * It contains a synchronization, therefore has to be used on all ranks.
139
+ * Only the main process runs evaluation.
140
+ """
141
+
142
+ def process(self, inputs, outputs):
143
+ from cityscapesscripts.helpers.labels import trainId2label
144
+
145
+ for input, output in zip(inputs, outputs):
146
+ file_name = input["file_name"]
147
+ basename = os.path.splitext(os.path.basename(file_name))[0]
148
+ pred_filename = os.path.join(self._temp_dir, basename + "_pred.png")
149
+
150
+ output = output["sem_seg"].argmax(dim=0).to(self._cpu_device).numpy()
151
+ pred = 255 * np.ones(output.shape, dtype=np.uint8)
152
+ for train_id, label in trainId2label.items():
153
+ if label.ignoreInEval:
154
+ continue
155
+ pred[output == train_id] = label.id
156
+ Image.fromarray(pred).save(pred_filename)
157
+
158
+ def evaluate(self):
159
+ comm.synchronize()
160
+ if comm.get_rank() > 0:
161
+ return
162
+ # Load the Cityscapes eval script *after* setting the required env var,
163
+ # since the script reads CITYSCAPES_DATASET into global variables at load time.
164
+ import cityscapesscripts.evaluation.evalPixelLevelSemanticLabeling as cityscapes_eval
165
+
166
+ self._logger.info("Evaluating results under {} ...".format(self._temp_dir))
167
+
168
+ # set some global states in cityscapes evaluation API, before evaluating
169
+ cityscapes_eval.args.predictionPath = os.path.abspath(self._temp_dir)
170
+ cityscapes_eval.args.predictionWalk = None
171
+ cityscapes_eval.args.JSONOutput = False
172
+ cityscapes_eval.args.colorized = False
173
+
174
+ # These lines are adopted from
175
+ # https://github.com/mcordts/cityscapesScripts/blob/master/cityscapesscripts/evaluation/evalPixelLevelSemanticLabeling.py # noqa
176
+ gt_dir = PathManager.get_local_path(self._metadata.gt_dir)
177
+ groundTruthImgList = glob.glob(os.path.join(gt_dir, "*", "*_gtFine_labelIds.png"))
178
+ assert len(
179
+ groundTruthImgList
180
+ ), "Cannot find any ground truth images to use for evaluation. Searched for: {}".format(
181
+ cityscapes_eval.args.groundTruthSearch
182
+ )
183
+ predictionImgList = []
184
+ for gt in groundTruthImgList:
185
+ predictionImgList.append(cityscapes_eval.getPrediction(cityscapes_eval.args, gt))
186
+ results = cityscapes_eval.evaluateImgLists(
187
+ predictionImgList, groundTruthImgList, cityscapes_eval.args
188
+ )
189
+ ret = OrderedDict()
190
+ ret["sem_seg"] = {
191
+ "IoU": 100.0 * results["averageScoreClasses"],
192
+ "iIoU": 100.0 * results["averageScoreInstClasses"],
193
+ "IoU_sup": 100.0 * results["averageScoreCategories"],
194
+ "iIoU_sup": 100.0 * results["averageScoreInstCategories"],
195
+ }
196
+ self._working_dir.cleanup()
197
+ return ret
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/coco_evaluation.py ADDED
@@ -0,0 +1,722 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import contextlib
3
+ import copy
4
+ import io
5
+ import itertools
6
+ import json
7
+ import logging
8
+ import numpy as np
9
+ import os
10
+ import pickle
11
+ from collections import OrderedDict
12
+ import pycocotools.mask as mask_util
13
+ import torch
14
+ from pycocotools.coco import COCO
15
+ from pycocotools.cocoeval import COCOeval
16
+ from tabulate import tabulate
17
+
18
+ import annotator.oneformer.detectron2.utils.comm as comm
19
+ from annotator.oneformer.detectron2.config import CfgNode
20
+ from annotator.oneformer.detectron2.data import MetadataCatalog
21
+ from annotator.oneformer.detectron2.data.datasets.coco import convert_to_coco_json
22
+ from annotator.oneformer.detectron2.structures import Boxes, BoxMode, pairwise_iou
23
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
24
+ from annotator.oneformer.detectron2.utils.logger import create_small_table
25
+
26
+ from .evaluator import DatasetEvaluator
27
+
28
+ try:
29
+ from annotator.oneformer.detectron2.evaluation.fast_eval_api import COCOeval_opt
30
+ except ImportError:
31
+ COCOeval_opt = COCOeval
32
+
33
+
34
+ class COCOEvaluator(DatasetEvaluator):
35
+ """
36
+ Evaluate AR for object proposals, AP for instance detection/segmentation, AP
37
+ for keypoint detection outputs using COCO's metrics.
38
+ See http://cocodataset.org/#detection-eval and
39
+ http://cocodataset.org/#keypoints-eval to understand its metrics.
40
+ The metrics range from 0 to 100 (instead of 0 to 1), where a -1 or NaN means
41
+ the metric cannot be computed (e.g. due to no predictions made).
42
+
43
+ In addition to COCO, this evaluator is able to support any bounding box detection,
44
+ instance segmentation, or keypoint detection dataset.
45
+ """
46
+
47
+ def __init__(
48
+ self,
49
+ dataset_name,
50
+ tasks=None,
51
+ distributed=True,
52
+ output_dir=None,
53
+ *,
54
+ max_dets_per_image=None,
55
+ use_fast_impl=True,
56
+ kpt_oks_sigmas=(),
57
+ allow_cached_coco=True,
58
+ ):
59
+ """
60
+ Args:
61
+ dataset_name (str): name of the dataset to be evaluated.
62
+ It must have either the following corresponding metadata:
63
+
64
+ "json_file": the path to the COCO format annotation
65
+
66
+ Or it must be in detectron2's standard dataset format
67
+ so it can be converted to COCO format automatically.
68
+ tasks (tuple[str]): tasks that can be evaluated under the given
69
+ configuration. A task is one of "bbox", "segm", "keypoints".
70
+ By default, will infer this automatically from predictions.
71
+ distributed (True): if True, will collect results from all ranks and run evaluation
72
+ in the main process.
73
+ Otherwise, will only evaluate the results in the current process.
74
+ output_dir (str): optional, an output directory to dump all
75
+ results predicted on the dataset. The dump contains two files:
76
+
77
+ 1. "instances_predictions.pth" a file that can be loaded with `torch.load` and
78
+ contains all the results in the format they are produced by the model.
79
+ 2. "coco_instances_results.json" a json file in COCO's result format.
80
+ max_dets_per_image (int): limit on the maximum number of detections per image.
81
+ By default in COCO, this limit is to 100, but this can be customized
82
+ to be greater, as is needed in evaluation metrics AP fixed and AP pool
83
+ (see https://arxiv.org/pdf/2102.01066.pdf)
84
+ This doesn't affect keypoint evaluation.
85
+ use_fast_impl (bool): use a fast but **unofficial** implementation to compute AP.
86
+ Although the results should be very close to the official implementation in COCO
87
+ API, it is still recommended to compute results with the official API for use in
88
+ papers. The faster implementation also uses more RAM.
89
+ kpt_oks_sigmas (list[float]): The sigmas used to calculate keypoint OKS.
90
+ See http://cocodataset.org/#keypoints-eval
91
+ When empty, it will use the defaults in COCO.
92
+ Otherwise it should be the same length as ROI_KEYPOINT_HEAD.NUM_KEYPOINTS.
93
+ allow_cached_coco (bool): Whether to use cached coco json from previous validation
94
+ runs. You should set this to False if you need to use different validation data.
95
+ Defaults to True.
96
+ """
97
+ self._logger = logging.getLogger(__name__)
98
+ self._distributed = distributed
99
+ self._output_dir = output_dir
100
+
101
+ if use_fast_impl and (COCOeval_opt is COCOeval):
102
+ self._logger.info("Fast COCO eval is not built. Falling back to official COCO eval.")
103
+ use_fast_impl = False
104
+ self._use_fast_impl = use_fast_impl
105
+
106
+ # COCOeval requires the limit on the number of detections per image (maxDets) to be a list
107
+ # with at least 3 elements. The default maxDets in COCOeval is [1, 10, 100], in which the
108
+ # 3rd element (100) is used as the limit on the number of detections per image when
109
+ # evaluating AP. COCOEvaluator expects an integer for max_dets_per_image, so for COCOeval,
110
+ # we reformat max_dets_per_image into [1, 10, max_dets_per_image], based on the defaults.
111
+ if max_dets_per_image is None:
112
+ max_dets_per_image = [1, 10, 100]
113
+ else:
114
+ max_dets_per_image = [1, 10, max_dets_per_image]
115
+ self._max_dets_per_image = max_dets_per_image
116
+
117
+ if tasks is not None and isinstance(tasks, CfgNode):
118
+ kpt_oks_sigmas = (
119
+ tasks.TEST.KEYPOINT_OKS_SIGMAS if not kpt_oks_sigmas else kpt_oks_sigmas
120
+ )
121
+ self._logger.warn(
122
+ "COCO Evaluator instantiated using config, this is deprecated behavior."
123
+ " Please pass in explicit arguments instead."
124
+ )
125
+ self._tasks = None # Infering it from predictions should be better
126
+ else:
127
+ self._tasks = tasks
128
+
129
+ self._cpu_device = torch.device("cpu")
130
+
131
+ self._metadata = MetadataCatalog.get(dataset_name)
132
+ if not hasattr(self._metadata, "json_file"):
133
+ if output_dir is None:
134
+ raise ValueError(
135
+ "output_dir must be provided to COCOEvaluator "
136
+ "for datasets not in COCO format."
137
+ )
138
+ self._logger.info(f"Trying to convert '{dataset_name}' to COCO format ...")
139
+
140
+ cache_path = os.path.join(output_dir, f"{dataset_name}_coco_format.json")
141
+ self._metadata.json_file = cache_path
142
+ convert_to_coco_json(dataset_name, cache_path, allow_cached=allow_cached_coco)
143
+
144
+ json_file = PathManager.get_local_path(self._metadata.json_file)
145
+ with contextlib.redirect_stdout(io.StringIO()):
146
+ self._coco_api = COCO(json_file)
147
+
148
+ # Test set json files do not contain annotations (evaluation must be
149
+ # performed using the COCO evaluation server).
150
+ self._do_evaluation = "annotations" in self._coco_api.dataset
151
+ if self._do_evaluation:
152
+ self._kpt_oks_sigmas = kpt_oks_sigmas
153
+
154
+ def reset(self):
155
+ self._predictions = []
156
+
157
+ def process(self, inputs, outputs):
158
+ """
159
+ Args:
160
+ inputs: the inputs to a COCO model (e.g., GeneralizedRCNN).
161
+ It is a list of dict. Each dict corresponds to an image and
162
+ contains keys like "height", "width", "file_name", "image_id".
163
+ outputs: the outputs of a COCO model. It is a list of dicts with key
164
+ "instances" that contains :class:`Instances`.
165
+ """
166
+ for input, output in zip(inputs, outputs):
167
+ prediction = {"image_id": input["image_id"]}
168
+
169
+ if "instances" in output:
170
+ instances = output["instances"].to(self._cpu_device)
171
+ prediction["instances"] = instances_to_coco_json(instances, input["image_id"])
172
+ if "proposals" in output:
173
+ prediction["proposals"] = output["proposals"].to(self._cpu_device)
174
+ if len(prediction) > 1:
175
+ self._predictions.append(prediction)
176
+
177
+ def evaluate(self, img_ids=None):
178
+ """
179
+ Args:
180
+ img_ids: a list of image IDs to evaluate on. Default to None for the whole dataset
181
+ """
182
+ if self._distributed:
183
+ comm.synchronize()
184
+ predictions = comm.gather(self._predictions, dst=0)
185
+ predictions = list(itertools.chain(*predictions))
186
+
187
+ if not comm.is_main_process():
188
+ return {}
189
+ else:
190
+ predictions = self._predictions
191
+
192
+ if len(predictions) == 0:
193
+ self._logger.warning("[COCOEvaluator] Did not receive valid predictions.")
194
+ return {}
195
+
196
+ if self._output_dir:
197
+ PathManager.mkdirs(self._output_dir)
198
+ file_path = os.path.join(self._output_dir, "instances_predictions.pth")
199
+ with PathManager.open(file_path, "wb") as f:
200
+ torch.save(predictions, f)
201
+
202
+ self._results = OrderedDict()
203
+ if "proposals" in predictions[0]:
204
+ self._eval_box_proposals(predictions)
205
+ if "instances" in predictions[0]:
206
+ self._eval_predictions(predictions, img_ids=img_ids)
207
+ # Copy so the caller can do whatever with results
208
+ return copy.deepcopy(self._results)
209
+
210
+ def _tasks_from_predictions(self, predictions):
211
+ """
212
+ Get COCO API "tasks" (i.e. iou_type) from COCO-format predictions.
213
+ """
214
+ tasks = {"bbox"}
215
+ for pred in predictions:
216
+ if "segmentation" in pred:
217
+ tasks.add("segm")
218
+ if "keypoints" in pred:
219
+ tasks.add("keypoints")
220
+ return sorted(tasks)
221
+
222
+ def _eval_predictions(self, predictions, img_ids=None):
223
+ """
224
+ Evaluate predictions. Fill self._results with the metrics of the tasks.
225
+ """
226
+ self._logger.info("Preparing results for COCO format ...")
227
+ coco_results = list(itertools.chain(*[x["instances"] for x in predictions]))
228
+ tasks = self._tasks or self._tasks_from_predictions(coco_results)
229
+
230
+ # unmap the category ids for COCO
231
+ if hasattr(self._metadata, "thing_dataset_id_to_contiguous_id"):
232
+ dataset_id_to_contiguous_id = self._metadata.thing_dataset_id_to_contiguous_id
233
+ all_contiguous_ids = list(dataset_id_to_contiguous_id.values())
234
+ num_classes = len(all_contiguous_ids)
235
+ assert min(all_contiguous_ids) == 0 and max(all_contiguous_ids) == num_classes - 1
236
+
237
+ reverse_id_mapping = {v: k for k, v in dataset_id_to_contiguous_id.items()}
238
+ for result in coco_results:
239
+ category_id = result["category_id"]
240
+ assert category_id < num_classes, (
241
+ f"A prediction has class={category_id}, "
242
+ f"but the dataset only has {num_classes} classes and "
243
+ f"predicted class id should be in [0, {num_classes - 1}]."
244
+ )
245
+ result["category_id"] = reverse_id_mapping[category_id]
246
+
247
+ if self._output_dir:
248
+ file_path = os.path.join(self._output_dir, "coco_instances_results.json")
249
+ self._logger.info("Saving results to {}".format(file_path))
250
+ with PathManager.open(file_path, "w") as f:
251
+ f.write(json.dumps(coco_results))
252
+ f.flush()
253
+
254
+ if not self._do_evaluation:
255
+ self._logger.info("Annotations are not available for evaluation.")
256
+ return
257
+
258
+ self._logger.info(
259
+ "Evaluating predictions with {} COCO API...".format(
260
+ "unofficial" if self._use_fast_impl else "official"
261
+ )
262
+ )
263
+ for task in sorted(tasks):
264
+ assert task in {"bbox", "segm", "keypoints"}, f"Got unknown task: {task}!"
265
+ coco_eval = (
266
+ _evaluate_predictions_on_coco(
267
+ self._coco_api,
268
+ coco_results,
269
+ task,
270
+ kpt_oks_sigmas=self._kpt_oks_sigmas,
271
+ cocoeval_fn=COCOeval_opt if self._use_fast_impl else COCOeval,
272
+ img_ids=img_ids,
273
+ max_dets_per_image=self._max_dets_per_image,
274
+ )
275
+ if len(coco_results) > 0
276
+ else None # cocoapi does not handle empty results very well
277
+ )
278
+
279
+ res = self._derive_coco_results(
280
+ coco_eval, task, class_names=self._metadata.get("thing_classes")
281
+ )
282
+ self._results[task] = res
283
+
284
+ def _eval_box_proposals(self, predictions):
285
+ """
286
+ Evaluate the box proposals in predictions.
287
+ Fill self._results with the metrics for "box_proposals" task.
288
+ """
289
+ if self._output_dir:
290
+ # Saving generated box proposals to file.
291
+ # Predicted box_proposals are in XYXY_ABS mode.
292
+ bbox_mode = BoxMode.XYXY_ABS.value
293
+ ids, boxes, objectness_logits = [], [], []
294
+ for prediction in predictions:
295
+ ids.append(prediction["image_id"])
296
+ boxes.append(prediction["proposals"].proposal_boxes.tensor.numpy())
297
+ objectness_logits.append(prediction["proposals"].objectness_logits.numpy())
298
+
299
+ proposal_data = {
300
+ "boxes": boxes,
301
+ "objectness_logits": objectness_logits,
302
+ "ids": ids,
303
+ "bbox_mode": bbox_mode,
304
+ }
305
+ with PathManager.open(os.path.join(self._output_dir, "box_proposals.pkl"), "wb") as f:
306
+ pickle.dump(proposal_data, f)
307
+
308
+ if not self._do_evaluation:
309
+ self._logger.info("Annotations are not available for evaluation.")
310
+ return
311
+
312
+ self._logger.info("Evaluating bbox proposals ...")
313
+ res = {}
314
+ areas = {"all": "", "small": "s", "medium": "m", "large": "l"}
315
+ for limit in [100, 1000]:
316
+ for area, suffix in areas.items():
317
+ stats = _evaluate_box_proposals(predictions, self._coco_api, area=area, limit=limit)
318
+ key = "AR{}@{:d}".format(suffix, limit)
319
+ res[key] = float(stats["ar"].item() * 100)
320
+ self._logger.info("Proposal metrics: \n" + create_small_table(res))
321
+ self._results["box_proposals"] = res
322
+
323
+ def _derive_coco_results(self, coco_eval, iou_type, class_names=None):
324
+ """
325
+ Derive the desired score numbers from summarized COCOeval.
326
+
327
+ Args:
328
+ coco_eval (None or COCOEval): None represents no predictions from model.
329
+ iou_type (str):
330
+ class_names (None or list[str]): if provided, will use it to predict
331
+ per-category AP.
332
+
333
+ Returns:
334
+ a dict of {metric name: score}
335
+ """
336
+
337
+ metrics = {
338
+ "bbox": ["AP", "AP50", "AP75", "APs", "APm", "APl"],
339
+ "segm": ["AP", "AP50", "AP75", "APs", "APm", "APl"],
340
+ "keypoints": ["AP", "AP50", "AP75", "APm", "APl"],
341
+ }[iou_type]
342
+
343
+ if coco_eval is None:
344
+ self._logger.warn("No predictions from the model!")
345
+ return {metric: float("nan") for metric in metrics}
346
+
347
+ # the standard metrics
348
+ results = {
349
+ metric: float(coco_eval.stats[idx] * 100 if coco_eval.stats[idx] >= 0 else "nan")
350
+ for idx, metric in enumerate(metrics)
351
+ }
352
+ self._logger.info(
353
+ "Evaluation results for {}: \n".format(iou_type) + create_small_table(results)
354
+ )
355
+ if not np.isfinite(sum(results.values())):
356
+ self._logger.info("Some metrics cannot be computed and is shown as NaN.")
357
+
358
+ if class_names is None or len(class_names) <= 1:
359
+ return results
360
+ # Compute per-category AP
361
+ # from https://github.com/facebookresearch/Detectron/blob/a6a835f5b8208c45d0dce217ce9bbda915f44df7/detectron/datasets/json_dataset_evaluator.py#L222-L252 # noqa
362
+ precisions = coco_eval.eval["precision"]
363
+ # precision has dims (iou, recall, cls, area range, max dets)
364
+ assert len(class_names) == precisions.shape[2]
365
+
366
+ results_per_category = []
367
+ for idx, name in enumerate(class_names):
368
+ # area range index 0: all area ranges
369
+ # max dets index -1: typically 100 per image
370
+ precision = precisions[:, :, idx, 0, -1]
371
+ precision = precision[precision > -1]
372
+ ap = np.mean(precision) if precision.size else float("nan")
373
+ results_per_category.append(("{}".format(name), float(ap * 100)))
374
+
375
+ # tabulate it
376
+ N_COLS = min(6, len(results_per_category) * 2)
377
+ results_flatten = list(itertools.chain(*results_per_category))
378
+ results_2d = itertools.zip_longest(*[results_flatten[i::N_COLS] for i in range(N_COLS)])
379
+ table = tabulate(
380
+ results_2d,
381
+ tablefmt="pipe",
382
+ floatfmt=".3f",
383
+ headers=["category", "AP"] * (N_COLS // 2),
384
+ numalign="left",
385
+ )
386
+ self._logger.info("Per-category {} AP: \n".format(iou_type) + table)
387
+
388
+ results.update({"AP-" + name: ap for name, ap in results_per_category})
389
+ return results
390
+
391
+
392
+ def instances_to_coco_json(instances, img_id):
393
+ """
394
+ Dump an "Instances" object to a COCO-format json that's used for evaluation.
395
+
396
+ Args:
397
+ instances (Instances):
398
+ img_id (int): the image id
399
+
400
+ Returns:
401
+ list[dict]: list of json annotations in COCO format.
402
+ """
403
+ num_instance = len(instances)
404
+ if num_instance == 0:
405
+ return []
406
+
407
+ boxes = instances.pred_boxes.tensor.numpy()
408
+ boxes = BoxMode.convert(boxes, BoxMode.XYXY_ABS, BoxMode.XYWH_ABS)
409
+ boxes = boxes.tolist()
410
+ scores = instances.scores.tolist()
411
+ classes = instances.pred_classes.tolist()
412
+
413
+ has_mask = instances.has("pred_masks")
414
+ if has_mask:
415
+ # use RLE to encode the masks, because they are too large and takes memory
416
+ # since this evaluator stores outputs of the entire dataset
417
+ rles = [
418
+ mask_util.encode(np.array(mask[:, :, None], order="F", dtype="uint8"))[0]
419
+ for mask in instances.pred_masks
420
+ ]
421
+ for rle in rles:
422
+ # "counts" is an array encoded by mask_util as a byte-stream. Python3's
423
+ # json writer which always produces strings cannot serialize a bytestream
424
+ # unless you decode it. Thankfully, utf-8 works out (which is also what
425
+ # the pycocotools/_mask.pyx does).
426
+ rle["counts"] = rle["counts"].decode("utf-8")
427
+
428
+ has_keypoints = instances.has("pred_keypoints")
429
+ if has_keypoints:
430
+ keypoints = instances.pred_keypoints
431
+
432
+ results = []
433
+ for k in range(num_instance):
434
+ result = {
435
+ "image_id": img_id,
436
+ "category_id": classes[k],
437
+ "bbox": boxes[k],
438
+ "score": scores[k],
439
+ }
440
+ if has_mask:
441
+ result["segmentation"] = rles[k]
442
+ if has_keypoints:
443
+ # In COCO annotations,
444
+ # keypoints coordinates are pixel indices.
445
+ # However our predictions are floating point coordinates.
446
+ # Therefore we subtract 0.5 to be consistent with the annotation format.
447
+ # This is the inverse of data loading logic in `datasets/coco.py`.
448
+ keypoints[k][:, :2] -= 0.5
449
+ result["keypoints"] = keypoints[k].flatten().tolist()
450
+ results.append(result)
451
+ return results
452
+
453
+
454
+ # inspired from Detectron:
455
+ # https://github.com/facebookresearch/Detectron/blob/a6a835f5b8208c45d0dce217ce9bbda915f44df7/detectron/datasets/json_dataset_evaluator.py#L255 # noqa
456
+ def _evaluate_box_proposals(dataset_predictions, coco_api, thresholds=None, area="all", limit=None):
457
+ """
458
+ Evaluate detection proposal recall metrics. This function is a much
459
+ faster alternative to the official COCO API recall evaluation code. However,
460
+ it produces slightly different results.
461
+ """
462
+ # Record max overlap value for each gt box
463
+ # Return vector of overlap values
464
+ areas = {
465
+ "all": 0,
466
+ "small": 1,
467
+ "medium": 2,
468
+ "large": 3,
469
+ "96-128": 4,
470
+ "128-256": 5,
471
+ "256-512": 6,
472
+ "512-inf": 7,
473
+ }
474
+ area_ranges = [
475
+ [0**2, 1e5**2], # all
476
+ [0**2, 32**2], # small
477
+ [32**2, 96**2], # medium
478
+ [96**2, 1e5**2], # large
479
+ [96**2, 128**2], # 96-128
480
+ [128**2, 256**2], # 128-256
481
+ [256**2, 512**2], # 256-512
482
+ [512**2, 1e5**2],
483
+ ] # 512-inf
484
+ assert area in areas, "Unknown area range: {}".format(area)
485
+ area_range = area_ranges[areas[area]]
486
+ gt_overlaps = []
487
+ num_pos = 0
488
+
489
+ for prediction_dict in dataset_predictions:
490
+ predictions = prediction_dict["proposals"]
491
+
492
+ # sort predictions in descending order
493
+ # TODO maybe remove this and make it explicit in the documentation
494
+ inds = predictions.objectness_logits.sort(descending=True)[1]
495
+ predictions = predictions[inds]
496
+
497
+ ann_ids = coco_api.getAnnIds(imgIds=prediction_dict["image_id"])
498
+ anno = coco_api.loadAnns(ann_ids)
499
+ gt_boxes = [
500
+ BoxMode.convert(obj["bbox"], BoxMode.XYWH_ABS, BoxMode.XYXY_ABS)
501
+ for obj in anno
502
+ if obj["iscrowd"] == 0
503
+ ]
504
+ gt_boxes = torch.as_tensor(gt_boxes).reshape(-1, 4) # guard against no boxes
505
+ gt_boxes = Boxes(gt_boxes)
506
+ gt_areas = torch.as_tensor([obj["area"] for obj in anno if obj["iscrowd"] == 0])
507
+
508
+ if len(gt_boxes) == 0 or len(predictions) == 0:
509
+ continue
510
+
511
+ valid_gt_inds = (gt_areas >= area_range[0]) & (gt_areas <= area_range[1])
512
+ gt_boxes = gt_boxes[valid_gt_inds]
513
+
514
+ num_pos += len(gt_boxes)
515
+
516
+ if len(gt_boxes) == 0:
517
+ continue
518
+
519
+ if limit is not None and len(predictions) > limit:
520
+ predictions = predictions[:limit]
521
+
522
+ overlaps = pairwise_iou(predictions.proposal_boxes, gt_boxes)
523
+
524
+ _gt_overlaps = torch.zeros(len(gt_boxes))
525
+ for j in range(min(len(predictions), len(gt_boxes))):
526
+ # find which proposal box maximally covers each gt box
527
+ # and get the iou amount of coverage for each gt box
528
+ max_overlaps, argmax_overlaps = overlaps.max(dim=0)
529
+
530
+ # find which gt box is 'best' covered (i.e. 'best' = most iou)
531
+ gt_ovr, gt_ind = max_overlaps.max(dim=0)
532
+ assert gt_ovr >= 0
533
+ # find the proposal box that covers the best covered gt box
534
+ box_ind = argmax_overlaps[gt_ind]
535
+ # record the iou coverage of this gt box
536
+ _gt_overlaps[j] = overlaps[box_ind, gt_ind]
537
+ assert _gt_overlaps[j] == gt_ovr
538
+ # mark the proposal box and the gt box as used
539
+ overlaps[box_ind, :] = -1
540
+ overlaps[:, gt_ind] = -1
541
+
542
+ # append recorded iou coverage level
543
+ gt_overlaps.append(_gt_overlaps)
544
+ gt_overlaps = (
545
+ torch.cat(gt_overlaps, dim=0) if len(gt_overlaps) else torch.zeros(0, dtype=torch.float32)
546
+ )
547
+ gt_overlaps, _ = torch.sort(gt_overlaps)
548
+
549
+ if thresholds is None:
550
+ step = 0.05
551
+ thresholds = torch.arange(0.5, 0.95 + 1e-5, step, dtype=torch.float32)
552
+ recalls = torch.zeros_like(thresholds)
553
+ # compute recall for each iou threshold
554
+ for i, t in enumerate(thresholds):
555
+ recalls[i] = (gt_overlaps >= t).float().sum() / float(num_pos)
556
+ # ar = 2 * np.trapz(recalls, thresholds)
557
+ ar = recalls.mean()
558
+ return {
559
+ "ar": ar,
560
+ "recalls": recalls,
561
+ "thresholds": thresholds,
562
+ "gt_overlaps": gt_overlaps,
563
+ "num_pos": num_pos,
564
+ }
565
+
566
+
567
+ def _evaluate_predictions_on_coco(
568
+ coco_gt,
569
+ coco_results,
570
+ iou_type,
571
+ kpt_oks_sigmas=None,
572
+ cocoeval_fn=COCOeval_opt,
573
+ img_ids=None,
574
+ max_dets_per_image=None,
575
+ ):
576
+ """
577
+ Evaluate the coco results using COCOEval API.
578
+ """
579
+ assert len(coco_results) > 0
580
+
581
+ if iou_type == "segm":
582
+ coco_results = copy.deepcopy(coco_results)
583
+ # When evaluating mask AP, if the results contain bbox, cocoapi will
584
+ # use the box area as the area of the instance, instead of the mask area.
585
+ # This leads to a different definition of small/medium/large.
586
+ # We remove the bbox field to let mask AP use mask area.
587
+ for c in coco_results:
588
+ c.pop("bbox", None)
589
+
590
+ coco_dt = coco_gt.loadRes(coco_results)
591
+ coco_eval = cocoeval_fn(coco_gt, coco_dt, iou_type)
592
+ # For COCO, the default max_dets_per_image is [1, 10, 100].
593
+ if max_dets_per_image is None:
594
+ max_dets_per_image = [1, 10, 100] # Default from COCOEval
595
+ else:
596
+ assert (
597
+ len(max_dets_per_image) >= 3
598
+ ), "COCOeval requires maxDets (and max_dets_per_image) to have length at least 3"
599
+ # In the case that user supplies a custom input for max_dets_per_image,
600
+ # apply COCOevalMaxDets to evaluate AP with the custom input.
601
+ if max_dets_per_image[2] != 100:
602
+ coco_eval = COCOevalMaxDets(coco_gt, coco_dt, iou_type)
603
+ if iou_type != "keypoints":
604
+ coco_eval.params.maxDets = max_dets_per_image
605
+
606
+ if img_ids is not None:
607
+ coco_eval.params.imgIds = img_ids
608
+
609
+ if iou_type == "keypoints":
610
+ # Use the COCO default keypoint OKS sigmas unless overrides are specified
611
+ if kpt_oks_sigmas:
612
+ assert hasattr(coco_eval.params, "kpt_oks_sigmas"), "pycocotools is too old!"
613
+ coco_eval.params.kpt_oks_sigmas = np.array(kpt_oks_sigmas)
614
+ # COCOAPI requires every detection and every gt to have keypoints, so
615
+ # we just take the first entry from both
616
+ num_keypoints_dt = len(coco_results[0]["keypoints"]) // 3
617
+ num_keypoints_gt = len(next(iter(coco_gt.anns.values()))["keypoints"]) // 3
618
+ num_keypoints_oks = len(coco_eval.params.kpt_oks_sigmas)
619
+ assert num_keypoints_oks == num_keypoints_dt == num_keypoints_gt, (
620
+ f"[COCOEvaluator] Prediction contain {num_keypoints_dt} keypoints. "
621
+ f"Ground truth contains {num_keypoints_gt} keypoints. "
622
+ f"The length of cfg.TEST.KEYPOINT_OKS_SIGMAS is {num_keypoints_oks}. "
623
+ "They have to agree with each other. For meaning of OKS, please refer to "
624
+ "http://cocodataset.org/#keypoints-eval."
625
+ )
626
+
627
+ coco_eval.evaluate()
628
+ coco_eval.accumulate()
629
+ coco_eval.summarize()
630
+
631
+ return coco_eval
632
+
633
+
634
+ class COCOevalMaxDets(COCOeval):
635
+ """
636
+ Modified version of COCOeval for evaluating AP with a custom
637
+ maxDets (by default for COCO, maxDets is 100)
638
+ """
639
+
640
+ def summarize(self):
641
+ """
642
+ Compute and display summary metrics for evaluation results given
643
+ a custom value for max_dets_per_image
644
+ """
645
+
646
+ def _summarize(ap=1, iouThr=None, areaRng="all", maxDets=100):
647
+ p = self.params
648
+ iStr = " {:<18} {} @[ IoU={:<9} | area={:>6s} | maxDets={:>3d} ] = {:0.3f}"
649
+ titleStr = "Average Precision" if ap == 1 else "Average Recall"
650
+ typeStr = "(AP)" if ap == 1 else "(AR)"
651
+ iouStr = (
652
+ "{:0.2f}:{:0.2f}".format(p.iouThrs[0], p.iouThrs[-1])
653
+ if iouThr is None
654
+ else "{:0.2f}".format(iouThr)
655
+ )
656
+
657
+ aind = [i for i, aRng in enumerate(p.areaRngLbl) if aRng == areaRng]
658
+ mind = [i for i, mDet in enumerate(p.maxDets) if mDet == maxDets]
659
+ if ap == 1:
660
+ # dimension of precision: [TxRxKxAxM]
661
+ s = self.eval["precision"]
662
+ # IoU
663
+ if iouThr is not None:
664
+ t = np.where(iouThr == p.iouThrs)[0]
665
+ s = s[t]
666
+ s = s[:, :, :, aind, mind]
667
+ else:
668
+ # dimension of recall: [TxKxAxM]
669
+ s = self.eval["recall"]
670
+ if iouThr is not None:
671
+ t = np.where(iouThr == p.iouThrs)[0]
672
+ s = s[t]
673
+ s = s[:, :, aind, mind]
674
+ if len(s[s > -1]) == 0:
675
+ mean_s = -1
676
+ else:
677
+ mean_s = np.mean(s[s > -1])
678
+ print(iStr.format(titleStr, typeStr, iouStr, areaRng, maxDets, mean_s))
679
+ return mean_s
680
+
681
+ def _summarizeDets():
682
+ stats = np.zeros((12,))
683
+ # Evaluate AP using the custom limit on maximum detections per image
684
+ stats[0] = _summarize(1, maxDets=self.params.maxDets[2])
685
+ stats[1] = _summarize(1, iouThr=0.5, maxDets=self.params.maxDets[2])
686
+ stats[2] = _summarize(1, iouThr=0.75, maxDets=self.params.maxDets[2])
687
+ stats[3] = _summarize(1, areaRng="small", maxDets=self.params.maxDets[2])
688
+ stats[4] = _summarize(1, areaRng="medium", maxDets=self.params.maxDets[2])
689
+ stats[5] = _summarize(1, areaRng="large", maxDets=self.params.maxDets[2])
690
+ stats[6] = _summarize(0, maxDets=self.params.maxDets[0])
691
+ stats[7] = _summarize(0, maxDets=self.params.maxDets[1])
692
+ stats[8] = _summarize(0, maxDets=self.params.maxDets[2])
693
+ stats[9] = _summarize(0, areaRng="small", maxDets=self.params.maxDets[2])
694
+ stats[10] = _summarize(0, areaRng="medium", maxDets=self.params.maxDets[2])
695
+ stats[11] = _summarize(0, areaRng="large", maxDets=self.params.maxDets[2])
696
+ return stats
697
+
698
+ def _summarizeKps():
699
+ stats = np.zeros((10,))
700
+ stats[0] = _summarize(1, maxDets=20)
701
+ stats[1] = _summarize(1, maxDets=20, iouThr=0.5)
702
+ stats[2] = _summarize(1, maxDets=20, iouThr=0.75)
703
+ stats[3] = _summarize(1, maxDets=20, areaRng="medium")
704
+ stats[4] = _summarize(1, maxDets=20, areaRng="large")
705
+ stats[5] = _summarize(0, maxDets=20)
706
+ stats[6] = _summarize(0, maxDets=20, iouThr=0.5)
707
+ stats[7] = _summarize(0, maxDets=20, iouThr=0.75)
708
+ stats[8] = _summarize(0, maxDets=20, areaRng="medium")
709
+ stats[9] = _summarize(0, maxDets=20, areaRng="large")
710
+ return stats
711
+
712
+ if not self.eval:
713
+ raise Exception("Please run accumulate() first")
714
+ iouType = self.params.iouType
715
+ if iouType == "segm" or iouType == "bbox":
716
+ summarize = _summarizeDets
717
+ elif iouType == "keypoints":
718
+ summarize = _summarizeKps
719
+ self.stats = summarize()
720
+
721
+ def __str__(self):
722
+ self.summarize()
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/evaluator.py ADDED
@@ -0,0 +1,224 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import datetime
3
+ import logging
4
+ import time
5
+ from collections import OrderedDict, abc
6
+ from contextlib import ExitStack, contextmanager
7
+ from typing import List, Union
8
+ import torch
9
+ from torch import nn
10
+
11
+ from annotator.oneformer.detectron2.utils.comm import get_world_size, is_main_process
12
+ from annotator.oneformer.detectron2.utils.logger import log_every_n_seconds
13
+
14
+
15
+ class DatasetEvaluator:
16
+ """
17
+ Base class for a dataset evaluator.
18
+
19
+ The function :func:`inference_on_dataset` runs the model over
20
+ all samples in the dataset, and have a DatasetEvaluator to process the inputs/outputs.
21
+
22
+ This class will accumulate information of the inputs/outputs (by :meth:`process`),
23
+ and produce evaluation results in the end (by :meth:`evaluate`).
24
+ """
25
+
26
+ def reset(self):
27
+ """
28
+ Preparation for a new round of evaluation.
29
+ Should be called before starting a round of evaluation.
30
+ """
31
+ pass
32
+
33
+ def process(self, inputs, outputs):
34
+ """
35
+ Process the pair of inputs and outputs.
36
+ If they contain batches, the pairs can be consumed one-by-one using `zip`:
37
+
38
+ .. code-block:: python
39
+
40
+ for input_, output in zip(inputs, outputs):
41
+ # do evaluation on single input/output pair
42
+ ...
43
+
44
+ Args:
45
+ inputs (list): the inputs that's used to call the model.
46
+ outputs (list): the return value of `model(inputs)`
47
+ """
48
+ pass
49
+
50
+ def evaluate(self):
51
+ """
52
+ Evaluate/summarize the performance, after processing all input/output pairs.
53
+
54
+ Returns:
55
+ dict:
56
+ A new evaluator class can return a dict of arbitrary format
57
+ as long as the user can process the results.
58
+ In our train_net.py, we expect the following format:
59
+
60
+ * key: the name of the task (e.g., bbox)
61
+ * value: a dict of {metric name: score}, e.g.: {"AP50": 80}
62
+ """
63
+ pass
64
+
65
+
66
+ class DatasetEvaluators(DatasetEvaluator):
67
+ """
68
+ Wrapper class to combine multiple :class:`DatasetEvaluator` instances.
69
+
70
+ This class dispatches every evaluation call to
71
+ all of its :class:`DatasetEvaluator`.
72
+ """
73
+
74
+ def __init__(self, evaluators):
75
+ """
76
+ Args:
77
+ evaluators (list): the evaluators to combine.
78
+ """
79
+ super().__init__()
80
+ self._evaluators = evaluators
81
+
82
+ def reset(self):
83
+ for evaluator in self._evaluators:
84
+ evaluator.reset()
85
+
86
+ def process(self, inputs, outputs):
87
+ for evaluator in self._evaluators:
88
+ evaluator.process(inputs, outputs)
89
+
90
+ def evaluate(self):
91
+ results = OrderedDict()
92
+ for evaluator in self._evaluators:
93
+ result = evaluator.evaluate()
94
+ if is_main_process() and result is not None:
95
+ for k, v in result.items():
96
+ assert (
97
+ k not in results
98
+ ), "Different evaluators produce results with the same key {}".format(k)
99
+ results[k] = v
100
+ return results
101
+
102
+
103
+ def inference_on_dataset(
104
+ model, data_loader, evaluator: Union[DatasetEvaluator, List[DatasetEvaluator], None]
105
+ ):
106
+ """
107
+ Run model on the data_loader and evaluate the metrics with evaluator.
108
+ Also benchmark the inference speed of `model.__call__` accurately.
109
+ The model will be used in eval mode.
110
+
111
+ Args:
112
+ model (callable): a callable which takes an object from
113
+ `data_loader` and returns some outputs.
114
+
115
+ If it's an nn.Module, it will be temporarily set to `eval` mode.
116
+ If you wish to evaluate a model in `training` mode instead, you can
117
+ wrap the given model and override its behavior of `.eval()` and `.train()`.
118
+ data_loader: an iterable object with a length.
119
+ The elements it generates will be the inputs to the model.
120
+ evaluator: the evaluator(s) to run. Use `None` if you only want to benchmark,
121
+ but don't want to do any evaluation.
122
+
123
+ Returns:
124
+ The return value of `evaluator.evaluate()`
125
+ """
126
+ num_devices = get_world_size()
127
+ logger = logging.getLogger(__name__)
128
+ logger.info("Start inference on {} batches".format(len(data_loader)))
129
+
130
+ total = len(data_loader) # inference data loader must have a fixed length
131
+ if evaluator is None:
132
+ # create a no-op evaluator
133
+ evaluator = DatasetEvaluators([])
134
+ if isinstance(evaluator, abc.MutableSequence):
135
+ evaluator = DatasetEvaluators(evaluator)
136
+ evaluator.reset()
137
+
138
+ num_warmup = min(5, total - 1)
139
+ start_time = time.perf_counter()
140
+ total_data_time = 0
141
+ total_compute_time = 0
142
+ total_eval_time = 0
143
+ with ExitStack() as stack:
144
+ if isinstance(model, nn.Module):
145
+ stack.enter_context(inference_context(model))
146
+ stack.enter_context(torch.no_grad())
147
+
148
+ start_data_time = time.perf_counter()
149
+ for idx, inputs in enumerate(data_loader):
150
+ total_data_time += time.perf_counter() - start_data_time
151
+ if idx == num_warmup:
152
+ start_time = time.perf_counter()
153
+ total_data_time = 0
154
+ total_compute_time = 0
155
+ total_eval_time = 0
156
+
157
+ start_compute_time = time.perf_counter()
158
+ outputs = model(inputs)
159
+ if torch.cuda.is_available():
160
+ torch.cuda.synchronize()
161
+ total_compute_time += time.perf_counter() - start_compute_time
162
+
163
+ start_eval_time = time.perf_counter()
164
+ evaluator.process(inputs, outputs)
165
+ total_eval_time += time.perf_counter() - start_eval_time
166
+
167
+ iters_after_start = idx + 1 - num_warmup * int(idx >= num_warmup)
168
+ data_seconds_per_iter = total_data_time / iters_after_start
169
+ compute_seconds_per_iter = total_compute_time / iters_after_start
170
+ eval_seconds_per_iter = total_eval_time / iters_after_start
171
+ total_seconds_per_iter = (time.perf_counter() - start_time) / iters_after_start
172
+ if idx >= num_warmup * 2 or compute_seconds_per_iter > 5:
173
+ eta = datetime.timedelta(seconds=int(total_seconds_per_iter * (total - idx - 1)))
174
+ log_every_n_seconds(
175
+ logging.INFO,
176
+ (
177
+ f"Inference done {idx + 1}/{total}. "
178
+ f"Dataloading: {data_seconds_per_iter:.4f} s/iter. "
179
+ f"Inference: {compute_seconds_per_iter:.4f} s/iter. "
180
+ f"Eval: {eval_seconds_per_iter:.4f} s/iter. "
181
+ f"Total: {total_seconds_per_iter:.4f} s/iter. "
182
+ f"ETA={eta}"
183
+ ),
184
+ n=5,
185
+ )
186
+ start_data_time = time.perf_counter()
187
+
188
+ # Measure the time only for this worker (before the synchronization barrier)
189
+ total_time = time.perf_counter() - start_time
190
+ total_time_str = str(datetime.timedelta(seconds=total_time))
191
+ # NOTE this format is parsed by grep
192
+ logger.info(
193
+ "Total inference time: {} ({:.6f} s / iter per device, on {} devices)".format(
194
+ total_time_str, total_time / (total - num_warmup), num_devices
195
+ )
196
+ )
197
+ total_compute_time_str = str(datetime.timedelta(seconds=int(total_compute_time)))
198
+ logger.info(
199
+ "Total inference pure compute time: {} ({:.6f} s / iter per device, on {} devices)".format(
200
+ total_compute_time_str, total_compute_time / (total - num_warmup), num_devices
201
+ )
202
+ )
203
+
204
+ results = evaluator.evaluate()
205
+ # An evaluator may return None when not in main process.
206
+ # Replace it by an empty dict instead to make it easier for downstream code to handle
207
+ if results is None:
208
+ results = {}
209
+ return results
210
+
211
+
212
+ @contextmanager
213
+ def inference_context(model):
214
+ """
215
+ A context where the model is temporarily changed to eval mode,
216
+ and restored to previous mode afterwards.
217
+
218
+ Args:
219
+ model: a torch Module
220
+ """
221
+ training_mode = model.training
222
+ model.eval()
223
+ yield
224
+ model.train(training_mode)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/fast_eval_api.py ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import copy
3
+ import logging
4
+ import numpy as np
5
+ import time
6
+ from pycocotools.cocoeval import COCOeval
7
+
8
+ from annotator.oneformer.detectron2 import _C
9
+
10
+ logger = logging.getLogger(__name__)
11
+
12
+
13
+ class COCOeval_opt(COCOeval):
14
+ """
15
+ This is a slightly modified version of the original COCO API, where the functions evaluateImg()
16
+ and accumulate() are implemented in C++ to speedup evaluation
17
+ """
18
+
19
+ def evaluate(self):
20
+ """
21
+ Run per image evaluation on given images and store results in self.evalImgs_cpp, a
22
+ datastructure that isn't readable from Python but is used by a c++ implementation of
23
+ accumulate(). Unlike the original COCO PythonAPI, we don't populate the datastructure
24
+ self.evalImgs because this datastructure is a computational bottleneck.
25
+ :return: None
26
+ """
27
+ tic = time.time()
28
+
29
+ p = self.params
30
+ # add backward compatibility if useSegm is specified in params
31
+ if p.useSegm is not None:
32
+ p.iouType = "segm" if p.useSegm == 1 else "bbox"
33
+ logger.info("Evaluate annotation type *{}*".format(p.iouType))
34
+ p.imgIds = list(np.unique(p.imgIds))
35
+ if p.useCats:
36
+ p.catIds = list(np.unique(p.catIds))
37
+ p.maxDets = sorted(p.maxDets)
38
+ self.params = p
39
+
40
+ self._prepare() # bottleneck
41
+
42
+ # loop through images, area range, max detection number
43
+ catIds = p.catIds if p.useCats else [-1]
44
+
45
+ if p.iouType == "segm" or p.iouType == "bbox":
46
+ computeIoU = self.computeIoU
47
+ elif p.iouType == "keypoints":
48
+ computeIoU = self.computeOks
49
+ self.ious = {
50
+ (imgId, catId): computeIoU(imgId, catId) for imgId in p.imgIds for catId in catIds
51
+ } # bottleneck
52
+
53
+ maxDet = p.maxDets[-1]
54
+
55
+ # <<<< Beginning of code differences with original COCO API
56
+ def convert_instances_to_cpp(instances, is_det=False):
57
+ # Convert annotations for a list of instances in an image to a format that's fast
58
+ # to access in C++
59
+ instances_cpp = []
60
+ for instance in instances:
61
+ instance_cpp = _C.InstanceAnnotation(
62
+ int(instance["id"]),
63
+ instance["score"] if is_det else instance.get("score", 0.0),
64
+ instance["area"],
65
+ bool(instance.get("iscrowd", 0)),
66
+ bool(instance.get("ignore", 0)),
67
+ )
68
+ instances_cpp.append(instance_cpp)
69
+ return instances_cpp
70
+
71
+ # Convert GT annotations, detections, and IOUs to a format that's fast to access in C++
72
+ ground_truth_instances = [
73
+ [convert_instances_to_cpp(self._gts[imgId, catId]) for catId in p.catIds]
74
+ for imgId in p.imgIds
75
+ ]
76
+ detected_instances = [
77
+ [convert_instances_to_cpp(self._dts[imgId, catId], is_det=True) for catId in p.catIds]
78
+ for imgId in p.imgIds
79
+ ]
80
+ ious = [[self.ious[imgId, catId] for catId in catIds] for imgId in p.imgIds]
81
+
82
+ if not p.useCats:
83
+ # For each image, flatten per-category lists into a single list
84
+ ground_truth_instances = [[[o for c in i for o in c]] for i in ground_truth_instances]
85
+ detected_instances = [[[o for c in i for o in c]] for i in detected_instances]
86
+
87
+ # Call C++ implementation of self.evaluateImgs()
88
+ self._evalImgs_cpp = _C.COCOevalEvaluateImages(
89
+ p.areaRng, maxDet, p.iouThrs, ious, ground_truth_instances, detected_instances
90
+ )
91
+ self._evalImgs = None
92
+
93
+ self._paramsEval = copy.deepcopy(self.params)
94
+ toc = time.time()
95
+ logger.info("COCOeval_opt.evaluate() finished in {:0.2f} seconds.".format(toc - tic))
96
+ # >>>> End of code differences with original COCO API
97
+
98
+ def accumulate(self):
99
+ """
100
+ Accumulate per image evaluation results and store the result in self.eval. Does not
101
+ support changing parameter settings from those used by self.evaluate()
102
+ """
103
+ logger.info("Accumulating evaluation results...")
104
+ tic = time.time()
105
+ assert hasattr(
106
+ self, "_evalImgs_cpp"
107
+ ), "evaluate() must be called before accmulate() is called."
108
+
109
+ self.eval = _C.COCOevalAccumulate(self._paramsEval, self._evalImgs_cpp)
110
+
111
+ # recall is num_iou_thresholds X num_categories X num_area_ranges X num_max_detections
112
+ self.eval["recall"] = np.array(self.eval["recall"]).reshape(
113
+ self.eval["counts"][:1] + self.eval["counts"][2:]
114
+ )
115
+
116
+ # precision and scores are num_iou_thresholds X num_recall_thresholds X num_categories X
117
+ # num_area_ranges X num_max_detections
118
+ self.eval["precision"] = np.array(self.eval["precision"]).reshape(self.eval["counts"])
119
+ self.eval["scores"] = np.array(self.eval["scores"]).reshape(self.eval["counts"])
120
+ toc = time.time()
121
+ logger.info("COCOeval_opt.accumulate() finished in {:0.2f} seconds.".format(toc - tic))
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/lvis_evaluation.py ADDED
@@ -0,0 +1,380 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import copy
3
+ import itertools
4
+ import json
5
+ import logging
6
+ import os
7
+ import pickle
8
+ from collections import OrderedDict
9
+ import torch
10
+
11
+ import annotator.oneformer.detectron2.utils.comm as comm
12
+ from annotator.oneformer.detectron2.config import CfgNode
13
+ from annotator.oneformer.detectron2.data import MetadataCatalog
14
+ from annotator.oneformer.detectron2.structures import Boxes, BoxMode, pairwise_iou
15
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
16
+ from annotator.oneformer.detectron2.utils.logger import create_small_table
17
+
18
+ from .coco_evaluation import instances_to_coco_json
19
+ from .evaluator import DatasetEvaluator
20
+
21
+
22
+ class LVISEvaluator(DatasetEvaluator):
23
+ """
24
+ Evaluate object proposal and instance detection/segmentation outputs using
25
+ LVIS's metrics and evaluation API.
26
+ """
27
+
28
+ def __init__(
29
+ self,
30
+ dataset_name,
31
+ tasks=None,
32
+ distributed=True,
33
+ output_dir=None,
34
+ *,
35
+ max_dets_per_image=None,
36
+ ):
37
+ """
38
+ Args:
39
+ dataset_name (str): name of the dataset to be evaluated.
40
+ It must have the following corresponding metadata:
41
+ "json_file": the path to the LVIS format annotation
42
+ tasks (tuple[str]): tasks that can be evaluated under the given
43
+ configuration. A task is one of "bbox", "segm".
44
+ By default, will infer this automatically from predictions.
45
+ distributed (True): if True, will collect results from all ranks for evaluation.
46
+ Otherwise, will evaluate the results in the current process.
47
+ output_dir (str): optional, an output directory to dump results.
48
+ max_dets_per_image (None or int): limit on maximum detections per image in evaluating AP
49
+ This limit, by default of the LVIS dataset, is 300.
50
+ """
51
+ from lvis import LVIS
52
+
53
+ self._logger = logging.getLogger(__name__)
54
+
55
+ if tasks is not None and isinstance(tasks, CfgNode):
56
+ self._logger.warn(
57
+ "COCO Evaluator instantiated using config, this is deprecated behavior."
58
+ " Please pass in explicit arguments instead."
59
+ )
60
+ self._tasks = None # Infering it from predictions should be better
61
+ else:
62
+ self._tasks = tasks
63
+
64
+ self._distributed = distributed
65
+ self._output_dir = output_dir
66
+ self._max_dets_per_image = max_dets_per_image
67
+
68
+ self._cpu_device = torch.device("cpu")
69
+
70
+ self._metadata = MetadataCatalog.get(dataset_name)
71
+ json_file = PathManager.get_local_path(self._metadata.json_file)
72
+ self._lvis_api = LVIS(json_file)
73
+ # Test set json files do not contain annotations (evaluation must be
74
+ # performed using the LVIS evaluation server).
75
+ self._do_evaluation = len(self._lvis_api.get_ann_ids()) > 0
76
+
77
+ def reset(self):
78
+ self._predictions = []
79
+
80
+ def process(self, inputs, outputs):
81
+ """
82
+ Args:
83
+ inputs: the inputs to a LVIS model (e.g., GeneralizedRCNN).
84
+ It is a list of dict. Each dict corresponds to an image and
85
+ contains keys like "height", "width", "file_name", "image_id".
86
+ outputs: the outputs of a LVIS model. It is a list of dicts with key
87
+ "instances" that contains :class:`Instances`.
88
+ """
89
+ for input, output in zip(inputs, outputs):
90
+ prediction = {"image_id": input["image_id"]}
91
+
92
+ if "instances" in output:
93
+ instances = output["instances"].to(self._cpu_device)
94
+ prediction["instances"] = instances_to_coco_json(instances, input["image_id"])
95
+ if "proposals" in output:
96
+ prediction["proposals"] = output["proposals"].to(self._cpu_device)
97
+ self._predictions.append(prediction)
98
+
99
+ def evaluate(self):
100
+ if self._distributed:
101
+ comm.synchronize()
102
+ predictions = comm.gather(self._predictions, dst=0)
103
+ predictions = list(itertools.chain(*predictions))
104
+
105
+ if not comm.is_main_process():
106
+ return
107
+ else:
108
+ predictions = self._predictions
109
+
110
+ if len(predictions) == 0:
111
+ self._logger.warning("[LVISEvaluator] Did not receive valid predictions.")
112
+ return {}
113
+
114
+ if self._output_dir:
115
+ PathManager.mkdirs(self._output_dir)
116
+ file_path = os.path.join(self._output_dir, "instances_predictions.pth")
117
+ with PathManager.open(file_path, "wb") as f:
118
+ torch.save(predictions, f)
119
+
120
+ self._results = OrderedDict()
121
+ if "proposals" in predictions[0]:
122
+ self._eval_box_proposals(predictions)
123
+ if "instances" in predictions[0]:
124
+ self._eval_predictions(predictions)
125
+ # Copy so the caller can do whatever with results
126
+ return copy.deepcopy(self._results)
127
+
128
+ def _tasks_from_predictions(self, predictions):
129
+ for pred in predictions:
130
+ if "segmentation" in pred:
131
+ return ("bbox", "segm")
132
+ return ("bbox",)
133
+
134
+ def _eval_predictions(self, predictions):
135
+ """
136
+ Evaluate predictions. Fill self._results with the metrics of the tasks.
137
+
138
+ Args:
139
+ predictions (list[dict]): list of outputs from the model
140
+ """
141
+ self._logger.info("Preparing results in the LVIS format ...")
142
+ lvis_results = list(itertools.chain(*[x["instances"] for x in predictions]))
143
+ tasks = self._tasks or self._tasks_from_predictions(lvis_results)
144
+
145
+ # LVIS evaluator can be used to evaluate results for COCO dataset categories.
146
+ # In this case `_metadata` variable will have a field with COCO-specific category mapping.
147
+ if hasattr(self._metadata, "thing_dataset_id_to_contiguous_id"):
148
+ reverse_id_mapping = {
149
+ v: k for k, v in self._metadata.thing_dataset_id_to_contiguous_id.items()
150
+ }
151
+ for result in lvis_results:
152
+ result["category_id"] = reverse_id_mapping[result["category_id"]]
153
+ else:
154
+ # unmap the category ids for LVIS (from 0-indexed to 1-indexed)
155
+ for result in lvis_results:
156
+ result["category_id"] += 1
157
+
158
+ if self._output_dir:
159
+ file_path = os.path.join(self._output_dir, "lvis_instances_results.json")
160
+ self._logger.info("Saving results to {}".format(file_path))
161
+ with PathManager.open(file_path, "w") as f:
162
+ f.write(json.dumps(lvis_results))
163
+ f.flush()
164
+
165
+ if not self._do_evaluation:
166
+ self._logger.info("Annotations are not available for evaluation.")
167
+ return
168
+
169
+ self._logger.info("Evaluating predictions ...")
170
+ for task in sorted(tasks):
171
+ res = _evaluate_predictions_on_lvis(
172
+ self._lvis_api,
173
+ lvis_results,
174
+ task,
175
+ max_dets_per_image=self._max_dets_per_image,
176
+ class_names=self._metadata.get("thing_classes"),
177
+ )
178
+ self._results[task] = res
179
+
180
+ def _eval_box_proposals(self, predictions):
181
+ """
182
+ Evaluate the box proposals in predictions.
183
+ Fill self._results with the metrics for "box_proposals" task.
184
+ """
185
+ if self._output_dir:
186
+ # Saving generated box proposals to file.
187
+ # Predicted box_proposals are in XYXY_ABS mode.
188
+ bbox_mode = BoxMode.XYXY_ABS.value
189
+ ids, boxes, objectness_logits = [], [], []
190
+ for prediction in predictions:
191
+ ids.append(prediction["image_id"])
192
+ boxes.append(prediction["proposals"].proposal_boxes.tensor.numpy())
193
+ objectness_logits.append(prediction["proposals"].objectness_logits.numpy())
194
+
195
+ proposal_data = {
196
+ "boxes": boxes,
197
+ "objectness_logits": objectness_logits,
198
+ "ids": ids,
199
+ "bbox_mode": bbox_mode,
200
+ }
201
+ with PathManager.open(os.path.join(self._output_dir, "box_proposals.pkl"), "wb") as f:
202
+ pickle.dump(proposal_data, f)
203
+
204
+ if not self._do_evaluation:
205
+ self._logger.info("Annotations are not available for evaluation.")
206
+ return
207
+
208
+ self._logger.info("Evaluating bbox proposals ...")
209
+ res = {}
210
+ areas = {"all": "", "small": "s", "medium": "m", "large": "l"}
211
+ for limit in [100, 1000]:
212
+ for area, suffix in areas.items():
213
+ stats = _evaluate_box_proposals(predictions, self._lvis_api, area=area, limit=limit)
214
+ key = "AR{}@{:d}".format(suffix, limit)
215
+ res[key] = float(stats["ar"].item() * 100)
216
+ self._logger.info("Proposal metrics: \n" + create_small_table(res))
217
+ self._results["box_proposals"] = res
218
+
219
+
220
+ # inspired from Detectron:
221
+ # https://github.com/facebookresearch/Detectron/blob/a6a835f5b8208c45d0dce217ce9bbda915f44df7/detectron/datasets/json_dataset_evaluator.py#L255 # noqa
222
+ def _evaluate_box_proposals(dataset_predictions, lvis_api, thresholds=None, area="all", limit=None):
223
+ """
224
+ Evaluate detection proposal recall metrics. This function is a much
225
+ faster alternative to the official LVIS API recall evaluation code. However,
226
+ it produces slightly different results.
227
+ """
228
+ # Record max overlap value for each gt box
229
+ # Return vector of overlap values
230
+ areas = {
231
+ "all": 0,
232
+ "small": 1,
233
+ "medium": 2,
234
+ "large": 3,
235
+ "96-128": 4,
236
+ "128-256": 5,
237
+ "256-512": 6,
238
+ "512-inf": 7,
239
+ }
240
+ area_ranges = [
241
+ [0**2, 1e5**2], # all
242
+ [0**2, 32**2], # small
243
+ [32**2, 96**2], # medium
244
+ [96**2, 1e5**2], # large
245
+ [96**2, 128**2], # 96-128
246
+ [128**2, 256**2], # 128-256
247
+ [256**2, 512**2], # 256-512
248
+ [512**2, 1e5**2],
249
+ ] # 512-inf
250
+ assert area in areas, "Unknown area range: {}".format(area)
251
+ area_range = area_ranges[areas[area]]
252
+ gt_overlaps = []
253
+ num_pos = 0
254
+
255
+ for prediction_dict in dataset_predictions:
256
+ predictions = prediction_dict["proposals"]
257
+
258
+ # sort predictions in descending order
259
+ # TODO maybe remove this and make it explicit in the documentation
260
+ inds = predictions.objectness_logits.sort(descending=True)[1]
261
+ predictions = predictions[inds]
262
+
263
+ ann_ids = lvis_api.get_ann_ids(img_ids=[prediction_dict["image_id"]])
264
+ anno = lvis_api.load_anns(ann_ids)
265
+ gt_boxes = [
266
+ BoxMode.convert(obj["bbox"], BoxMode.XYWH_ABS, BoxMode.XYXY_ABS) for obj in anno
267
+ ]
268
+ gt_boxes = torch.as_tensor(gt_boxes).reshape(-1, 4) # guard against no boxes
269
+ gt_boxes = Boxes(gt_boxes)
270
+ gt_areas = torch.as_tensor([obj["area"] for obj in anno])
271
+
272
+ if len(gt_boxes) == 0 or len(predictions) == 0:
273
+ continue
274
+
275
+ valid_gt_inds = (gt_areas >= area_range[0]) & (gt_areas <= area_range[1])
276
+ gt_boxes = gt_boxes[valid_gt_inds]
277
+
278
+ num_pos += len(gt_boxes)
279
+
280
+ if len(gt_boxes) == 0:
281
+ continue
282
+
283
+ if limit is not None and len(predictions) > limit:
284
+ predictions = predictions[:limit]
285
+
286
+ overlaps = pairwise_iou(predictions.proposal_boxes, gt_boxes)
287
+
288
+ _gt_overlaps = torch.zeros(len(gt_boxes))
289
+ for j in range(min(len(predictions), len(gt_boxes))):
290
+ # find which proposal box maximally covers each gt box
291
+ # and get the iou amount of coverage for each gt box
292
+ max_overlaps, argmax_overlaps = overlaps.max(dim=0)
293
+
294
+ # find which gt box is 'best' covered (i.e. 'best' = most iou)
295
+ gt_ovr, gt_ind = max_overlaps.max(dim=0)
296
+ assert gt_ovr >= 0
297
+ # find the proposal box that covers the best covered gt box
298
+ box_ind = argmax_overlaps[gt_ind]
299
+ # record the iou coverage of this gt box
300
+ _gt_overlaps[j] = overlaps[box_ind, gt_ind]
301
+ assert _gt_overlaps[j] == gt_ovr
302
+ # mark the proposal box and the gt box as used
303
+ overlaps[box_ind, :] = -1
304
+ overlaps[:, gt_ind] = -1
305
+
306
+ # append recorded iou coverage level
307
+ gt_overlaps.append(_gt_overlaps)
308
+ gt_overlaps = (
309
+ torch.cat(gt_overlaps, dim=0) if len(gt_overlaps) else torch.zeros(0, dtype=torch.float32)
310
+ )
311
+ gt_overlaps, _ = torch.sort(gt_overlaps)
312
+
313
+ if thresholds is None:
314
+ step = 0.05
315
+ thresholds = torch.arange(0.5, 0.95 + 1e-5, step, dtype=torch.float32)
316
+ recalls = torch.zeros_like(thresholds)
317
+ # compute recall for each iou threshold
318
+ for i, t in enumerate(thresholds):
319
+ recalls[i] = (gt_overlaps >= t).float().sum() / float(num_pos)
320
+ # ar = 2 * np.trapz(recalls, thresholds)
321
+ ar = recalls.mean()
322
+ return {
323
+ "ar": ar,
324
+ "recalls": recalls,
325
+ "thresholds": thresholds,
326
+ "gt_overlaps": gt_overlaps,
327
+ "num_pos": num_pos,
328
+ }
329
+
330
+
331
+ def _evaluate_predictions_on_lvis(
332
+ lvis_gt, lvis_results, iou_type, max_dets_per_image=None, class_names=None
333
+ ):
334
+ """
335
+ Args:
336
+ iou_type (str):
337
+ max_dets_per_image (None or int): limit on maximum detections per image in evaluating AP
338
+ This limit, by default of the LVIS dataset, is 300.
339
+ class_names (None or list[str]): if provided, will use it to predict
340
+ per-category AP.
341
+
342
+ Returns:
343
+ a dict of {metric name: score}
344
+ """
345
+ metrics = {
346
+ "bbox": ["AP", "AP50", "AP75", "APs", "APm", "APl", "APr", "APc", "APf"],
347
+ "segm": ["AP", "AP50", "AP75", "APs", "APm", "APl", "APr", "APc", "APf"],
348
+ }[iou_type]
349
+
350
+ logger = logging.getLogger(__name__)
351
+
352
+ if len(lvis_results) == 0: # TODO: check if needed
353
+ logger.warn("No predictions from the model!")
354
+ return {metric: float("nan") for metric in metrics}
355
+
356
+ if iou_type == "segm":
357
+ lvis_results = copy.deepcopy(lvis_results)
358
+ # When evaluating mask AP, if the results contain bbox, LVIS API will
359
+ # use the box area as the area of the instance, instead of the mask area.
360
+ # This leads to a different definition of small/medium/large.
361
+ # We remove the bbox field to let mask AP use mask area.
362
+ for c in lvis_results:
363
+ c.pop("bbox", None)
364
+
365
+ if max_dets_per_image is None:
366
+ max_dets_per_image = 300 # Default for LVIS dataset
367
+
368
+ from lvis import LVISEval, LVISResults
369
+
370
+ logger.info(f"Evaluating with max detections per image = {max_dets_per_image}")
371
+ lvis_results = LVISResults(lvis_gt, lvis_results, max_dets=max_dets_per_image)
372
+ lvis_eval = LVISEval(lvis_gt, lvis_results, iou_type)
373
+ lvis_eval.run()
374
+ lvis_eval.print_results()
375
+
376
+ # Pull the standard metrics from the LVIS results
377
+ results = lvis_eval.get_results()
378
+ results = {metric: float(results[metric] * 100) for metric in metrics}
379
+ logger.info("Evaluation results for {}: \n".format(iou_type) + create_small_table(results))
380
+ return results
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/panoptic_evaluation.py ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import contextlib
3
+ import io
4
+ import itertools
5
+ import json
6
+ import logging
7
+ import numpy as np
8
+ import os
9
+ import tempfile
10
+ from collections import OrderedDict
11
+ from typing import Optional
12
+ from PIL import Image
13
+ from tabulate import tabulate
14
+
15
+ from annotator.oneformer.detectron2.data import MetadataCatalog
16
+ from annotator.oneformer.detectron2.utils import comm
17
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
18
+
19
+ from .evaluator import DatasetEvaluator
20
+
21
+ logger = logging.getLogger(__name__)
22
+
23
+
24
+ class COCOPanopticEvaluator(DatasetEvaluator):
25
+ """
26
+ Evaluate Panoptic Quality metrics on COCO using PanopticAPI.
27
+ It saves panoptic segmentation prediction in `output_dir`
28
+
29
+ It contains a synchronize call and has to be called from all workers.
30
+ """
31
+
32
+ def __init__(self, dataset_name: str, output_dir: Optional[str] = None):
33
+ """
34
+ Args:
35
+ dataset_name: name of the dataset
36
+ output_dir: output directory to save results for evaluation.
37
+ """
38
+ self._metadata = MetadataCatalog.get(dataset_name)
39
+ self._thing_contiguous_id_to_dataset_id = {
40
+ v: k for k, v in self._metadata.thing_dataset_id_to_contiguous_id.items()
41
+ }
42
+ self._stuff_contiguous_id_to_dataset_id = {
43
+ v: k for k, v in self._metadata.stuff_dataset_id_to_contiguous_id.items()
44
+ }
45
+
46
+ self._output_dir = output_dir
47
+ if self._output_dir is not None:
48
+ PathManager.mkdirs(self._output_dir)
49
+
50
+ def reset(self):
51
+ self._predictions = []
52
+
53
+ def _convert_category_id(self, segment_info):
54
+ isthing = segment_info.pop("isthing", None)
55
+ if isthing is None:
56
+ # the model produces panoptic category id directly. No more conversion needed
57
+ return segment_info
58
+ if isthing is True:
59
+ segment_info["category_id"] = self._thing_contiguous_id_to_dataset_id[
60
+ segment_info["category_id"]
61
+ ]
62
+ else:
63
+ segment_info["category_id"] = self._stuff_contiguous_id_to_dataset_id[
64
+ segment_info["category_id"]
65
+ ]
66
+ return segment_info
67
+
68
+ def process(self, inputs, outputs):
69
+ from panopticapi.utils import id2rgb
70
+
71
+ for input, output in zip(inputs, outputs):
72
+ panoptic_img, segments_info = output["panoptic_seg"]
73
+ panoptic_img = panoptic_img.cpu().numpy()
74
+ if segments_info is None:
75
+ # If "segments_info" is None, we assume "panoptic_img" is a
76
+ # H*W int32 image storing the panoptic_id in the format of
77
+ # category_id * label_divisor + instance_id. We reserve -1 for
78
+ # VOID label, and add 1 to panoptic_img since the official
79
+ # evaluation script uses 0 for VOID label.
80
+ label_divisor = self._metadata.label_divisor
81
+ segments_info = []
82
+ for panoptic_label in np.unique(panoptic_img):
83
+ if panoptic_label == -1:
84
+ # VOID region.
85
+ continue
86
+ pred_class = panoptic_label // label_divisor
87
+ isthing = (
88
+ pred_class in self._metadata.thing_dataset_id_to_contiguous_id.values()
89
+ )
90
+ segments_info.append(
91
+ {
92
+ "id": int(panoptic_label) + 1,
93
+ "category_id": int(pred_class),
94
+ "isthing": bool(isthing),
95
+ }
96
+ )
97
+ # Official evaluation script uses 0 for VOID label.
98
+ panoptic_img += 1
99
+
100
+ file_name = os.path.basename(input["file_name"])
101
+ file_name_png = os.path.splitext(file_name)[0] + ".png"
102
+ with io.BytesIO() as out:
103
+ Image.fromarray(id2rgb(panoptic_img)).save(out, format="PNG")
104
+ segments_info = [self._convert_category_id(x) for x in segments_info]
105
+ self._predictions.append(
106
+ {
107
+ "image_id": input["image_id"],
108
+ "file_name": file_name_png,
109
+ "png_string": out.getvalue(),
110
+ "segments_info": segments_info,
111
+ }
112
+ )
113
+
114
+ def evaluate(self):
115
+ comm.synchronize()
116
+
117
+ self._predictions = comm.gather(self._predictions)
118
+ self._predictions = list(itertools.chain(*self._predictions))
119
+ if not comm.is_main_process():
120
+ return
121
+
122
+ # PanopticApi requires local files
123
+ gt_json = PathManager.get_local_path(self._metadata.panoptic_json)
124
+ gt_folder = PathManager.get_local_path(self._metadata.panoptic_root)
125
+
126
+ with tempfile.TemporaryDirectory(prefix="panoptic_eval") as pred_dir:
127
+ logger.info("Writing all panoptic predictions to {} ...".format(pred_dir))
128
+ for p in self._predictions:
129
+ with open(os.path.join(pred_dir, p["file_name"]), "wb") as f:
130
+ f.write(p.pop("png_string"))
131
+
132
+ with open(gt_json, "r") as f:
133
+ json_data = json.load(f)
134
+ json_data["annotations"] = self._predictions
135
+
136
+ output_dir = self._output_dir or pred_dir
137
+ predictions_json = os.path.join(output_dir, "predictions.json")
138
+ with PathManager.open(predictions_json, "w") as f:
139
+ f.write(json.dumps(json_data))
140
+
141
+ from panopticapi.evaluation import pq_compute
142
+
143
+ with contextlib.redirect_stdout(io.StringIO()):
144
+ pq_res = pq_compute(
145
+ gt_json,
146
+ PathManager.get_local_path(predictions_json),
147
+ gt_folder=gt_folder,
148
+ pred_folder=pred_dir,
149
+ )
150
+
151
+ res = {}
152
+ res["PQ"] = 100 * pq_res["All"]["pq"]
153
+ res["SQ"] = 100 * pq_res["All"]["sq"]
154
+ res["RQ"] = 100 * pq_res["All"]["rq"]
155
+ res["PQ_th"] = 100 * pq_res["Things"]["pq"]
156
+ res["SQ_th"] = 100 * pq_res["Things"]["sq"]
157
+ res["RQ_th"] = 100 * pq_res["Things"]["rq"]
158
+ res["PQ_st"] = 100 * pq_res["Stuff"]["pq"]
159
+ res["SQ_st"] = 100 * pq_res["Stuff"]["sq"]
160
+ res["RQ_st"] = 100 * pq_res["Stuff"]["rq"]
161
+
162
+ results = OrderedDict({"panoptic_seg": res})
163
+ _print_panoptic_results(pq_res)
164
+
165
+ return results
166
+
167
+
168
+ def _print_panoptic_results(pq_res):
169
+ headers = ["", "PQ", "SQ", "RQ", "#categories"]
170
+ data = []
171
+ for name in ["All", "Things", "Stuff"]:
172
+ row = [name] + [pq_res[name][k] * 100 for k in ["pq", "sq", "rq"]] + [pq_res[name]["n"]]
173
+ data.append(row)
174
+ table = tabulate(
175
+ data, headers=headers, tablefmt="pipe", floatfmt=".3f", stralign="center", numalign="center"
176
+ )
177
+ logger.info("Panoptic Evaluation Results:\n" + table)
178
+
179
+
180
+ if __name__ == "__main__":
181
+ from annotator.oneformer.detectron2.utils.logger import setup_logger
182
+
183
+ logger = setup_logger()
184
+ import argparse
185
+
186
+ parser = argparse.ArgumentParser()
187
+ parser.add_argument("--gt-json")
188
+ parser.add_argument("--gt-dir")
189
+ parser.add_argument("--pred-json")
190
+ parser.add_argument("--pred-dir")
191
+ args = parser.parse_args()
192
+
193
+ from panopticapi.evaluation import pq_compute
194
+
195
+ with contextlib.redirect_stdout(io.StringIO()):
196
+ pq_res = pq_compute(
197
+ args.gt_json, args.pred_json, gt_folder=args.gt_dir, pred_folder=args.pred_dir
198
+ )
199
+ _print_panoptic_results(pq_res)
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/pascal_voc_evaluation.py ADDED
@@ -0,0 +1,300 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # -*- coding: utf-8 -*-
2
+ # Copyright (c) Facebook, Inc. and its affiliates.
3
+
4
+ import logging
5
+ import numpy as np
6
+ import os
7
+ import tempfile
8
+ import xml.etree.ElementTree as ET
9
+ from collections import OrderedDict, defaultdict
10
+ from functools import lru_cache
11
+ import torch
12
+
13
+ from annotator.oneformer.detectron2.data import MetadataCatalog
14
+ from annotator.oneformer.detectron2.utils import comm
15
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
16
+
17
+ from .evaluator import DatasetEvaluator
18
+
19
+
20
+ class PascalVOCDetectionEvaluator(DatasetEvaluator):
21
+ """
22
+ Evaluate Pascal VOC style AP for Pascal VOC dataset.
23
+ It contains a synchronization, therefore has to be called from all ranks.
24
+
25
+ Note that the concept of AP can be implemented in different ways and may not
26
+ produce identical results. This class mimics the implementation of the official
27
+ Pascal VOC Matlab API, and should produce similar but not identical results to the
28
+ official API.
29
+ """
30
+
31
+ def __init__(self, dataset_name):
32
+ """
33
+ Args:
34
+ dataset_name (str): name of the dataset, e.g., "voc_2007_test"
35
+ """
36
+ self._dataset_name = dataset_name
37
+ meta = MetadataCatalog.get(dataset_name)
38
+
39
+ # Too many tiny files, download all to local for speed.
40
+ annotation_dir_local = PathManager.get_local_path(
41
+ os.path.join(meta.dirname, "Annotations/")
42
+ )
43
+ self._anno_file_template = os.path.join(annotation_dir_local, "{}.xml")
44
+ self._image_set_path = os.path.join(meta.dirname, "ImageSets", "Main", meta.split + ".txt")
45
+ self._class_names = meta.thing_classes
46
+ assert meta.year in [2007, 2012], meta.year
47
+ self._is_2007 = meta.year == 2007
48
+ self._cpu_device = torch.device("cpu")
49
+ self._logger = logging.getLogger(__name__)
50
+
51
+ def reset(self):
52
+ self._predictions = defaultdict(list) # class name -> list of prediction strings
53
+
54
+ def process(self, inputs, outputs):
55
+ for input, output in zip(inputs, outputs):
56
+ image_id = input["image_id"]
57
+ instances = output["instances"].to(self._cpu_device)
58
+ boxes = instances.pred_boxes.tensor.numpy()
59
+ scores = instances.scores.tolist()
60
+ classes = instances.pred_classes.tolist()
61
+ for box, score, cls in zip(boxes, scores, classes):
62
+ xmin, ymin, xmax, ymax = box
63
+ # The inverse of data loading logic in `datasets/pascal_voc.py`
64
+ xmin += 1
65
+ ymin += 1
66
+ self._predictions[cls].append(
67
+ f"{image_id} {score:.3f} {xmin:.1f} {ymin:.1f} {xmax:.1f} {ymax:.1f}"
68
+ )
69
+
70
+ def evaluate(self):
71
+ """
72
+ Returns:
73
+ dict: has a key "segm", whose value is a dict of "AP", "AP50", and "AP75".
74
+ """
75
+ all_predictions = comm.gather(self._predictions, dst=0)
76
+ if not comm.is_main_process():
77
+ return
78
+ predictions = defaultdict(list)
79
+ for predictions_per_rank in all_predictions:
80
+ for clsid, lines in predictions_per_rank.items():
81
+ predictions[clsid].extend(lines)
82
+ del all_predictions
83
+
84
+ self._logger.info(
85
+ "Evaluating {} using {} metric. "
86
+ "Note that results do not use the official Matlab API.".format(
87
+ self._dataset_name, 2007 if self._is_2007 else 2012
88
+ )
89
+ )
90
+
91
+ with tempfile.TemporaryDirectory(prefix="pascal_voc_eval_") as dirname:
92
+ res_file_template = os.path.join(dirname, "{}.txt")
93
+
94
+ aps = defaultdict(list) # iou -> ap per class
95
+ for cls_id, cls_name in enumerate(self._class_names):
96
+ lines = predictions.get(cls_id, [""])
97
+
98
+ with open(res_file_template.format(cls_name), "w") as f:
99
+ f.write("\n".join(lines))
100
+
101
+ for thresh in range(50, 100, 5):
102
+ rec, prec, ap = voc_eval(
103
+ res_file_template,
104
+ self._anno_file_template,
105
+ self._image_set_path,
106
+ cls_name,
107
+ ovthresh=thresh / 100.0,
108
+ use_07_metric=self._is_2007,
109
+ )
110
+ aps[thresh].append(ap * 100)
111
+
112
+ ret = OrderedDict()
113
+ mAP = {iou: np.mean(x) for iou, x in aps.items()}
114
+ ret["bbox"] = {"AP": np.mean(list(mAP.values())), "AP50": mAP[50], "AP75": mAP[75]}
115
+ return ret
116
+
117
+
118
+ ##############################################################################
119
+ #
120
+ # Below code is modified from
121
+ # https://github.com/rbgirshick/py-faster-rcnn/blob/master/lib/datasets/voc_eval.py
122
+ # --------------------------------------------------------
123
+ # Fast/er R-CNN
124
+ # Licensed under The MIT License [see LICENSE for details]
125
+ # Written by Bharath Hariharan
126
+ # --------------------------------------------------------
127
+
128
+ """Python implementation of the PASCAL VOC devkit's AP evaluation code."""
129
+
130
+
131
+ @lru_cache(maxsize=None)
132
+ def parse_rec(filename):
133
+ """Parse a PASCAL VOC xml file."""
134
+ with PathManager.open(filename) as f:
135
+ tree = ET.parse(f)
136
+ objects = []
137
+ for obj in tree.findall("object"):
138
+ obj_struct = {}
139
+ obj_struct["name"] = obj.find("name").text
140
+ obj_struct["pose"] = obj.find("pose").text
141
+ obj_struct["truncated"] = int(obj.find("truncated").text)
142
+ obj_struct["difficult"] = int(obj.find("difficult").text)
143
+ bbox = obj.find("bndbox")
144
+ obj_struct["bbox"] = [
145
+ int(bbox.find("xmin").text),
146
+ int(bbox.find("ymin").text),
147
+ int(bbox.find("xmax").text),
148
+ int(bbox.find("ymax").text),
149
+ ]
150
+ objects.append(obj_struct)
151
+
152
+ return objects
153
+
154
+
155
+ def voc_ap(rec, prec, use_07_metric=False):
156
+ """Compute VOC AP given precision and recall. If use_07_metric is true, uses
157
+ the VOC 07 11-point method (default:False).
158
+ """
159
+ if use_07_metric:
160
+ # 11 point metric
161
+ ap = 0.0
162
+ for t in np.arange(0.0, 1.1, 0.1):
163
+ if np.sum(rec >= t) == 0:
164
+ p = 0
165
+ else:
166
+ p = np.max(prec[rec >= t])
167
+ ap = ap + p / 11.0
168
+ else:
169
+ # correct AP calculation
170
+ # first append sentinel values at the end
171
+ mrec = np.concatenate(([0.0], rec, [1.0]))
172
+ mpre = np.concatenate(([0.0], prec, [0.0]))
173
+
174
+ # compute the precision envelope
175
+ for i in range(mpre.size - 1, 0, -1):
176
+ mpre[i - 1] = np.maximum(mpre[i - 1], mpre[i])
177
+
178
+ # to calculate area under PR curve, look for points
179
+ # where X axis (recall) changes value
180
+ i = np.where(mrec[1:] != mrec[:-1])[0]
181
+
182
+ # and sum (\Delta recall) * prec
183
+ ap = np.sum((mrec[i + 1] - mrec[i]) * mpre[i + 1])
184
+ return ap
185
+
186
+
187
+ def voc_eval(detpath, annopath, imagesetfile, classname, ovthresh=0.5, use_07_metric=False):
188
+ """rec, prec, ap = voc_eval(detpath,
189
+ annopath,
190
+ imagesetfile,
191
+ classname,
192
+ [ovthresh],
193
+ [use_07_metric])
194
+
195
+ Top level function that does the PASCAL VOC evaluation.
196
+
197
+ detpath: Path to detections
198
+ detpath.format(classname) should produce the detection results file.
199
+ annopath: Path to annotations
200
+ annopath.format(imagename) should be the xml annotations file.
201
+ imagesetfile: Text file containing the list of images, one image per line.
202
+ classname: Category name (duh)
203
+ [ovthresh]: Overlap threshold (default = 0.5)
204
+ [use_07_metric]: Whether to use VOC07's 11 point AP computation
205
+ (default False)
206
+ """
207
+ # assumes detections are in detpath.format(classname)
208
+ # assumes annotations are in annopath.format(imagename)
209
+ # assumes imagesetfile is a text file with each line an image name
210
+
211
+ # first load gt
212
+ # read list of images
213
+ with PathManager.open(imagesetfile, "r") as f:
214
+ lines = f.readlines()
215
+ imagenames = [x.strip() for x in lines]
216
+
217
+ # load annots
218
+ recs = {}
219
+ for imagename in imagenames:
220
+ recs[imagename] = parse_rec(annopath.format(imagename))
221
+
222
+ # extract gt objects for this class
223
+ class_recs = {}
224
+ npos = 0
225
+ for imagename in imagenames:
226
+ R = [obj for obj in recs[imagename] if obj["name"] == classname]
227
+ bbox = np.array([x["bbox"] for x in R])
228
+ difficult = np.array([x["difficult"] for x in R]).astype(bool)
229
+ # difficult = np.array([False for x in R]).astype(bool) # treat all "difficult" as GT
230
+ det = [False] * len(R)
231
+ npos = npos + sum(~difficult)
232
+ class_recs[imagename] = {"bbox": bbox, "difficult": difficult, "det": det}
233
+
234
+ # read dets
235
+ detfile = detpath.format(classname)
236
+ with open(detfile, "r") as f:
237
+ lines = f.readlines()
238
+
239
+ splitlines = [x.strip().split(" ") for x in lines]
240
+ image_ids = [x[0] for x in splitlines]
241
+ confidence = np.array([float(x[1]) for x in splitlines])
242
+ BB = np.array([[float(z) for z in x[2:]] for x in splitlines]).reshape(-1, 4)
243
+
244
+ # sort by confidence
245
+ sorted_ind = np.argsort(-confidence)
246
+ BB = BB[sorted_ind, :]
247
+ image_ids = [image_ids[x] for x in sorted_ind]
248
+
249
+ # go down dets and mark TPs and FPs
250
+ nd = len(image_ids)
251
+ tp = np.zeros(nd)
252
+ fp = np.zeros(nd)
253
+ for d in range(nd):
254
+ R = class_recs[image_ids[d]]
255
+ bb = BB[d, :].astype(float)
256
+ ovmax = -np.inf
257
+ BBGT = R["bbox"].astype(float)
258
+
259
+ if BBGT.size > 0:
260
+ # compute overlaps
261
+ # intersection
262
+ ixmin = np.maximum(BBGT[:, 0], bb[0])
263
+ iymin = np.maximum(BBGT[:, 1], bb[1])
264
+ ixmax = np.minimum(BBGT[:, 2], bb[2])
265
+ iymax = np.minimum(BBGT[:, 3], bb[3])
266
+ iw = np.maximum(ixmax - ixmin + 1.0, 0.0)
267
+ ih = np.maximum(iymax - iymin + 1.0, 0.0)
268
+ inters = iw * ih
269
+
270
+ # union
271
+ uni = (
272
+ (bb[2] - bb[0] + 1.0) * (bb[3] - bb[1] + 1.0)
273
+ + (BBGT[:, 2] - BBGT[:, 0] + 1.0) * (BBGT[:, 3] - BBGT[:, 1] + 1.0)
274
+ - inters
275
+ )
276
+
277
+ overlaps = inters / uni
278
+ ovmax = np.max(overlaps)
279
+ jmax = np.argmax(overlaps)
280
+
281
+ if ovmax > ovthresh:
282
+ if not R["difficult"][jmax]:
283
+ if not R["det"][jmax]:
284
+ tp[d] = 1.0
285
+ R["det"][jmax] = 1
286
+ else:
287
+ fp[d] = 1.0
288
+ else:
289
+ fp[d] = 1.0
290
+
291
+ # compute precision recall
292
+ fp = np.cumsum(fp)
293
+ tp = np.cumsum(tp)
294
+ rec = tp / float(npos)
295
+ # avoid divide by zero in case the first detection matches a difficult
296
+ # ground truth
297
+ prec = tp / np.maximum(tp + fp, np.finfo(np.float64).eps)
298
+ ap = voc_ap(rec, prec, use_07_metric)
299
+
300
+ return rec, prec, ap
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/rotated_coco_evaluation.py ADDED
@@ -0,0 +1,207 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import itertools
3
+ import json
4
+ import numpy as np
5
+ import os
6
+ import torch
7
+ from pycocotools.cocoeval import COCOeval, maskUtils
8
+
9
+ from annotator.oneformer.detectron2.structures import BoxMode, RotatedBoxes, pairwise_iou_rotated
10
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
11
+
12
+ from .coco_evaluation import COCOEvaluator
13
+
14
+
15
+ class RotatedCOCOeval(COCOeval):
16
+ @staticmethod
17
+ def is_rotated(box_list):
18
+ if type(box_list) == np.ndarray:
19
+ return box_list.shape[1] == 5
20
+ elif type(box_list) == list:
21
+ if box_list == []: # cannot decide the box_dim
22
+ return False
23
+ return np.all(
24
+ np.array(
25
+ [
26
+ (len(obj) == 5) and ((type(obj) == list) or (type(obj) == np.ndarray))
27
+ for obj in box_list
28
+ ]
29
+ )
30
+ )
31
+ return False
32
+
33
+ @staticmethod
34
+ def boxlist_to_tensor(boxlist, output_box_dim):
35
+ if type(boxlist) == np.ndarray:
36
+ box_tensor = torch.from_numpy(boxlist)
37
+ elif type(boxlist) == list:
38
+ if boxlist == []:
39
+ return torch.zeros((0, output_box_dim), dtype=torch.float32)
40
+ else:
41
+ box_tensor = torch.FloatTensor(boxlist)
42
+ else:
43
+ raise Exception("Unrecognized boxlist type")
44
+
45
+ input_box_dim = box_tensor.shape[1]
46
+ if input_box_dim != output_box_dim:
47
+ if input_box_dim == 4 and output_box_dim == 5:
48
+ box_tensor = BoxMode.convert(box_tensor, BoxMode.XYWH_ABS, BoxMode.XYWHA_ABS)
49
+ else:
50
+ raise Exception(
51
+ "Unable to convert from {}-dim box to {}-dim box".format(
52
+ input_box_dim, output_box_dim
53
+ )
54
+ )
55
+ return box_tensor
56
+
57
+ def compute_iou_dt_gt(self, dt, gt, is_crowd):
58
+ if self.is_rotated(dt) or self.is_rotated(gt):
59
+ # TODO: take is_crowd into consideration
60
+ assert all(c == 0 for c in is_crowd)
61
+ dt = RotatedBoxes(self.boxlist_to_tensor(dt, output_box_dim=5))
62
+ gt = RotatedBoxes(self.boxlist_to_tensor(gt, output_box_dim=5))
63
+ return pairwise_iou_rotated(dt, gt)
64
+ else:
65
+ # This is the same as the classical COCO evaluation
66
+ return maskUtils.iou(dt, gt, is_crowd)
67
+
68
+ def computeIoU(self, imgId, catId):
69
+ p = self.params
70
+ if p.useCats:
71
+ gt = self._gts[imgId, catId]
72
+ dt = self._dts[imgId, catId]
73
+ else:
74
+ gt = [_ for cId in p.catIds for _ in self._gts[imgId, cId]]
75
+ dt = [_ for cId in p.catIds for _ in self._dts[imgId, cId]]
76
+ if len(gt) == 0 and len(dt) == 0:
77
+ return []
78
+ inds = np.argsort([-d["score"] for d in dt], kind="mergesort")
79
+ dt = [dt[i] for i in inds]
80
+ if len(dt) > p.maxDets[-1]:
81
+ dt = dt[0 : p.maxDets[-1]]
82
+
83
+ assert p.iouType == "bbox", "unsupported iouType for iou computation"
84
+
85
+ g = [g["bbox"] for g in gt]
86
+ d = [d["bbox"] for d in dt]
87
+
88
+ # compute iou between each dt and gt region
89
+ iscrowd = [int(o["iscrowd"]) for o in gt]
90
+
91
+ # Note: this function is copied from cocoeval.py in cocoapi
92
+ # and the major difference is here.
93
+ ious = self.compute_iou_dt_gt(d, g, iscrowd)
94
+ return ious
95
+
96
+
97
+ class RotatedCOCOEvaluator(COCOEvaluator):
98
+ """
99
+ Evaluate object proposal/instance detection outputs using COCO-like metrics and APIs,
100
+ with rotated boxes support.
101
+ Note: this uses IOU only and does not consider angle differences.
102
+ """
103
+
104
+ def process(self, inputs, outputs):
105
+ """
106
+ Args:
107
+ inputs: the inputs to a COCO model (e.g., GeneralizedRCNN).
108
+ It is a list of dict. Each dict corresponds to an image and
109
+ contains keys like "height", "width", "file_name", "image_id".
110
+ outputs: the outputs of a COCO model. It is a list of dicts with key
111
+ "instances" that contains :class:`Instances`.
112
+ """
113
+ for input, output in zip(inputs, outputs):
114
+ prediction = {"image_id": input["image_id"]}
115
+
116
+ if "instances" in output:
117
+ instances = output["instances"].to(self._cpu_device)
118
+
119
+ prediction["instances"] = self.instances_to_json(instances, input["image_id"])
120
+ if "proposals" in output:
121
+ prediction["proposals"] = output["proposals"].to(self._cpu_device)
122
+ self._predictions.append(prediction)
123
+
124
+ def instances_to_json(self, instances, img_id):
125
+ num_instance = len(instances)
126
+ if num_instance == 0:
127
+ return []
128
+
129
+ boxes = instances.pred_boxes.tensor.numpy()
130
+ if boxes.shape[1] == 4:
131
+ boxes = BoxMode.convert(boxes, BoxMode.XYXY_ABS, BoxMode.XYWH_ABS)
132
+ boxes = boxes.tolist()
133
+ scores = instances.scores.tolist()
134
+ classes = instances.pred_classes.tolist()
135
+
136
+ results = []
137
+ for k in range(num_instance):
138
+ result = {
139
+ "image_id": img_id,
140
+ "category_id": classes[k],
141
+ "bbox": boxes[k],
142
+ "score": scores[k],
143
+ }
144
+
145
+ results.append(result)
146
+ return results
147
+
148
+ def _eval_predictions(self, predictions, img_ids=None): # img_ids: unused
149
+ """
150
+ Evaluate predictions on the given tasks.
151
+ Fill self._results with the metrics of the tasks.
152
+ """
153
+ self._logger.info("Preparing results for COCO format ...")
154
+ coco_results = list(itertools.chain(*[x["instances"] for x in predictions]))
155
+
156
+ # unmap the category ids for COCO
157
+ if hasattr(self._metadata, "thing_dataset_id_to_contiguous_id"):
158
+ reverse_id_mapping = {
159
+ v: k for k, v in self._metadata.thing_dataset_id_to_contiguous_id.items()
160
+ }
161
+ for result in coco_results:
162
+ result["category_id"] = reverse_id_mapping[result["category_id"]]
163
+
164
+ if self._output_dir:
165
+ file_path = os.path.join(self._output_dir, "coco_instances_results.json")
166
+ self._logger.info("Saving results to {}".format(file_path))
167
+ with PathManager.open(file_path, "w") as f:
168
+ f.write(json.dumps(coco_results))
169
+ f.flush()
170
+
171
+ if not self._do_evaluation:
172
+ self._logger.info("Annotations are not available for evaluation.")
173
+ return
174
+
175
+ self._logger.info("Evaluating predictions ...")
176
+
177
+ assert self._tasks is None or set(self._tasks) == {
178
+ "bbox"
179
+ }, "[RotatedCOCOEvaluator] Only bbox evaluation is supported"
180
+ coco_eval = (
181
+ self._evaluate_predictions_on_coco(self._coco_api, coco_results)
182
+ if len(coco_results) > 0
183
+ else None # cocoapi does not handle empty results very well
184
+ )
185
+
186
+ task = "bbox"
187
+ res = self._derive_coco_results(
188
+ coco_eval, task, class_names=self._metadata.get("thing_classes")
189
+ )
190
+ self._results[task] = res
191
+
192
+ def _evaluate_predictions_on_coco(self, coco_gt, coco_results):
193
+ """
194
+ Evaluate the coco results using COCOEval API.
195
+ """
196
+ assert len(coco_results) > 0
197
+
198
+ coco_dt = coco_gt.loadRes(coco_results)
199
+
200
+ # Only bbox is supported for now
201
+ coco_eval = RotatedCOCOeval(coco_gt, coco_dt, iouType="bbox")
202
+
203
+ coco_eval.evaluate()
204
+ coco_eval.accumulate()
205
+ coco_eval.summarize()
206
+
207
+ return coco_eval
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/sem_seg_evaluation.py ADDED
@@ -0,0 +1,265 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import itertools
3
+ import json
4
+ import logging
5
+ import numpy as np
6
+ import os
7
+ from collections import OrderedDict
8
+ from typing import Optional, Union
9
+ import pycocotools.mask as mask_util
10
+ import torch
11
+ from PIL import Image
12
+
13
+ from annotator.oneformer.detectron2.data import DatasetCatalog, MetadataCatalog
14
+ from annotator.oneformer.detectron2.utils.comm import all_gather, is_main_process, synchronize
15
+ from annotator.oneformer.detectron2.utils.file_io import PathManager
16
+
17
+ from .evaluator import DatasetEvaluator
18
+
19
+ _CV2_IMPORTED = True
20
+ try:
21
+ import cv2 # noqa
22
+ except ImportError:
23
+ # OpenCV is an optional dependency at the moment
24
+ _CV2_IMPORTED = False
25
+
26
+
27
+ def load_image_into_numpy_array(
28
+ filename: str,
29
+ copy: bool = False,
30
+ dtype: Optional[Union[np.dtype, str]] = None,
31
+ ) -> np.ndarray:
32
+ with PathManager.open(filename, "rb") as f:
33
+ array = np.array(Image.open(f), copy=copy, dtype=dtype)
34
+ return array
35
+
36
+
37
+ class SemSegEvaluator(DatasetEvaluator):
38
+ """
39
+ Evaluate semantic segmentation metrics.
40
+ """
41
+
42
+ def __init__(
43
+ self,
44
+ dataset_name,
45
+ distributed=True,
46
+ output_dir=None,
47
+ *,
48
+ sem_seg_loading_fn=load_image_into_numpy_array,
49
+ num_classes=None,
50
+ ignore_label=None,
51
+ ):
52
+ """
53
+ Args:
54
+ dataset_name (str): name of the dataset to be evaluated.
55
+ distributed (bool): if True, will collect results from all ranks for evaluation.
56
+ Otherwise, will evaluate the results in the current process.
57
+ output_dir (str): an output directory to dump results.
58
+ sem_seg_loading_fn: function to read sem seg file and load into numpy array.
59
+ Default provided, but projects can customize.
60
+ num_classes, ignore_label: deprecated argument
61
+ """
62
+ self._logger = logging.getLogger(__name__)
63
+ if num_classes is not None:
64
+ self._logger.warn(
65
+ "SemSegEvaluator(num_classes) is deprecated! It should be obtained from metadata."
66
+ )
67
+ if ignore_label is not None:
68
+ self._logger.warn(
69
+ "SemSegEvaluator(ignore_label) is deprecated! It should be obtained from metadata."
70
+ )
71
+ self._dataset_name = dataset_name
72
+ self._distributed = distributed
73
+ self._output_dir = output_dir
74
+
75
+ self._cpu_device = torch.device("cpu")
76
+
77
+ self.input_file_to_gt_file = {
78
+ dataset_record["file_name"]: dataset_record["sem_seg_file_name"]
79
+ for dataset_record in DatasetCatalog.get(dataset_name)
80
+ }
81
+
82
+ meta = MetadataCatalog.get(dataset_name)
83
+ # Dict that maps contiguous training ids to COCO category ids
84
+ try:
85
+ c2d = meta.stuff_dataset_id_to_contiguous_id
86
+ self._contiguous_id_to_dataset_id = {v: k for k, v in c2d.items()}
87
+ except AttributeError:
88
+ self._contiguous_id_to_dataset_id = None
89
+ self._class_names = meta.stuff_classes
90
+ self.sem_seg_loading_fn = sem_seg_loading_fn
91
+ self._num_classes = len(meta.stuff_classes)
92
+ if num_classes is not None:
93
+ assert self._num_classes == num_classes, f"{self._num_classes} != {num_classes}"
94
+ self._ignore_label = ignore_label if ignore_label is not None else meta.ignore_label
95
+
96
+ # This is because cv2.erode did not work for int datatype. Only works for uint8.
97
+ self._compute_boundary_iou = True
98
+ if not _CV2_IMPORTED:
99
+ self._compute_boundary_iou = False
100
+ self._logger.warn(
101
+ """Boundary IoU calculation requires OpenCV. B-IoU metrics are
102
+ not going to be computed because OpenCV is not available to import."""
103
+ )
104
+ if self._num_classes >= np.iinfo(np.uint8).max:
105
+ self._compute_boundary_iou = False
106
+ self._logger.warn(
107
+ f"""SemSegEvaluator(num_classes) is more than supported value for Boundary IoU calculation!
108
+ B-IoU metrics are not going to be computed. Max allowed value (exclusive)
109
+ for num_classes for calculating Boundary IoU is {np.iinfo(np.uint8).max}.
110
+ The number of classes of dataset {self._dataset_name} is {self._num_classes}"""
111
+ )
112
+
113
+ def reset(self):
114
+ self._conf_matrix = np.zeros((self._num_classes + 1, self._num_classes + 1), dtype=np.int64)
115
+ self._b_conf_matrix = np.zeros(
116
+ (self._num_classes + 1, self._num_classes + 1), dtype=np.int64
117
+ )
118
+ self._predictions = []
119
+
120
+ def process(self, inputs, outputs):
121
+ """
122
+ Args:
123
+ inputs: the inputs to a model.
124
+ It is a list of dicts. Each dict corresponds to an image and
125
+ contains keys like "height", "width", "file_name".
126
+ outputs: the outputs of a model. It is either list of semantic segmentation predictions
127
+ (Tensor [H, W]) or list of dicts with key "sem_seg" that contains semantic
128
+ segmentation prediction in the same format.
129
+ """
130
+ for input, output in zip(inputs, outputs):
131
+ output = output["sem_seg"].argmax(dim=0).to(self._cpu_device)
132
+ pred = np.array(output, dtype=np.int)
133
+ gt_filename = self.input_file_to_gt_file[input["file_name"]]
134
+ gt = self.sem_seg_loading_fn(gt_filename, dtype=np.int)
135
+
136
+ gt[gt == self._ignore_label] = self._num_classes
137
+
138
+ self._conf_matrix += np.bincount(
139
+ (self._num_classes + 1) * pred.reshape(-1) + gt.reshape(-1),
140
+ minlength=self._conf_matrix.size,
141
+ ).reshape(self._conf_matrix.shape)
142
+
143
+ if self._compute_boundary_iou:
144
+ b_gt = self._mask_to_boundary(gt.astype(np.uint8))
145
+ b_pred = self._mask_to_boundary(pred.astype(np.uint8))
146
+
147
+ self._b_conf_matrix += np.bincount(
148
+ (self._num_classes + 1) * b_pred.reshape(-1) + b_gt.reshape(-1),
149
+ minlength=self._conf_matrix.size,
150
+ ).reshape(self._conf_matrix.shape)
151
+
152
+ self._predictions.extend(self.encode_json_sem_seg(pred, input["file_name"]))
153
+
154
+ def evaluate(self):
155
+ """
156
+ Evaluates standard semantic segmentation metrics (http://cocodataset.org/#stuff-eval):
157
+
158
+ * Mean intersection-over-union averaged across classes (mIoU)
159
+ * Frequency Weighted IoU (fwIoU)
160
+ * Mean pixel accuracy averaged across classes (mACC)
161
+ * Pixel Accuracy (pACC)
162
+ """
163
+ if self._distributed:
164
+ synchronize()
165
+ conf_matrix_list = all_gather(self._conf_matrix)
166
+ b_conf_matrix_list = all_gather(self._b_conf_matrix)
167
+ self._predictions = all_gather(self._predictions)
168
+ self._predictions = list(itertools.chain(*self._predictions))
169
+ if not is_main_process():
170
+ return
171
+
172
+ self._conf_matrix = np.zeros_like(self._conf_matrix)
173
+ for conf_matrix in conf_matrix_list:
174
+ self._conf_matrix += conf_matrix
175
+
176
+ self._b_conf_matrix = np.zeros_like(self._b_conf_matrix)
177
+ for b_conf_matrix in b_conf_matrix_list:
178
+ self._b_conf_matrix += b_conf_matrix
179
+
180
+ if self._output_dir:
181
+ PathManager.mkdirs(self._output_dir)
182
+ file_path = os.path.join(self._output_dir, "sem_seg_predictions.json")
183
+ with PathManager.open(file_path, "w") as f:
184
+ f.write(json.dumps(self._predictions))
185
+
186
+ acc = np.full(self._num_classes, np.nan, dtype=np.float)
187
+ iou = np.full(self._num_classes, np.nan, dtype=np.float)
188
+ tp = self._conf_matrix.diagonal()[:-1].astype(np.float)
189
+ pos_gt = np.sum(self._conf_matrix[:-1, :-1], axis=0).astype(np.float)
190
+ class_weights = pos_gt / np.sum(pos_gt)
191
+ pos_pred = np.sum(self._conf_matrix[:-1, :-1], axis=1).astype(np.float)
192
+ acc_valid = pos_gt > 0
193
+ acc[acc_valid] = tp[acc_valid] / pos_gt[acc_valid]
194
+ union = pos_gt + pos_pred - tp
195
+ iou_valid = np.logical_and(acc_valid, union > 0)
196
+ iou[iou_valid] = tp[iou_valid] / union[iou_valid]
197
+ macc = np.sum(acc[acc_valid]) / np.sum(acc_valid)
198
+ miou = np.sum(iou[iou_valid]) / np.sum(iou_valid)
199
+ fiou = np.sum(iou[iou_valid] * class_weights[iou_valid])
200
+ pacc = np.sum(tp) / np.sum(pos_gt)
201
+
202
+ if self._compute_boundary_iou:
203
+ b_iou = np.full(self._num_classes, np.nan, dtype=np.float)
204
+ b_tp = self._b_conf_matrix.diagonal()[:-1].astype(np.float)
205
+ b_pos_gt = np.sum(self._b_conf_matrix[:-1, :-1], axis=0).astype(np.float)
206
+ b_pos_pred = np.sum(self._b_conf_matrix[:-1, :-1], axis=1).astype(np.float)
207
+ b_union = b_pos_gt + b_pos_pred - b_tp
208
+ b_iou_valid = b_union > 0
209
+ b_iou[b_iou_valid] = b_tp[b_iou_valid] / b_union[b_iou_valid]
210
+
211
+ res = {}
212
+ res["mIoU"] = 100 * miou
213
+ res["fwIoU"] = 100 * fiou
214
+ for i, name in enumerate(self._class_names):
215
+ res[f"IoU-{name}"] = 100 * iou[i]
216
+ if self._compute_boundary_iou:
217
+ res[f"BoundaryIoU-{name}"] = 100 * b_iou[i]
218
+ res[f"min(IoU, B-Iou)-{name}"] = 100 * min(iou[i], b_iou[i])
219
+ res["mACC"] = 100 * macc
220
+ res["pACC"] = 100 * pacc
221
+ for i, name in enumerate(self._class_names):
222
+ res[f"ACC-{name}"] = 100 * acc[i]
223
+
224
+ if self._output_dir:
225
+ file_path = os.path.join(self._output_dir, "sem_seg_evaluation.pth")
226
+ with PathManager.open(file_path, "wb") as f:
227
+ torch.save(res, f)
228
+ results = OrderedDict({"sem_seg": res})
229
+ self._logger.info(results)
230
+ return results
231
+
232
+ def encode_json_sem_seg(self, sem_seg, input_file_name):
233
+ """
234
+ Convert semantic segmentation to COCO stuff format with segments encoded as RLEs.
235
+ See http://cocodataset.org/#format-results
236
+ """
237
+ json_list = []
238
+ for label in np.unique(sem_seg):
239
+ if self._contiguous_id_to_dataset_id is not None:
240
+ assert (
241
+ label in self._contiguous_id_to_dataset_id
242
+ ), "Label {} is not in the metadata info for {}".format(label, self._dataset_name)
243
+ dataset_id = self._contiguous_id_to_dataset_id[label]
244
+ else:
245
+ dataset_id = int(label)
246
+ mask = (sem_seg == label).astype(np.uint8)
247
+ mask_rle = mask_util.encode(np.array(mask[:, :, None], order="F"))[0]
248
+ mask_rle["counts"] = mask_rle["counts"].decode("utf-8")
249
+ json_list.append(
250
+ {"file_name": input_file_name, "category_id": dataset_id, "segmentation": mask_rle}
251
+ )
252
+ return json_list
253
+
254
+ def _mask_to_boundary(self, mask: np.ndarray, dilation_ratio=0.02):
255
+ assert mask.ndim == 2, "mask_to_boundary expects a 2-dimensional image"
256
+ h, w = mask.shape
257
+ diag_len = np.sqrt(h**2 + w**2)
258
+ dilation = max(1, int(round(dilation_ratio * diag_len)))
259
+ kernel = np.ones((3, 3), dtype=np.uint8)
260
+
261
+ padded_mask = cv2.copyMakeBorder(mask, 1, 1, 1, 1, cv2.BORDER_CONSTANT, value=0)
262
+ eroded_mask_with_padding = cv2.erode(padded_mask, kernel, iterations=dilation)
263
+ eroded_mask = eroded_mask_with_padding[1:-1, 1:-1]
264
+ boundary = mask - eroded_mask
265
+ return boundary
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/evaluation/testing.py ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import logging
3
+ import numpy as np
4
+ import pprint
5
+ import sys
6
+ from collections.abc import Mapping
7
+
8
+
9
+ def print_csv_format(results):
10
+ """
11
+ Print main metrics in a format similar to Detectron,
12
+ so that they are easy to copypaste into a spreadsheet.
13
+
14
+ Args:
15
+ results (OrderedDict[dict]): task_name -> {metric -> score}
16
+ unordered dict can also be printed, but in arbitrary order
17
+ """
18
+ assert isinstance(results, Mapping) or not len(results), results
19
+ logger = logging.getLogger(__name__)
20
+ for task, res in results.items():
21
+ if isinstance(res, Mapping):
22
+ # Don't print "AP-category" metrics since they are usually not tracked.
23
+ important_res = [(k, v) for k, v in res.items() if "-" not in k]
24
+ logger.info("copypaste: Task: {}".format(task))
25
+ logger.info("copypaste: " + ",".join([k[0] for k in important_res]))
26
+ logger.info("copypaste: " + ",".join(["{0:.4f}".format(k[1]) for k in important_res]))
27
+ else:
28
+ logger.info(f"copypaste: {task}={res}")
29
+
30
+
31
+ def verify_results(cfg, results):
32
+ """
33
+ Args:
34
+ results (OrderedDict[dict]): task_name -> {metric -> score}
35
+
36
+ Returns:
37
+ bool: whether the verification succeeds or not
38
+ """
39
+ expected_results = cfg.TEST.EXPECTED_RESULTS
40
+ if not len(expected_results):
41
+ return True
42
+
43
+ ok = True
44
+ for task, metric, expected, tolerance in expected_results:
45
+ actual = results[task].get(metric, None)
46
+ if actual is None:
47
+ ok = False
48
+ continue
49
+ if not np.isfinite(actual):
50
+ ok = False
51
+ continue
52
+ diff = abs(actual - expected)
53
+ if diff > tolerance:
54
+ ok = False
55
+
56
+ logger = logging.getLogger(__name__)
57
+ if not ok:
58
+ logger.error("Result verification failed!")
59
+ logger.error("Expected Results: " + str(expected_results))
60
+ logger.error("Actual Results: " + pprint.pformat(results))
61
+
62
+ sys.exit(1)
63
+ else:
64
+ logger.info("Results verification passed.")
65
+ return ok
66
+
67
+
68
+ def flatten_results_dict(results):
69
+ """
70
+ Expand a hierarchical dict of scalars into a flat dict of scalars.
71
+ If results[k1][k2][k3] = v, the returned dict will have the entry
72
+ {"k1/k2/k3": v}.
73
+
74
+ Args:
75
+ results (dict):
76
+ """
77
+ r = {}
78
+ for k, v in results.items():
79
+ if isinstance(v, Mapping):
80
+ v = flatten_results_dict(v)
81
+ for kk, vv in v.items():
82
+ r[k + "/" + kk] = vv
83
+ else:
84
+ r[k] = v
85
+ return r
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/model_zoo/__init__.py ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ """
3
+ Model Zoo API for Detectron2: a collection of functions to create common model architectures
4
+ listed in `MODEL_ZOO.md <https://github.com/facebookresearch/detectron2/blob/main/MODEL_ZOO.md>`_,
5
+ and optionally load their pre-trained weights.
6
+ """
7
+
8
+ from .model_zoo import get, get_config_file, get_checkpoint_url, get_config
9
+
10
+ __all__ = ["get_checkpoint_url", "get", "get_config_file", "get_config"]
CCEdit-main/src/controlnet11/annotator/oneformer/detectron2/model_zoo/model_zoo.py ADDED
@@ -0,0 +1,213 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Copyright (c) Facebook, Inc. and its affiliates.
2
+ import os
3
+ from typing import Optional
4
+ import pkg_resources
5
+ import torch
6
+
7
+ from annotator.oneformer.detectron2.checkpoint import DetectionCheckpointer
8
+ from annotator.oneformer.detectron2.config import CfgNode, LazyConfig, get_cfg, instantiate
9
+ from annotator.oneformer.detectron2.modeling import build_model
10
+
11
+
12
+ class _ModelZooUrls(object):
13
+ """
14
+ Mapping from names to officially released Detectron2 pre-trained models.
15
+ """
16
+
17
+ S3_PREFIX = "https://dl.fbaipublicfiles.com/detectron2/"
18
+
19
+ # format: {config_path.yaml} -> model_id/model_final_{commit}.pkl
20
+ CONFIG_PATH_TO_URL_SUFFIX = {
21
+ # COCO Detection with Faster R-CNN
22
+ "COCO-Detection/faster_rcnn_R_50_C4_1x": "137257644/model_final_721ade.pkl",
23
+ "COCO-Detection/faster_rcnn_R_50_DC5_1x": "137847829/model_final_51d356.pkl",
24
+ "COCO-Detection/faster_rcnn_R_50_FPN_1x": "137257794/model_final_b275ba.pkl",
25
+ "COCO-Detection/faster_rcnn_R_50_C4_3x": "137849393/model_final_f97cb7.pkl",
26
+ "COCO-Detection/faster_rcnn_R_50_DC5_3x": "137849425/model_final_68d202.pkl",
27
+ "COCO-Detection/faster_rcnn_R_50_FPN_3x": "137849458/model_final_280758.pkl",
28
+ "COCO-Detection/faster_rcnn_R_101_C4_3x": "138204752/model_final_298dad.pkl",
29
+ "COCO-Detection/faster_rcnn_R_101_DC5_3x": "138204841/model_final_3e0943.pkl",
30
+ "COCO-Detection/faster_rcnn_R_101_FPN_3x": "137851257/model_final_f6e8b1.pkl",
31
+ "COCO-Detection/faster_rcnn_X_101_32x8d_FPN_3x": "139173657/model_final_68b088.pkl",
32
+ # COCO Detection with RetinaNet
33
+ "COCO-Detection/retinanet_R_50_FPN_1x": "190397773/model_final_bfca0b.pkl",
34
+ "COCO-Detection/retinanet_R_50_FPN_3x": "190397829/model_final_5bd44e.pkl",
35
+ "COCO-Detection/retinanet_R_101_FPN_3x": "190397697/model_final_971ab9.pkl",
36
+ # COCO Detection with RPN and Fast R-CNN
37
+ "COCO-Detection/rpn_R_50_C4_1x": "137258005/model_final_450694.pkl",
38
+ "COCO-Detection/rpn_R_50_FPN_1x": "137258492/model_final_02ce48.pkl",
39
+ "COCO-Detection/fast_rcnn_R_50_FPN_1x": "137635226/model_final_e5f7ce.pkl",
40
+ # COCO Instance Segmentation Baselines with Mask R-CNN
41
+ "COCO-InstanceSegmentation/mask_rcnn_R_50_C4_1x": "137259246/model_final_9243eb.pkl",
42
+ "COCO-InstanceSegmentation/mask_rcnn_R_50_DC5_1x": "137260150/model_final_4f86c3.pkl",
43
+ "COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_1x": "137260431/model_final_a54504.pkl",
44
+ "COCO-InstanceSegmentation/mask_rcnn_R_50_C4_3x": "137849525/model_final_4ce675.pkl",
45
+ "COCO-InstanceSegmentation/mask_rcnn_R_50_DC5_3x": "137849551/model_final_84107b.pkl",
46
+ "COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_3x": "137849600/model_final_f10217.pkl",
47
+ "COCO-InstanceSegmentation/mask_rcnn_R_101_C4_3x": "138363239/model_final_a2914c.pkl",
48
+ "COCO-InstanceSegmentation/mask_rcnn_R_101_DC5_3x": "138363294/model_final_0464b7.pkl",
49
+ "COCO-InstanceSegmentation/mask_rcnn_R_101_FPN_3x": "138205316/model_final_a3ec72.pkl",
50
+ "COCO-InstanceSegmentation/mask_rcnn_X_101_32x8d_FPN_3x": "139653917/model_final_2d9806.pkl", # noqa
51
+ # New baselines using Large-Scale Jitter and Longer Training Schedule
52
+ "new_baselines/mask_rcnn_R_50_FPN_100ep_LSJ": "42047764/model_final_bb69de.pkl",
53
+ "new_baselines/mask_rcnn_R_50_FPN_200ep_LSJ": "42047638/model_final_89a8d3.pkl",
54
+ "new_baselines/mask_rcnn_R_50_FPN_400ep_LSJ": "42019571/model_final_14d201.pkl",
55
+ "new_baselines/mask_rcnn_R_101_FPN_100ep_LSJ": "42025812/model_final_4f7b58.pkl",
56
+ "new_baselines/mask_rcnn_R_101_FPN_200ep_LSJ": "42131867/model_final_0bb7ae.pkl",
57
+ "new_baselines/mask_rcnn_R_101_FPN_400ep_LSJ": "42073830/model_final_f96b26.pkl",
58
+ "new_baselines/mask_rcnn_regnetx_4gf_dds_FPN_100ep_LSJ": "42047771/model_final_b7fbab.pkl", # noqa
59
+ "new_baselines/mask_rcnn_regnetx_4gf_dds_FPN_200ep_LSJ": "42132721/model_final_5d87c1.pkl", # noqa
60
+ "new_baselines/mask_rcnn_regnetx_4gf_dds_FPN_400ep_LSJ": "42025447/model_final_f1362d.pkl", # noqa
61
+ "new_baselines/mask_rcnn_regnety_4gf_dds_FPN_100ep_LSJ": "42047784/model_final_6ba57e.pkl", # noqa
62
+ "new_baselines/mask_rcnn_regnety_4gf_dds_FPN_200ep_LSJ": "42047642/model_final_27b9c1.pkl", # noqa
63
+ "new_baselines/mask_rcnn_regnety_4gf_dds_FPN_400ep_LSJ": "42045954/model_final_ef3a80.pkl", # noqa
64
+ # COCO Person Keypoint Detection Baselines with Keypoint R-CNN
65
+ "COCO-Keypoints/keypoint_rcnn_R_50_FPN_1x": "137261548/model_final_04e291.pkl",
66
+ "COCO-Keypoints/keypoint_rcnn_R_50_FPN_3x": "137849621/model_final_a6e10b.pkl",
67
+ "COCO-Keypoints/keypoint_rcnn_R_101_FPN_3x": "138363331/model_final_997cc7.pkl",
68
+ "COCO-Keypoints/keypoint_rcnn_X_101_32x8d_FPN_3x": "139686956/model_final_5ad38f.pkl",
69
+ # COCO Panoptic Segmentation Baselines with Panoptic FPN
70
+ "COCO-PanopticSegmentation/panoptic_fpn_R_50_1x": "139514544/model_final_dbfeb4.pkl",
71
+ "COCO-PanopticSegmentation/panoptic_fpn_R_50_3x": "139514569/model_final_c10459.pkl",
72
+ "COCO-PanopticSegmentation/panoptic_fpn_R_101_3x": "139514519/model_final_cafdb1.pkl",
73
+ # LVIS Instance Segmentation Baselines with Mask R-CNN
74
+ "LVISv0.5-InstanceSegmentation/mask_rcnn_R_50_FPN_1x": "144219072/model_final_571f7c.pkl", # noqa
75
+ "LVISv0.5-InstanceSegmentation/mask_rcnn_R_101_FPN_1x": "144219035/model_final_824ab5.pkl", # noqa
76
+ "LVISv0.5-InstanceSegmentation/mask_rcnn_X_101_32x8d_FPN_1x": "144219108/model_final_5e3439.pkl", # noqa
77
+ # Cityscapes & Pascal VOC Baselines
78
+ "Cityscapes/mask_rcnn_R_50_FPN": "142423278/model_final_af9cf5.pkl",
79
+ "PascalVOC-Detection/faster_rcnn_R_50_C4": "142202221/model_final_b1acc2.pkl",
80
+ # Other Settings
81
+ "Misc/mask_rcnn_R_50_FPN_1x_dconv_c3-c5": "138602867/model_final_65c703.pkl",
82
+ "Misc/mask_rcnn_R_50_FPN_3x_dconv_c3-c5": "144998336/model_final_821d0b.pkl",
83
+ "Misc/cascade_mask_rcnn_R_50_FPN_1x": "138602847/model_final_e9d89b.pkl",
84
+ "Misc/cascade_mask_rcnn_R_50_FPN_3x": "144998488/model_final_480dd8.pkl",
85
+ "Misc/mask_rcnn_R_50_FPN_3x_syncbn": "169527823/model_final_3b3c51.pkl",
86
+ "Misc/mask_rcnn_R_50_FPN_3x_gn": "138602888/model_final_dc5d9e.pkl",
87
+ "Misc/scratch_mask_rcnn_R_50_FPN_3x_gn": "138602908/model_final_01ca85.pkl",
88
+ "Misc/scratch_mask_rcnn_R_50_FPN_9x_gn": "183808979/model_final_da7b4c.pkl",
89
+ "Misc/scratch_mask_rcnn_R_50_FPN_9x_syncbn": "184226666/model_final_5ce33e.pkl",
90
+ "Misc/panoptic_fpn_R_101_dconv_cascade_gn_3x": "139797668/model_final_be35db.pkl",
91
+ "Misc/cascade_mask_rcnn_X_152_32x8d_FPN_IN5k_gn_dconv": "18131413/model_0039999_e76410.pkl", # noqa
92
+ # D1 Comparisons
93
+ "Detectron1-Comparisons/faster_rcnn_R_50_FPN_noaug_1x": "137781054/model_final_7ab50c.pkl", # noqa
94
+ "Detectron1-Comparisons/mask_rcnn_R_50_FPN_noaug_1x": "137781281/model_final_62ca52.pkl", # noqa
95
+ "Detectron1-Comparisons/keypoint_rcnn_R_50_FPN_1x": "137781195/model_final_cce136.pkl",
96
+ }
97
+
98
+ @staticmethod
99
+ def query(config_path: str) -> Optional[str]:
100
+ """
101
+ Args:
102
+ config_path: relative config filename
103
+ """
104
+ name = config_path.replace(".yaml", "").replace(".py", "")
105
+ if name in _ModelZooUrls.CONFIG_PATH_TO_URL_SUFFIX:
106
+ suffix = _ModelZooUrls.CONFIG_PATH_TO_URL_SUFFIX[name]
107
+ return _ModelZooUrls.S3_PREFIX + name + "/" + suffix
108
+ return None
109
+
110
+
111
+ def get_checkpoint_url(config_path):
112
+ """
113
+ Returns the URL to the model trained using the given config
114
+
115
+ Args:
116
+ config_path (str): config file name relative to detectron2's "configs/"
117
+ directory, e.g., "COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_1x.yaml"
118
+
119
+ Returns:
120
+ str: a URL to the model
121
+ """
122
+ url = _ModelZooUrls.query(config_path)
123
+ if url is None:
124
+ raise RuntimeError("Pretrained model for {} is not available!".format(config_path))
125
+ return url
126
+
127
+
128
+ def get_config_file(config_path):
129
+ """
130
+ Returns path to a builtin config file.
131
+
132
+ Args:
133
+ config_path (str): config file name relative to detectron2's "configs/"
134
+ directory, e.g., "COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_1x.yaml"
135
+
136
+ Returns:
137
+ str: the real path to the config file.
138
+ """
139
+ cfg_file = pkg_resources.resource_filename(
140
+ "detectron2.model_zoo", os.path.join("configs", config_path)
141
+ )
142
+ if not os.path.exists(cfg_file):
143
+ raise RuntimeError("{} not available in Model Zoo!".format(config_path))
144
+ return cfg_file
145
+
146
+
147
+ def get_config(config_path, trained: bool = False):
148
+ """
149
+ Returns a config object for a model in model zoo.
150
+
151
+ Args:
152
+ config_path (str): config file name relative to detectron2's "configs/"
153
+ directory, e.g., "COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_1x.yaml"
154
+ trained (bool): If True, will set ``MODEL.WEIGHTS`` to trained model zoo weights.
155
+ If False, the checkpoint specified in the config file's ``MODEL.WEIGHTS`` is used
156
+ instead; this will typically (though not always) initialize a subset of weights using
157
+ an ImageNet pre-trained model, while randomly initializing the other weights.
158
+
159
+ Returns:
160
+ CfgNode or omegaconf.DictConfig: a config object
161
+ """
162
+ cfg_file = get_config_file(config_path)
163
+ if cfg_file.endswith(".yaml"):
164
+ cfg = get_cfg()
165
+ cfg.merge_from_file(cfg_file)
166
+ if trained:
167
+ cfg.MODEL.WEIGHTS = get_checkpoint_url(config_path)
168
+ return cfg
169
+ elif cfg_file.endswith(".py"):
170
+ cfg = LazyConfig.load(cfg_file)
171
+ if trained:
172
+ url = get_checkpoint_url(config_path)
173
+ if "train" in cfg and "init_checkpoint" in cfg.train:
174
+ cfg.train.init_checkpoint = url
175
+ else:
176
+ raise NotImplementedError
177
+ return cfg
178
+
179
+
180
+ def get(config_path, trained: bool = False, device: Optional[str] = None):
181
+ """
182
+ Get a model specified by relative path under Detectron2's official ``configs/`` directory.
183
+
184
+ Args:
185
+ config_path (str): config file name relative to detectron2's "configs/"
186
+ directory, e.g., "COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_1x.yaml"
187
+ trained (bool): see :func:`get_config`.
188
+ device (str or None): overwrite the device in config, if given.
189
+
190
+ Returns:
191
+ nn.Module: a detectron2 model. Will be in training mode.
192
+
193
+ Example:
194
+ ::
195
+ from annotator.oneformer.detectron2 import model_zoo
196
+ model = model_zoo.get("COCO-InstanceSegmentation/mask_rcnn_R_50_FPN_1x.yaml", trained=True)
197
+ """
198
+ cfg = get_config(config_path, trained)
199
+ if device is None and not torch.cuda.is_available():
200
+ device = "cpu"
201
+ if device is not None and isinstance(cfg, CfgNode):
202
+ cfg.MODEL.DEVICE = device
203
+
204
+ if isinstance(cfg, CfgNode):
205
+ model = build_model(cfg)
206
+ DetectionCheckpointer(model).load(cfg.MODEL.WEIGHTS)
207
+ else:
208
+ model = instantiate(cfg.model)
209
+ if device is not None:
210
+ model = model.to(device)
211
+ if "train" in cfg and "init_checkpoint" in cfg.train:
212
+ DetectionCheckpointer(model).load(cfg.train.init_checkpoint)
213
+ return model