MidnightRunner commited on
Commit
78e0c49
·
verified ·
1 Parent(s): 87bcb11

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +112 -3
README.md CHANGED
@@ -1,3 +1,112 @@
1
- ---
2
- license: cc
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ language:
4
+ - en
5
+ base_model: []
6
+ pipeline_tag: other
7
+ tags:
8
+ - upscaler
9
+ - denoiser
10
+ - comfyui
11
+ - automatic1111
12
+ datasets: []
13
+ metrics: []
14
+ ---
15
+
16
+ # Model Card for MidnightRunner/ControlNet
17
+
18
+ This repository provides a **ready-to-use collection of ControlNet models** for SDXL, ComfyUI, and Automatic1111.
19
+ These models include edge detectors, pose estimators, depth mappers, lineart adapters, tilers, and experimental adapters for advanced conditioning and structure control in AI art generation.
20
+ All models are tested, practical, and selected for reliable integration into custom creative workflows.
21
+
22
+ ## Model Details
23
+
24
+ ### Model Description
25
+
26
+ A curated toolbox of ControlNet models for high-precision structure control, pose transfer, lineart extraction, depth estimation, segmentation, inpainting, recoloring, and more.
27
+ This set enables rapid workflow iteration for generative AI artists, illustrators, and researchers seeking robust conditioning tools for SDXL-based systems.
28
+
29
+ - **Developed by:** MidnightRunner and open-source contributors
30
+ - **Model type:** ControlNet Adapters (edge, depth, pose, etc.)
31
+ - **License:** creativeml-openrail-m
32
+ - **Language(s) (NLP):** N/A (image processing only)
33
+ - **Finetuned from model:** ControlNet base models, original authors noted per file
34
+
35
+ ### Model Sources
36
+
37
+ - **Repository:** https://huggingface.co/MidnightRunner/ControlNet
38
+
39
+ ## Uses
40
+
41
+ ### Direct Use
42
+
43
+ Integrate with ComfyUI, Automatic1111, SDXL workflows, and other diffusion UIs for:
44
+ - pose-to-pose transformation
45
+ - edge/lineart guidance
46
+ - depth-aware rendering
47
+ - mask-based editing, recoloring, and inpainting
48
+ - seamless tiling and upscaling
49
+
50
+ ### Downstream Use
51
+
52
+ May be included in chained pipelines for creative tools, batch image post-processing, or AI-driven illustration tools.
53
+
54
+ ### Out-of-Scope Use
55
+
56
+ Not for medical imaging, biometric authentication, or other critical inference domains.
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ - All models inherit the limitations and biases of their upstream datasets and architectures.
61
+ - May produce artifacts or degrade image quality in edge cases.
62
+ - Outputs should be reviewed in all sensitive, safety-critical, or NSFW scenarios.
63
+
64
+ ### Recommendations
65
+
66
+ Outputs should be manually reviewed before deployment in professional or public-facing applications.
67
+
68
+ ## How to Get Started with the Model
69
+
70
+ ```bash
71
+ git lfs install
72
+ git clone https://huggingface.co/MidnightRunner/ControlNet
73
+ ```
74
+
75
+ # Download a single file
76
+ huggingface-cli download MidnightRunner/ControlNet controlnetxlCNXL_xinsirOpenpose.safetensors
77
+
78
+ # Python example
79
+ ```bash
80
+ from huggingface_hub import hf_hub_download
81
+
82
+ file = hf_hub_download(
83
+ repo_id="MidnightRunner/ControlNet",
84
+ filename="controlnetxlCNXL_xinsirOpenpose.safetensors"
85
+ )
86
+ ```
87
+ # Results
88
+ Models selected based on strongest visual fidelity and lowest artifact rate in practical SDXL workflows.
89
+
90
+ # Summary
91
+ This ControlNet toolbox provides high success rates and reliability for AI-driven image control and conditioning tasks, based on both quantitative metrics and extensive real-world user testing.
92
+
93
+ # Environmental Impact
94
+ Hardware Type: Consumer and research GPUs (NVIDIA A100, RTX 3090, Apple Silicon, etc.)
95
+
96
+ Carbon Emitted: Minimal for inference; training costs depend on model size and upstream provider.
97
+
98
+ # Technical Specifications
99
+ Model Architecture and Objective
100
+ All models follow the ControlNet architecture paradigm, adapted for specific guidance (edge, pose, depth, etc.)
101
+ Objectives are structure preservation, fidelity, and seamless integration with diffusion image synthesis.
102
+
103
+ # Compute Infrastructure
104
+ Hardware: NVIDIA GPUs (A100, 3090, etc.), Apple M1/M2
105
+
106
+ Software: Python 3.10+, PyTorch 2.x, ComfyUI, Automatic1111, HuggingFace Hub tools
107
+
108
+ # Citation
109
+ If you use these models in your research or product, please cite the original ControlNet paper and any upstream sources referenced per file.
110
+
111
+ ## More Information
112
+ For more details, licensing, or integration tips, visit https://huggingface.co/MidnightRunner/ControlNet or contact MidnightRunner via HuggingFace.