Update README.md
Browse files
README.md
CHANGED
|
@@ -10,20 +10,14 @@ pipeline_tag: image-segmentation
|
|
| 10 |
---
|
| 11 |
# Model Card for Model ID
|
| 12 |
|
| 13 |
-
<!-- Provide a quick summary of what the model is/does. -->
|
| 14 |
-
|
| 15 |
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
|
| 16 |
|
| 17 |
## Model Details
|
| 18 |
|
| 19 |
### Model Description
|
| 20 |
|
| 21 |
-
<!-- Provide a longer summary of what this model is. -->
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
- Developed by:Highsky7
|
| 26 |
-
- Model type:
|
| 27 |
- License:MIT
|
| 28 |
|
| 29 |
### Model Sources [optional]
|
|
@@ -34,25 +28,59 @@ This modelcard aims to be a base template for new models. It has been generated
|
|
| 34 |
|
| 35 |
## Uses
|
| 36 |
|
| 37 |
-
|
| 38 |
|
| 39 |
### Direct Use
|
| 40 |
|
| 41 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 42 |
|
| 43 |
-
[More Information Needed]
|
| 44 |
|
| 45 |
### Downstream Use [optional]
|
| 46 |
|
| 47 |
-
|
| 48 |
|
| 49 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 50 |
|
| 51 |
### Out-of-Scope Use
|
| 52 |
|
| 53 |
-
|
| 54 |
|
| 55 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 56 |
|
| 57 |
## Bias, Risks, and Limitations
|
| 58 |
|
|
|
|
| 10 |
---
|
| 11 |
# Model Card for Model ID
|
| 12 |
|
|
|
|
|
|
|
| 13 |
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
|
| 14 |
|
| 15 |
## Model Details
|
| 16 |
|
| 17 |
### Model Description
|
| 18 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
- Developed by:Highsky7
|
| 20 |
+
- Model type:Image-segmentation
|
| 21 |
- License:MIT
|
| 22 |
|
| 23 |
### Model Sources [optional]
|
|
|
|
| 28 |
|
| 29 |
## Uses
|
| 30 |
|
| 31 |
+
This model is a lane recognition model created for use by Konkuk University Team 2 in The 4th International University Student EV Autonomous Driving Competition (2025-07-11). The accompanying roboflow_final.py code is a ROS-based driving node that utilizes this model to control the vehicle's steering angle in real-time.
|
| 32 |
|
| 33 |
### Direct Use
|
| 34 |
|
| 35 |
+
This model can be used directly via the ultralytics library.
|
| 36 |
+
|
| 37 |
+
First, install the necessary libraries:
|
| 38 |
+
|
| 39 |
+
Bash
|
| 40 |
+
|
| 41 |
+
pip install ultralytics opencv-python
|
| 42 |
+
The following is example code to load the model from Hugging Face and perform lane recognition on a single image.
|
| 43 |
+
|
| 44 |
+
Python
|
| 45 |
+
|
| 46 |
+
from ultralytics import YOLO
|
| 47 |
+
import cv2
|
| 48 |
+
|
| 49 |
+
model = YOLO("YourUsername/YourModelName")
|
| 50 |
+
|
| 51 |
+
image_path = "path/to/your/test_image.jpg"
|
| 52 |
+
frame = cv2.imread(image_path)
|
| 53 |
+
|
| 54 |
+
results = model(frame, conf=0.6)
|
| 55 |
+
|
| 56 |
+
result_plot = results[0].plot()
|
| 57 |
+
|
| 58 |
+
cv2.imshow("Lane Detection Result", result_plot)
|
| 59 |
+
cv2.waitKey(0)
|
| 60 |
+
cv2.destroyAllWindows()
|
| 61 |
|
|
|
|
| 62 |
|
| 63 |
### Downstream Use [optional]
|
| 64 |
|
| 65 |
+
The output of this model (lane masks) can be used as a key input for a larger autonomous driving system. For example, the roboflow_final.py code performs the following downstream tasks:
|
| 66 |
|
| 67 |
+
Generates a drivable center path based on the detected left/right lanes.
|
| 68 |
+
|
| 69 |
+
Calculates a dynamic lookahead point based on the generated path and the vehicle's speed (throttle).
|
| 70 |
+
|
| 71 |
+
Determines the steering angle using the Pure Pursuit algorithm, based on the lookahead point and the vehicle's current position.
|
| 72 |
+
|
| 73 |
+
Interfaces with other modules, such as triggering a forced RRT algorithm when an obstacle (cone) is detected on the driving path.
|
| 74 |
|
| 75 |
### Out-of-Scope Use
|
| 76 |
|
| 77 |
+
This model was designed for a specific purpose and environment. Its use in the following situations may be inappropriate or dangerous:
|
| 78 |
|
| 79 |
+
General Road Driving: Performance is not guaranteed on public roads with different lane markings, lighting, and weather conditions than those of the competition track.
|
| 80 |
+
|
| 81 |
+
Fully Autonomous System: This model only recognizes lanes. It cannot be used to build a complete autonomous system on its own, as it does not detect pedestrians, other vehicles, traffic lights, or signs.
|
| 82 |
+
|
| 83 |
+
Changes in Camera Setup: The model's Bird's-Eye-View (BEV) transformation logic is calibrated for a specific camera position and angle, stored in the bev_params_y_5.npz file. If the camera's mounting position or angle is changed, the coordinate transformation will be inaccurate, severely degrading model performance.
|
| 84 |
|
| 85 |
## Bias, Risks, and Limitations
|
| 86 |
|