Spaces:
Running
Running
Update README.md
Browse files
README.md
CHANGED
|
@@ -5,6 +5,8 @@ sdk: static
|
|
| 5 |
## Evaluation Pipeline
|
| 6 |
# Use eval_pipeline.py or the raw version of the code below to evaluate the model. Make sure to set the dataset and model path.
|
| 7 |
|
|
|
|
|
|
|
| 8 |
import pandas as pd
|
| 9 |
from sklearn.model_selection import train_test_split
|
| 10 |
from sklearn.metrics import accuracy_score, classification_report
|
|
@@ -160,3 +162,5 @@ y_true, y_pred = custom_model.evaluate(test_loader)
|
|
| 160 |
|
| 161 |
print(f"Accuracy: {accuracy_score(y_true, y_pred):.4f}")
|
| 162 |
print("Classification Report:\n", classification_report(y_true, y_pred))
|
|
|
|
|
|
|
|
|
| 5 |
## Evaluation Pipeline
|
| 6 |
# Use eval_pipeline.py or the raw version of the code below to evaluate the model. Make sure to set the dataset and model path.
|
| 7 |
|
| 8 |
+
```
|
| 9 |
+
|
| 10 |
import pandas as pd
|
| 11 |
from sklearn.model_selection import train_test_split
|
| 12 |
from sklearn.metrics import accuracy_score, classification_report
|
|
|
|
| 162 |
|
| 163 |
print(f"Accuracy: {accuracy_score(y_true, y_pred):.4f}")
|
| 164 |
print("Classification Report:\n", classification_report(y_true, y_pred))
|
| 165 |
+
|
| 166 |
+
```
|