File size: 2,642 Bytes
5f45186 90062ca 5f45186 71d1014 5f45186 cdee2a1 90062ca 5f45186 90062ca 5f45186 90062ca 5f45186 90062ca 5f45186 90062ca 5f45186 90062ca 5f45186 71d1014 5f45186 71d1014 5f45186 90062ca | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 | ---
library_name: transformers
license: apache-2.0
base_model: distilbert-base-uncased
tags:
- generated_from_trainer
model-index:
- name: help_classifier
results: []
datasets:
- King-8/help-request-messages
---
# help_classifier
This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the "King-8/help-request-messages" dataset.
It achieves the following results on the evaluation set:
- Loss: 1.3083
---
## π€ CIC Help Classifier Model
### Overview
This model is a fine-tuned text classification model designed to identify the type of help a user needs within the Coding in Color (CIC) ecosystem.
It enables AI systems to understand user challenges and provide structured support.
---
### π§ Model Details
* Base model: `distilbert-base-uncased`
* Task: Text classification
* Training data: CIC Help Classification Dataset
* Framework: Hugging Face Transformers
---
### π Labels
* learning_help
* project_help
* attendance_issue
* technical_issue
* general_guidance
---
### βοΈ Training
* Epochs: 3
* Dataset size: 100 samples
* Train/Validation/Test split used
---
### π Performance Notes
* Training and validation loss decreased across epochs
* Model performs well on common help scenarios
* Accuracy is limited due to small dataset size
---
### π§ͺ Example Usage
```python
predict("I'm stuck on my project and don't know what to do")
```
Output:
```json
{
"type": "project_help",
"confidence": 0.82
}
```
---
### π Use Case
This model is designed to be integrated into:
* MCP server tools
* Slack-based support systems
* AI assistants for CIC students
---
### π Future Improvements
* Fine-tune on larger CIC dataset
* Add real-time feedback learning
* Integrate with response generation models
* Improve classification accuracy with more edge cases
---
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: linear
- num_epochs: 3
---
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| 1.3887 | 1.0 | 9 | 1.4495 |
| 1.2613 | 2.0 | 18 | 1.3350 |
| 1.1704 | 3.0 | 27 | 1.3083 |
---
### Framework versions
- Transformers 5.0.0
- Pytorch 2.10.0+cpu
- Datasets 4.0.0
- Tokenizers 0.22.2 |