jatinmehra commited on
Commit
4a34ca9
·
verified ·
1 Parent(s): f1fae65

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +130 -7
README.md CHANGED
@@ -1,11 +1,134 @@
1
  ---
2
- license: apache-2.0
3
  language:
4
  - en
5
- base_model:
6
- - Qwen/Qwen3-8B
7
- pipeline_tag: text-classification
8
- library_name: transformers
9
  tags:
10
- - Math Misconception Annotation
11
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
2
  language:
3
  - en
4
+ license: apache-2.0
5
+ base_model: Qwen/Qwen3-8B
 
 
6
  tags:
7
+ - text-classification
8
+ - education
9
+ - math
10
+ - misconception-detection
11
+ - student-learning
12
+ metrics:
13
+ - map@3
14
+ model-index:
15
+ - name: Qwen3-8B-Math-Misconception-Classifier
16
+ results:
17
+ - task:
18
+ type: text-classification
19
+ name: Math Misconception Classification
20
+ metrics:
21
+ - type: map@3
22
+ value: 0.944
23
+ name: Mean Average Precision at 3
24
+ ---
25
+
26
+ # Qwen3-8B Math Misconception Classifier
27
+
28
+ ## Model Description
29
+
30
+ This model is a fine-tuned version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B) for identifying and classifying student mathematical misconceptions. The model analyzes student explanations of math problems and predicts the specific misconception category they exhibit.
31
+
32
+ **Model Architecture:** Qwen3-8B (8 billion parameters)
33
+ **Task:** Multi-class Text Classification (65 misconception classes)
34
+ **Performance:** MAP@3 Score of 0.944
35
+
36
+ ## Intended Use
37
+
38
+ ### Primary Use Cases
39
+ - Identifying mathematical misconceptions from student explanations
40
+ - Educational assessment and personalized learning
41
+ - Automated feedback systems for math education
42
+ - Research in mathematics education
43
+
44
+ ### Out-of-Scope Use
45
+ - General text classification tasks outside of math education
46
+ - Real-time production systems without human oversight
47
+ - Any application where misclassification could lead to harm
48
+
49
+ ## Training Details
50
+
51
+ ### Training Data
52
+ The model was trained on the MAP Charting Student Math Misunderstandings dataset, which includes:
53
+ - Mathematical questions with multiple choice answers
54
+ - Student explanations for their answer choices
55
+ - Labels indicating whether the answer was correct
56
+ - Misconception categories and specific misconceptions
57
+
58
+ ### Training Procedure
59
+
60
+ **Input Format:**
61
+ ```
62
+ Question: {QuestionText}
63
+ Answer: {MC_Answer}
64
+ Is Correct Answer: {Yes/No}
65
+ Student Explanation: {StudentExplanation}
66
+ ```
67
+
68
+ This structure provides the model with full context: the question, the student's answer choice, whether it's correct, and their reasoning.
69
+
70
+ **Preprocessing Steps:**
71
+ 1. Created target labels by combining `Category` and `Misconception` columns
72
+ 2. Transformed labels into numerical format using label encoding
73
+ 3. Identified correct answers and merged this information into the training data
74
+
75
+ **Training Configuration:**
76
+ - **Model:** Qwen 3 8B
77
+ - **Method:** Full Fine-tuning
78
+ - **Learning Rate:** 2e-5
79
+ - **Epochs:** 3
80
+ - **Batch Size:** 16
81
+ - **Precision:** Mixed precision (FP16/BF16)
82
+
83
+ ## Model Evaluation
84
+
85
+ The model was evaluated using the MAP@3 metric on the validation set from the competition, achieving a score of 0.944.
86
+
87
+ **Evaluation Procedure:**
88
+ - Predictions were generated for the validation set
89
+ - MAP@3 score was calculated based on the competition's evaluation script
90
+
91
+ ## Limitations & Bias
92
+
93
+ - The model is specifically tuned for the MAP competition dataset and may not generalize to other text classification tasks.
94
+ - There may be biases present in the training data that could affect the model's predictions.
95
+ - Misclassifications could occur, especially in cases of ambiguous or unclear student explanations.
96
+
97
+ ## Acknowledgments
98
+
99
+ This model was developed as part of the MAP (Misconception Annotation Project) competition on Kaggle. Special thanks to the competition hosts and the Kaggle community for their support and collaboration.
100
+
101
+ ## How to Use This Model
102
+
103
+ To use this model for predicting math misconceptions:
104
+
105
+ 1. Install the required libraries:
106
+ ```bash
107
+ pip install transformers torch
108
+ ```
109
+
110
+ 2. Load the model and tokenizer:
111
+ ```python
112
+ from transformers import AutoModelForSequenceClassification, AutoTokenizer
113
+
114
+ model_name = "Qwen3-8B-Math-Misconception-Classifier"
115
+ model = AutoModelForSequenceClassification.from_pretrained(model_name)
116
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
117
+ ```
118
+
119
+ 3. Prepare your input data in the required format.
120
+
121
+ 4. Make predictions:
122
+ ```python
123
+ from transformers import pipeline
124
+
125
+ classifier = pipeline("text-classification", model=model, tokenizer=tokenizer)
126
+ results = classifier(your_input_data)
127
+ ```
128
+
129
+ 5. Interprete the results, which will include the predicted misconception categories and their associated probabilities.
130
+
131
+ ## References
132
+
133
+ - [Hugging Face Transformers Documentation](https://huggingface.co/docs/transformers/index)
134
+ - [Kaggle MAP Competition](https://www.kaggle.com/competitions/map-charting-student-math-misunderstandings)