mrohith29 commited on
Commit
9a84d93
·
verified ·
1 Parent(s): 0066bfd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -12
README.md CHANGED
@@ -7,6 +7,13 @@ tags:
7
  model-index:
8
  - name: physisolver-gpt2
9
  results: []
 
 
 
 
 
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -14,21 +21,59 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # physisolver-gpt2
16
 
17
- This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an unknown dataset.
18
 
19
  ## Model description
20
 
21
- More information needed
22
 
23
- ## Intended uses & limitations
24
 
25
- More information needed
26
 
27
- ## Training and evaluation data
28
 
29
- More information needed
30
 
31
- ## Training procedure
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
32
 
33
  ### Training hyperparameters
34
 
@@ -44,13 +89,9 @@ The following hyperparameters were used during training:
44
  - num_epochs: 2
45
  - mixed_precision_training: Native AMP
46
 
47
- ### Training results
48
-
49
-
50
-
51
  ### Framework versions
52
 
53
  - Transformers 4.51.1
54
  - Pytorch 2.6.0+cu124
55
  - Datasets 3.5.0
56
- - Tokenizers 0.21.1
 
7
  model-index:
8
  - name: physisolver-gpt2
9
  results: []
10
+ datasets:
11
+ - mrohith29/high-school-physics
12
+ language:
13
+ - en
14
+ metrics:
15
+ - accuracy
16
+ pipeline_tag: text2text-generation
17
  ---
18
 
19
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
21
 
22
  # physisolver-gpt2
23
 
24
+ This model is a fine-tuned version of [gpt2](https://huggingface.co/gpt2) on an [high school physics](https://huggingface.co/datasets/mrohith29/high-school-physics) dataset.
25
 
26
  ## Model description
27
 
28
+ Physisolver-GPT2 is a fine-tuned version of the GPT-2 architecture, optimized for answering physics-related multiple-choice questions. This model has been trained on a custom dataset consisting of physics questions, multiple-choice options, and correct answers. The goal of the model is to generate accurate answers to physics questions by understanding the context and the provided choices.
29
 
30
+ ## Training Data:
31
 
32
+ The model was trained on a dataset of physics questions in a multiple-choice question format. Each question is paired with multiple possible answers, and the correct answer is also included. The dataset is formatted in JSON, where each entry contains a question, a list of choices, and the correct answer.
33
 
34
+ ## Training Objective:
35
 
36
+ The model has been fine-tuned to predict the correct answer given the context of the question and available choices. During training, the model learned to associate questions with their corresponding answers through supervised learning, using the transformer architecture’s language modeling objective.
37
 
38
+ ## Capabilities:
39
+
40
+ - The model can generate answers to a wide variety of physics-related questions.
41
+
42
+ - It understands the question prompt and the choices provided.
43
+
44
+ - It generates the most likely answer based on context and trained knowledge.
45
+
46
+ ## Intended Use: This model is suitable for interactive applications such as:
47
+
48
+ - Physics tutoring systems.
49
+
50
+ - AI-powered physics question-answering platforms.
51
+
52
+ - Integration with educational tools for physics learning.
53
+
54
+ ## Performance:
55
+
56
+ The model performs well on a variety of physics topics included in the dataset. However, since it is trained on a custom dataset, it may not generalize well to entirely different or unseen physics topics outside the scope of the training data.
57
+
58
+ ## Limitations:
59
+
60
+ - The model’s performance is limited to the scope and quality of the dataset it was trained on.
61
+
62
+ - It may struggle with complex or highly specialized physics concepts not covered in the training data.
63
+
64
+ - The model's ability to reason is limited to the patterns it learned from the dataset and may not always provide optimal answers for ambiguous questions.
65
+
66
+ ## Deployment:
67
+
68
+ Physisolver-GPT2 can be accessed and integrated into applications through the Hugging Face Model Hub, where it is available for both research and production use.
69
+
70
+ ## Future Improvements:
71
+
72
+ - Expanding the dataset with more diverse physics topics could improve the model’s generalization.
73
+
74
+ - Fine-tuning with a broader set of questions from different physics domains, such as thermodynamics, electromagnetism, etc.
75
+
76
+ - Adding capabilities for reasoning through multi-step problems or understanding context beyond single-choice questions.
77
 
78
  ### Training hyperparameters
79
 
 
89
  - num_epochs: 2
90
  - mixed_precision_training: Native AMP
91
 
 
 
 
 
92
  ### Framework versions
93
 
94
  - Transformers 4.51.1
95
  - Pytorch 2.6.0+cu124
96
  - Datasets 3.5.0
97
+ - Tokenizers 0.21.1