jaeyeol816 commited on
Commit
8c7ae6c
verified
1 Parent(s): c4e29e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -99
README.md CHANGED
@@ -11,8 +11,6 @@ pipeline_tag: question-answering
11
 
12
  <!-- Provide a quick summary of what the model is/does. -->
13
 
14
- This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
15
-
16
  ## Model Details
17
 
18
  ### Model Description
@@ -22,11 +20,8 @@ This modelcard aims to be a base template for new models. It has been generated
22
 
23
 
24
  - **Developed by:** Jaeyeol Choi, Yuchan Jung
25
- - **Funded by [optional]:** [More Information Needed]
26
- - **Shared by [optional]:** [More Information Needed]
27
- - **Model type:** [More Information Needed]
28
  - **Language(s) (NLP):** English
29
- - **License:** [More Information Needed]
30
  - **Finetuned from model [optional]:** https://huggingface.co/google/gemma-2-2b
31
 
32
  ### Model Sources [optional]
@@ -34,8 +29,6 @@ This modelcard aims to be a base template for new models. It has been generated
34
  <!-- Provide the basic links for the model. -->
35
 
36
  - **Repository:** https://github.com/YuchanJung/AI-Instructor
37
- - **Paper [optional]:** [More Information Needed]
38
- - **Demo [optional]:** [More Information Needed]
39
 
40
  ## Uses
41
 
@@ -83,10 +76,12 @@ Use the code below to get started with the model.
83
 
84
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
85
 
86
- [More Information Needed]
87
 
88
  ### Training Procedure
89
 
 
 
90
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
91
 
92
  #### Preprocessing [optional]
@@ -96,107 +91,20 @@ Use the code below to get started with the model.
96
 
97
  #### Training Hyperparameters
98
 
99
- - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
100
-
101
- #### Speeds, Sizes, Times [optional]
102
-
103
- <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
104
-
105
- [More Information Needed]
106
-
107
- ## Evaluation
108
 
109
- <!-- This section describes the evaluation protocols and provides the results. -->
110
 
111
- ### Testing Data, Factors & Metrics
112
 
113
- #### Testing Data
114
-
115
- <!-- This should link to a Dataset Card if possible. -->
116
-
117
- [More Information Needed]
118
-
119
- #### Factors
120
-
121
- <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
122
-
123
- [More Information Needed]
124
-
125
- #### Metrics
126
-
127
- <!-- These are the evaluation metrics being used, ideally with a description of why. -->
128
-
129
- [More Information Needed]
130
 
131
  ### Results
132
 
133
- [More Information Needed]
134
 
135
  #### Summary
 
136
 
137
 
138
 
139
- ## Model Examination [optional]
140
-
141
- <!-- Relevant interpretability work for the model goes here -->
142
-
143
- [More Information Needed]
144
-
145
- ## Environmental Impact
146
-
147
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
148
-
149
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
150
-
151
- - **Hardware Type:** [More Information Needed]
152
- - **Hours used:** [More Information Needed]
153
- - **Cloud Provider:** [More Information Needed]
154
- - **Compute Region:** [More Information Needed]
155
- - **Carbon Emitted:** [More Information Needed]
156
-
157
- ## Technical Specifications [optional]
158
-
159
- ### Model Architecture and Objective
160
-
161
- [More Information Needed]
162
-
163
- ### Compute Infrastructure
164
-
165
- [More Information Needed]
166
-
167
- #### Hardware
168
-
169
- [More Information Needed]
170
-
171
- #### Software
172
-
173
- [More Information Needed]
174
-
175
- ## Citation [optional]
176
-
177
- <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
178
-
179
- **BibTeX:**
180
-
181
- [More Information Needed]
182
-
183
- **APA:**
184
-
185
- [More Information Needed]
186
-
187
- ## Glossary [optional]
188
-
189
- <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
190
-
191
- [More Information Needed]
192
-
193
- ## More Information [optional]
194
-
195
- [More Information Needed]
196
-
197
- ## Model Card Authors [optional]
198
-
199
- [More Information Needed]
200
 
201
  ## Model Card Contact
202
 
 
11
 
12
  <!-- Provide a quick summary of what the model is/does. -->
13
 
 
 
14
  ## Model Details
15
 
16
  ### Model Description
 
20
 
21
 
22
  - **Developed by:** Jaeyeol Choi, Yuchan Jung
23
+ - **Model type:** LLM
 
 
24
  - **Language(s) (NLP):** English
 
25
  - **Finetuned from model [optional]:** https://huggingface.co/google/gemma-2-2b
26
 
27
  ### Model Sources [optional]
 
29
  <!-- Provide the basic links for the model. -->
30
 
31
  - **Repository:** https://github.com/YuchanJung/AI-Instructor
 
 
32
 
33
  ## Uses
34
 
 
76
 
77
  <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
78
 
79
+ https://huggingface.co/datasets/jaeyeol816/ai_lecture
80
 
81
  ### Training Procedure
82
 
83
+ https://github.com/YuchanJung/AI-Instructor?tab=readme-ov-file#model-training
84
+
85
  <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
86
 
87
  #### Preprocessing [optional]
 
91
 
92
  #### Training Hyperparameters
93
 
94
+ - **Training regime:** Basic <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
 
 
 
 
 
 
 
 
95
 
 
96
 
 
97
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
98
 
99
  ### Results
100
 
101
+ https://github.com/YuchanJung/AI-Instructor?tab=readme-ov-file#results
102
 
103
  #### Summary
104
+ This project focuses on building an AI Instructor, a Q&A bot, using transcripts from the Andrew Ng's Deep Learning course. It was created specifically for the juniors of the Google ML Bootcamp to provide them with an interactive tool to deepen their understanding of key machine learning concepts. The provided model was created by fine-tuning the Gemma-2B model on a custom-generated Q&A dataset derived from the lecture content.
105
 
106
 
107
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
108
 
109
  ## Model Card Contact
110