Shreyboiii commited on
Commit
23c5e7b
·
verified ·
1 Parent(s): 764f607

End of training

Browse files
README.md CHANGED
@@ -1,49 +1,60 @@
1
  ---
2
- license: mit
 
 
 
 
 
 
 
3
  ---
4
- # Question Answering Chatbot using Hugging Face Transformers
5
 
6
- This project fine-tunes a question answering model using the Hugging Face Transformers library and SQuAD dataset.
7
- It also includes a simple chat wrapper that allows users to ask questions interactively based on a given context.
8
 
9
- ---
10
 
11
- ## 📚 Project Overview
 
 
12
 
13
- - Fine-tunes a pretrained DistilBERT model (`distilbert-base-cased-distilled-squad`) on the SQuAD dataset.
14
- - Implements a chat interface where users can ask free-form questions related to a specific context.
15
- - Publishes the fine-tuned model to Hugging Face Hub.
16
 
17
- ---
18
 
19
- ## 🚀 Notebook Workflow
20
 
21
- 1. **Setup**
22
- Install required libraries and disable TensorFlow backend to ensure PyTorch is used.
23
 
24
- 2. **Dataset Loading and Preprocessing**
25
- Load the SQuAD dataset, tokenize the questions and contexts, and prepare input tensors.
26
 
27
- 3. **Model Fine-tuning**
28
- Fine-tune the pretrained model for 3 epochs using the Hugging Face `Trainer` API.
29
 
30
- 4. **Push to Hugging Face Hub**
31
- Upload the trained model and tokenizer to your Hugging Face profile.
32
 
33
- 5. **Chatbot Interface**
34
- Implement a simple wrapper allowing users to ask questions about a given context.
35
 
36
- ---
 
 
 
 
 
 
 
 
 
37
 
38
- ## 🛠️ Requirements
 
 
 
 
39
 
40
- - Python 3.8+
41
- - `transformers`
42
- - `datasets`
43
- - `torch`
44
- - `wandb` (optional, for experiment tracking)
45
 
46
- Install requirements in Colab or locally:
47
 
48
- ```bash
49
- pip install transformers datasets torch wandb
 
 
 
1
  ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ base_model: distilbert-base-uncased
5
+ tags:
6
+ - generated_from_trainer
7
+ model-index:
8
+ - name: my_awesome_qa_model
9
+ results: []
10
  ---
 
11
 
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # my_awesome_qa_model
16
 
17
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 1.8753
20
 
21
+ ## Model description
 
 
22
 
23
+ More information needed
24
 
25
+ ## Intended uses & limitations
26
 
27
+ More information needed
 
28
 
29
+ ## Training and evaluation data
 
30
 
31
+ More information needed
 
32
 
33
+ ## Training procedure
 
34
 
35
+ ### Training hyperparameters
 
36
 
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 2e-05
39
+ - train_batch_size: 16
40
+ - eval_batch_size: 16
41
+ - seed: 42
42
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
+ - lr_scheduler_type: linear
44
+ - num_epochs: 3
45
+
46
+ ### Training results
47
 
48
+ | Training Loss | Epoch | Step | Validation Loss |
49
+ |:-------------:|:-----:|:----:|:---------------:|
50
+ | No log | 1.0 | 250 | 2.3563 |
51
+ | 2.7737 | 2.0 | 500 | 1.9792 |
52
+ | 2.7737 | 3.0 | 750 | 1.8753 |
53
 
 
 
 
 
 
54
 
55
+ ### Framework versions
56
 
57
+ - Transformers 4.51.3
58
+ - Pytorch 2.6.0+cu124
59
+ - Datasets 3.6.0
60
+ - Tokenizers 0.21.1
runs/May12_14-32-41_29d1e8ae1181/events.out.tfevents.1747060365.29d1e8ae1181.477.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8f84662b0b969f9103076e6b0e48684cca092193849e9646ad3b948d85c9896a
3
- size 5610
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f83e1e5ec089e00e0fce413a47667ade733fd4405336c00b237539c3a88479d
3
+ size 6235