Update README.md
#3
by
suphanatwong
- opened
README.md
CHANGED
|
@@ -15,7 +15,6 @@ pipeline_tag: table-question-answering
|
|
| 15 |
|
| 16 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 17 |
|
| 18 |
-
This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1).
|
| 19 |
|
| 20 |
## Model Details
|
| 21 |
|
|
@@ -26,12 +25,10 @@ This modelcard aims to be a base template for new models. It has been generated
|
|
| 26 |
|
| 27 |
|
| 28 |
- **Developed by:** The Scamper
|
| 29 |
-
- **Funded by [optional]:** [More Information Needed]
|
| 30 |
-
- **Shared by [optional]:** [More Information Needed]
|
| 31 |
- **Model type:** Transformer
|
| 32 |
- **Language(s) (NLP):** Thai, English
|
| 33 |
-
- **License:**
|
| 34 |
-
- **Finetuned from model
|
| 35 |
|
| 36 |
|
| 37 |
## Uses
|
|
@@ -70,13 +67,22 @@ Use the code below to get started with the model.
|
|
| 70 |
|
| 71 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 72 |
|
| 73 |
-
|
| 74 |
|
| 75 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 76 |
|
|
|
|
| 77 |
|
| 78 |
-
|
| 79 |
|
| 80 |
-
|
| 81 |
|
| 82 |
|
|
|
|
| 15 |
|
| 16 |
<!-- Provide a quick summary of what the model is/does. -->
|
| 17 |
|
|
|
|
| 18 |
|
| 19 |
## Model Details
|
| 20 |
|
|
|
|
| 25 |
|
| 26 |
|
| 27 |
- **Developed by:** The Scamper
|
|
|
|
|
|
|
| 28 |
- **Model type:** Transformer
|
| 29 |
- **Language(s) (NLP):** Thai, English
|
| 30 |
+
- **License:** apache-2.0
|
| 31 |
+
- **Finetuned from model:** OpenThaiGPT-1.0.0 70B (https://huggingface.co/openthaigpt/openthaigpt-1.0.0-70b-chat)
|
| 32 |
|
| 33 |
|
| 34 |
## Uses
|
|
|
|
| 67 |
|
| 68 |
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 69 |
|
| 70 |
+
The methodology for fine-tuning involves a dataset with two columns: "question" and "SQL syntax". Here's a brief outline of the process:
|
| 71 |
|
| 72 |
+
1. **Data Collection**: Gather a dataset containing pairs of questions and their corresponding SQL queries. Ensure the questions cover various topics and query types, while the SQL queries represent the desired actions on a database.
|
| 73 |
+
|
| 74 |
+
2. **Pre-processing**: Clean and preprocess the data to remove noise, standardize formatting, and handle any inconsistencies. Tokenize the text and encode it into a format suitable for training.
|
| 75 |
+
|
| 76 |
+
3. **Model Architecture**: Utilize OpenThaiGPT 1.0.0 70B as the base model.
|
| 77 |
+
|
| 78 |
+
4. **Fine-tuning Setup**: Divide the dataset into training (90%) and test sets (10%). We define the training procedure, including hyperparameters such as learning rate, batch size, and number of training epochs.
|
| 79 |
+
|
| 80 |
+
5. **Fine-tuning Process**: Train the model on the question-SQL pairs using the defined setup. During training, the model learns to predict the SQL query corresponding to a given question by minimizing a suitable loss function.
|
| 81 |
|
| 82 |
+
6. **Testing**: Evaluate the final model on a held-out test set to assess its generalization performance on unseen data.
|
| 83 |
|
| 84 |
+
7. **Deployment**: Deploy the fine-tuned model for text-to-SQL tasks in real-world applications, where it can generate SQL queries from natural language questions effectively and efficiently.
|
| 85 |
|
| 86 |
+
By following this methodology, the model can be fine-tuned to accurately convert natural language questions into SQL syntax, enabling seamless interaction with structured databases.
|
| 87 |
|
| 88 |
|