omersaidd commited on
Commit
286f38c
·
verified ·
1 Parent(s): de06b8c

README.md

Browse files

The results of the Fine-Tuning process to improve the performance of the Mistral Nemo 12b parameterised LLM for the Turkish domain are as follows. I would like to thank those who contributed to the preparation of the Turkish data set.

🤖 Mistral Nemo 12b (Open-Source LLM)
🦥 Unsloth (2x faster with 60% less memory)
📄 Dataset (myzens/alpaca-turkish-combined)

Prompt : What is LLM and what are its application areas?

Response : Large language models, also known as LLM, large language model (LLM), are artificial intelligence models that are trained with large amounts of data and can be used for natural language processing (NLP) tasks. LLMs can be used to understand and produce texts written in natural language. Their applications include text classification, natural language processing, text summarisation, text generation and text translation.
Training hyperparameters
The following hyperparameters were used during training:


learning_rate: 2e-4
per_device_train_batch_size = 2
warmup_steps = 5
seed: 3407
optimizer: adamw_8bit
lr_scheduler_type: linear
dataset_num_proc = 2

![mistral.jpeg](https://cdn-uploads.huggingface.co/production/uploads/65327bbe910b844786524e3c/fM_rtFOp44HwxQjAsa4MV.jpeg)

Files changed (1) hide show
  1. README.md +1 -0
README.md CHANGED
@@ -8,4 +8,5 @@ pipeline_tag: text2text-generation
8
  tags:
9
  - sft
10
  - lora
 
11
  ---
 
8
  tags:
9
  - sft
10
  - lora
11
+ - generated_from_trainer
12
  ---