Spaces:
Runtime error
Runtime error
| data_downloading: | |
| download_directory: "quant_bio_retrieval/" # directory where the papers will be downloaded and the graph will be saved | |
| gexf_file: "test_graph.gexf" # name of the graph file that will be created only if downloading option is true | |
| processing: | |
| random_seed: 10 | |
| keep_unstructured_content: false # keep unstructured content of the papers as graph node attribute if true | |
| arxiv_rate_limit: 3 # time in seconds to wait between each arxiv api call to avoid ban | |
| retriever: | |
| embedder: BAAI/bge-large-en-v1.5 | |
| num_retrievals: 30000 | |
| load_arxiv_embeds: True # load arxiv embeddings from huggingface if true else generate them | |
| inference: | |
| base_model: meta-llama/Meta-Llama-3-8B | |
| pretrained_model: "models/Robotics/Meta-LLama-3-8B-Quantative-Robotics" # used only if training option is false | |
| generation_args: | |
| max_new_tokens: 1000 | |
| do_sample: True | |
| top_p: 0.9 | |
| top_k: 50 | |
| temperature: 0.7 | |
| no_repeat_ngram_size: 2 | |
| num_beams: 1 | |
| gen_related_work_instruct_model: meta-llama/Llama-3.1-8B-Instruct # Model assisting at the generation of related work instructions | |
| training: | |
| predefined_graph_path: "robotics.gexf" # path to the graph dataset used for fine-tuning only if downloading option is false | |
| trainer_args: | |
| per_device_train_batch_size: 4 | |
| warmup_steps: 100 | |
| num_train_epochs: 1 | |
| learning_rate: 0.0002 | |
| lr_scheduler_type: 'cosine' | |
| fp16: true | |
| logging_steps: 1 | |
| save_steps: 50 | |
| trainer_output_dir: trainer_outputs/ | |
| tokenizer: | |
| max_length: 1024 | |
| qlora: | |
| rank: 8 | |
| lora_alpha: 32 | |
| lora_dropout: 0.05 | |
| target_modules: # modules for which to train lora adapters | |
| - q_proj | |
| - k_proj | |
| - v_proj | |
| - o_proj | |
| # Used only if training option is true to save and load the fine-tuned model | |
| model_saving: | |
| model_name: llama_1b_qlora_uncensored | |
| model_output_dir: models # model saved in {model_output_dir}/{model_name}_{index} # model saved in {model_output_dir}/{model_name}_{index} after fine-tuning completion | |
| index: 1 | |