Llama_for_Finance / model_info.json
TimberGu's picture
Upload model_info.json with huggingface_hub
4d951b3 verified
{
"Project Name": "Financial LLaMA Fine-tuning",
"Base Model": "meta-llama/Meta-Llama-3.1-8B-Instruct",
"Training Dataset": "Josephgflowers/Finance-Instruct-500k",
"Fine-tuning Method": "LoRA (Low-Rank Adaptation)",
"Save Time": "2025-08-08 04:51:47",
"File List": [
"README.md",
"adapter_model.safetensors",
"adapter_config.json",
"training_args.bin",
"chat_template.jinja",
"tokenizer_config.json",
"special_tokens_map.json",
"tokenizer.json",
"training_config.json",
"test_results.json"
],
"Local Save Path": "C:\\Users\\Timber's Pad\\OneDrive\\Desktop\\JobHunting\\Project2_FineTune\\Project2_FineTune\\FineTuneSave",
"File Description": {
"adapter_config.json": "LoRA configuration file",
"adapter_model.safetensors": "LoRA weight file",
"tokenizer.json": "Tokenizer file",
"tokenizer_config.json": "Tokenizer configuration",
"special_tokens_map.json": "Special token mapping"
},
"Usage Instructions": [
"1. Extract zip file to target folder",
"2. Use the following code to load the model:",
" from peft import PeftModel",
" from transformers import AutoModelForCausalLM, AutoTokenizer",
" base_model = AutoModelForCausalLM.from_pretrained('meta-llama/Meta-Llama-3.1-8B-Instruct')",
" model = PeftModel.from_pretrained(base_model, 'path/to/model')",
" tokenizer = AutoTokenizer.from_pretrained('path/to/model')"
]
}