techpro-saida commited on
Commit
b97cb65
·
verified ·
1 Parent(s): 5ba9c14

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -11
README.md CHANGED
@@ -63,19 +63,15 @@ The model was fine-tuned on a custom dataset (`data.jsonl`) consisting of:
63
  ## Example Usage
64
 
65
  ```python
66
- from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
67
- from peft import PeftModel
68
 
69
- base_model = "mistralai/Mistral-7B-Instruct-v0.3"
70
- repo = "techpro-saida/msci_software_engineering_slm_v1"
71
 
72
- model = AutoModelForCausalLM.from_pretrained(base_model)
73
- model = PeftModel.from_pretrained(model, repo)
74
- tokenizer = AutoTokenizer.from_pretrained(base_model)
75
-
76
- slm = pipeline("text-generation", model=model, tokenizer=tokenizer)
77
- result = slm("Explain SOLID principles in OOP.", max_new_tokens=80)
78
- print(result[0]["generated_text"])
79
 
80
 
81
 
 
63
  ## Example Usage
64
 
65
  ```python
66
+ from transformers import AutoModelForCausalLM, AutoTokenizer
 
67
 
68
+ tokenizer = AutoTokenizer.from_pretrained("techpro-saida/banking-slm-v1")
69
+ model = AutoModelForCausalLM.from_pretrained("techpro-saida/banking-slm-v1")
70
 
71
+ prompt = "Explain SOLID principles in OOP?"
72
+ inputs = tokenizer(prompt, return_tensors="pt")
73
+ outputs = model.generate(**inputs, max_new_tokens=100, temperature=0.7)
74
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
 
 
 
75
 
76
 
77