namantjeaswi commited on
Commit
e132e5c
·
1 Parent(s): 01b8683

using llama 3

Browse files
Files changed (1) hide show
  1. app.py +6 -2
app.py CHANGED
@@ -2,8 +2,12 @@ import streamlit as st
2
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
 
4
  # Load the tokenizer and model
5
- tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")
6
- model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1")
 
 
 
 
7
 
8
  # Create text generation pipeline
9
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)
 
2
  from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
3
 
4
  # Load the tokenizer and model
5
+ #tokenizer = AutoTokenizer.from_pretrained("mistralai/Mistral-7B-v0.1")#
6
+ #model = AutoModelForCausalLM.from_pretrained("mistralai/Mistral-7B-v0.1")
7
+
8
+
9
+ tokenizer = AutoTokenizer.from_pretrained("meta-llama/Meta-Llama-3-8B")
10
+ model = AutoModelForCausalLM.from_pretrained("meta-llama/Meta-Llama-3-8B")
11
 
12
  # Create text generation pipeline
13
  pipe = pipeline("text-generation", model=model, tokenizer=tokenizer)