Prajjwalng commited on
Commit
efc351e
·
verified ·
1 Parent(s): b8df990

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +28 -10
app.py CHANGED
@@ -1,31 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  # Initialize model and tokenizer (load only once)
2
  @st.cache_resource
3
  def load_model():
4
- tokenizer = AutoTokenizer.from_pretrained("microsoft/DialoGPT-medium")
5
- model = AutoModelForCausalLM.from_pretrained("microsoft/DialoGPT-medium")
 
6
  return tokenizer, model
7
 
8
  tokenizer, model = load_model()
9
 
10
  # Function to generate chatbot response
11
  def generate_response(prompt, chat_history=""):
12
- inputs = tokenizer.encode(chat_history + prompt + tokenizer.eos_token, return_tensors="pt")
13
 
14
- # generate a response while limiting the total chat history to 1000 tokens,
15
- chat_history_ids = model.generate(
16
- inputs, max_length=1000,
 
17
  pad_token_id=tokenizer.eos_token_id,
18
- no_repeat_ngram_size=3,
19
  temperature=0.7,
20
  top_k=50,
21
  top_p=0.95,
22
  )
23
 
24
- response = tokenizer.decode(chat_history_ids[:, inputs.shape[-1]:][0], skip_special_tokens=True)
25
  return response
26
 
27
  # Streamlit app
28
- st.title("Simple Chatbot")
29
 
30
  # Initialize chat history
31
  if "messages" not in st.session_state:
@@ -64,4 +82,4 @@ if prompt := st.chat_input("What is up?"):
64
  # Add assistant response to chat history
65
  st.session_state.messages.append({"role": "assistant", "content": full_response})
66
  #update the chat history.
67
- st.session_state.chat_history += prompt + tokenizer.eos_token + response + tokenizer.eos_token
 
1
+ import streamlit as st
2
+ from transformers import AutoTokenizer, AutoModelForCausalLM
3
+ import torch
4
+ import os
5
+ from huggingface_hub import login
6
+
7
+ # Login with HF_TOKEN (if available)
8
+ hf_token = os.environ.get("HF_TOKEN")
9
+ if hf_token:
10
+ try:
11
+ login(token=hf_token, add_to_git_credential=False)
12
+ st.success("Hugging Face login successful!")
13
+ except Exception as e:
14
+ st.error(f"Hugging Face login failed: {e}")
15
+ else:
16
+ st.warning("HF_TOKEN environment variable not set. Some features may be limited.")
17
+
18
  # Initialize model and tokenizer (load only once)
19
  @st.cache_resource
20
  def load_model():
21
+ model_name = "google/gemma-2b-it"
22
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
23
+ model = AutoModelForCausalLM.from_pretrained(model_name)
24
  return tokenizer, model
25
 
26
  tokenizer, model = load_model()
27
 
28
  # Function to generate chatbot response
29
  def generate_response(prompt, chat_history=""):
30
+ inputs = tokenizer.encode(chat_history + prompt, return_tensors="pt")
31
 
32
+ # Generate a response
33
+ outputs = model.generate(
34
+ inputs,
35
+ max_length=1000,
36
  pad_token_id=tokenizer.eos_token_id,
 
37
  temperature=0.7,
38
  top_k=50,
39
  top_p=0.95,
40
  )
41
 
42
+ response = tokenizer.decode(outputs[0][inputs.shape[-1]:], skip_special_tokens=True)
43
  return response
44
 
45
  # Streamlit app
46
+ st.title("Gemma-2b-it Chatbot")
47
 
48
  # Initialize chat history
49
  if "messages" not in st.session_state:
 
82
  # Add assistant response to chat history
83
  st.session_state.messages.append({"role": "assistant", "content": full_response})
84
  #update the chat history.
85
+ st.session_state.chat_history += prompt + response