AbdulWahab14 commited on
Commit
eb77d48
·
verified ·
1 Parent(s): 6808dee

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +18 -30
app.py CHANGED
@@ -1,34 +1,21 @@
1
- import sys
2
  import streamlit as st
3
-
4
- HF_API_TOKEN = "your_huggingface_api_token_here" # <-- paste your token
5
- MODEL_ID = "mistralai/Mistral-7B-Instruct-v0.2"
6
- API_URL = f"https://api-inference.huggingface.co/models/{MODEL_ID}"
7
- HEADERS = {"Authorization": f"Bearer {HF_API_TOKEN}"}
8
-
9
- import streamlit as st
10
- from transformers import AutoTokenizer, AutoModelForCausalLM, AutoConfig
11
  import torch
 
12
 
13
- st.title("📚 AI Adaptive Learning (Local Small Model)")
14
 
15
- MODEL_ID = "microsoft/phi-2"
16
 
17
  @st.cache_resource
18
  def load_model():
19
- tokenizer = AutoTokenizer.from_pretrained(
20
- MODEL_ID,
21
- trust_remote_code=True
22
- )
23
 
24
  if tokenizer.pad_token is None:
25
  tokenizer.pad_token = tokenizer.eos_token
26
 
27
  model = AutoModelForCausalLM.from_pretrained(
28
  MODEL_ID,
29
- trust_remote_code=True,
30
- torch_dtype=torch.float32 # CPU safe
31
- # ❌ removed device_map="auto"
32
  )
33
 
34
  model.eval()
@@ -36,21 +23,22 @@ def load_model():
36
 
37
  tokenizer, model = load_model()
38
 
39
- # Input question
40
  user_input = st.text_input("Ask a question:")
41
 
42
  if st.button("Submit") and user_input:
43
- inputs = tokenizer(user_input, return_tensors="pt")
44
-
45
- with torch.no_grad():
46
- outputs = model.generate(
47
- **inputs,
48
- max_new_tokens=150,
49
- do_sample=True,
50
- temperature=0.7
51
- )
52
-
53
- answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
 
 
54
 
55
  st.subheader("AI Answer:")
56
  st.write(answer)
 
 
1
  import streamlit as st
 
 
 
 
 
 
 
 
2
  import torch
3
+ from transformers import AutoTokenizer, AutoModelForCausalLM
4
 
5
+ st.title("📚 AI Adaptive Learning")
6
 
7
+ MODEL_ID = "TinyLlama/TinyLlama-1.1B-Chat-v1.0"
8
 
9
  @st.cache_resource
10
  def load_model():
11
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
 
 
 
12
 
13
  if tokenizer.pad_token is None:
14
  tokenizer.pad_token = tokenizer.eos_token
15
 
16
  model = AutoModelForCausalLM.from_pretrained(
17
  MODEL_ID,
18
+ torch_dtype=torch.float32
 
 
19
  )
20
 
21
  model.eval()
 
23
 
24
  tokenizer, model = load_model()
25
 
 
26
  user_input = st.text_input("Ask a question:")
27
 
28
  if st.button("Submit") and user_input:
29
+ with st.spinner("Generating answer..."):
30
+ inputs = tokenizer(user_input, return_tensors="pt")
31
+
32
+ with torch.no_grad():
33
+ outputs = model.generate(
34
+ **inputs,
35
+ max_new_tokens=150,
36
+ temperature=0.7,
37
+ do_sample=True,
38
+ pad_token_id=tokenizer.eos_token_id
39
+ )
40
+
41
+ answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
42
 
43
  st.subheader("AI Answer:")
44
  st.write(answer)