Aditya757864 commited on
Commit
9ebda3d
·
verified ·
1 Parent(s): 928aece

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -7,22 +7,25 @@ from langchain.text_splitter import RecursiveCharacterTextSplitter
7
  from langchain.vectorstores import FAISS
8
  from langchain.memory import ConversationBufferMemory
9
  from langchain_community.document_loaders import PyPDFLoader
10
- from transformers import T5Tokenizer, T5ForConditionalGeneration
11
  import torch
12
  from transformers import pipeline
13
  import os
14
  import tempfile
15
 
 
 
 
16
 
17
  checkpoint = "LaMini-Flan-T5-783M"
18
- tokenizer = T5Tokenizer.from_pretrained(checkpoint)
19
 
20
  base_model = T5ForConditionalGeneration.from_pretrained( checkpoint, device_map = 'auto', torch_dtype = torch.float32 )
21
 
22
  def llm_pipeline():
23
  pipe = pipeline(
24
  'question-answering',
25
- model = base_model,
26
  tokenizer = tokenizer,
27
  do_sample = True,
28
  temperature = 0.5,
 
7
  from langchain.vectorstores import FAISS
8
  from langchain.memory import ConversationBufferMemory
9
  from langchain_community.document_loaders import PyPDFLoader
10
+ from transformers import pipeline, AutoModelForSeq2SeqLM, AutoTokenizer
11
  import torch
12
  from transformers import pipeline
13
  import os
14
  import tempfile
15
 
16
+ model = AutoModelForSeq2SeqLM.from_pretrained("LaMini-Flan-T5-783M")
17
+ tokenizer = AutoTokenizer.from_pretrained("LaMini-Flan-T5-783M", device_map = 'auto', torch_dtype = torch.float32)
18
+
19
 
20
  checkpoint = "LaMini-Flan-T5-783M"
21
+ #tokenizer = T5Tokenizer.from_pretrained(checkpoint)
22
 
23
  base_model = T5ForConditionalGeneration.from_pretrained( checkpoint, device_map = 'auto', torch_dtype = torch.float32 )
24
 
25
  def llm_pipeline():
26
  pipe = pipeline(
27
  'question-answering',
28
+ model = model,
29
  tokenizer = tokenizer,
30
  do_sample = True,
31
  temperature = 0.5,