AIPretender commited on
Commit
91a909a
·
verified ·
1 Parent(s): f852c0f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -2
app.py CHANGED
@@ -2,7 +2,18 @@ from langchain_community.document_loaders import PyPDFLoader
2
  import gradio as gr
3
  from langchain.chains.summarize import load_summarize_chain
4
  from huggingface_hub import InferenceClient
5
- TEXT_MODEL_NAME = "google/gemma-3-270m"
 
 
 
 
 
 
 
 
 
 
 
6
  loader = PyPDFLoader("http://arxiv.org/pdf/2508.13246v1")
7
  documents = loader.load()
8
  #llm = OpenAI(temperature=0)
@@ -10,7 +21,7 @@ documents = loader.load()
10
  def summarize_pdf (pdf_file_path, custom_prompt=""):
11
  loader = PyPDFLoader(pdf_file_path)
12
  docs = loader.load_and_split()
13
- chain = load_summarize_chain(TEXT_MODEL_NAME, chain_type="map_reduce")
14
  summary = chain.run(docs)
15
 
16
  return summary
 
2
  import gradio as gr
3
  from langchain.chains.summarize import load_summarize_chain
4
  from huggingface_hub import InferenceClient
5
+ # Use a pipeline as a high-level helper
6
+ from transformers import pipeline
7
+
8
+ pipe = pipeline("summarization", model="facebook/bart-large-cnn")
9
+
10
+ # Load model directly
11
+ from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
12
+
13
+ tokenizer = AutoTokenizer.from_pretrained("facebook/bart-large-cnn")
14
+ model = AutoModelForSeq2SeqLM.from_pretrained("facebook/bart-large-cnn")
15
+
16
+ #TEXT_MODEL_NAME = "google/gemma-3-270m"
17
  loader = PyPDFLoader("http://arxiv.org/pdf/2508.13246v1")
18
  documents = loader.load()
19
  #llm = OpenAI(temperature=0)
 
21
  def summarize_pdf (pdf_file_path, custom_prompt=""):
22
  loader = PyPDFLoader(pdf_file_path)
23
  docs = loader.load_and_split()
24
+ chain = load_summarize_chain(model, chain_type="map_reduce")
25
  summary = chain.run(docs)
26
 
27
  return summary