| | import gradio as gr |
| | import PyPDF2 |
| | from langchain.embeddings.openai import OpenAIEmbeddings |
| | from langchain.vectorstores.faiss import FAISS |
| | from langchain.text_splitter import RecursiveCharacterTextSplitter |
| | from langchain import OpenAI, VectorDBQA |
| |
|
| | import os |
| | openai_api_key = os.environ["OPENAI_API_KEY"] |
| |
|
| |
|
| | def pdf_to_text(pdf_file, query): |
| | |
| | with open(pdf_file.name, 'rb') as pdf_file: |
| | |
| | pdf_reader = PyPDF2.PdfReader(pdf_file) |
| |
|
| | |
| | text = "" |
| |
|
| | |
| | for page_num in range(len(pdf_reader.pages)): |
| | |
| | page = pdf_reader.pages[page_num] |
| | |
| | text += page.extract_text() |
| | |
| | text_splitter = RecursiveCharacterTextSplitter(chunk_size=1000, chunk_overlap=0) |
| | texts = text_splitter.split_text(text) |
| |
|
| | embeddings = OpenAIEmbeddings() |
| | |
| | vectorstore = FAISS.from_texts(texts, embeddings) |
| |
|
| | |
| | qa = VectorDBQA.from_chain_type(llm=OpenAI(), chain_type="stuff", vectorstore=vectorstore) |
| | return qa.run(query) |
| |
|
| | examples = [ |
| | [os.path.abspath("NASDAQ_AAPL_2020.pdf"), "how much are the outstanding shares ?"], |
| | [os.path.abspath("NASDAQ_AAPL_2020.pdf"), "what is competitors strategy ?"], |
| | [os.path.abspath("NASDAQ_AAPL_2020.pdf"), "who is the chief executive officer ?"], |
| | [os.path.abspath("NASDAQ_MSFT_2020.pdf"), "How much is the guided revenue for next quarter?"], |
| | [os.path.abspath("example_file.pdf"), "what are the name of all authors?"], |
| | [os.path.abspath("breast cancer.pdf"), "What Causes Breast Cancer?"], |
| | [os.path.abspath("v61n3a11.pdf"), "what is MDT?"] |
| | |
| | ] |
| | |
| |
|
| |
|
| | |
| | |
| | |
| | |
| |
|
| | interface = gr.Interface(fn=pdf_to_text, |
| | inputs= [gr.inputs.File(label="input pdf file"), gr.inputs.Textbox(label="Question:")], |
| | outputs =gr.outputs.Textbox(label="Chatbot Response"), |
| | examples = examples, |
| | ) |
| |
|
| | |
| | interface.launch(enable_queue = True) |