TimoTM commited on
Commit
38954a9
·
verified ·
1 Parent(s): f4009a8

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -9
app.py CHANGED
@@ -7,10 +7,10 @@ from langchain.chains import RetrievalQA
7
  from langchain.llms.base import LLM
8
  from transformers import pipeline
9
 
10
- # LeoLM-Wrapper-Klasse, die das LeoLM-Modell über die Transformers-Pipeline nutzt
11
- class LeoLM(LLM):
12
- def __init__(self, max_new_tokens=512, temperature=0.5):
13
- self.pipeline = pipeline("text-generation", model="LeoLM/leo-mistral-hessianai-7b")
14
  self.max_new_tokens = max_new_tokens
15
  self.temperature = temperature
16
 
@@ -20,13 +20,13 @@ class LeoLM(LLM):
20
 
21
  @property
22
  def _identifying_params(self):
23
- return {"model": "LeoLM/leo-mistral-hessianai-7b"}
24
 
25
  @property
26
  def _llm_type(self):
27
- return "custom_leolm"
28
 
29
- # PDF wird beim Start automatisch geladen und verarbeitet
30
  loader = PyPDFLoader("TrendingMedia_ChatbotBasis_FINAL.pdf")
31
  documents = loader.load()
32
  splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=100)
@@ -35,8 +35,8 @@ embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-
35
  db = FAISS.from_documents(texts, embeddings)
36
  retriever = db.as_retriever(search_kwargs={"k": 2})
37
 
38
- # Verwende den neuen LeoLM Wrapper als LLM
39
- llm = LeoLM(max_new_tokens=512, temperature=0.5)
40
 
41
  qa_chain = RetrievalQA.from_chain_type(llm=llm, chain_type="stuff", retriever=retriever)
42
 
 
7
  from langchain.llms.base import LLM
8
  from transformers import pipeline
9
 
10
+ # Wrapper-Klasse für das deutsche GPT-2 Modell
11
+ class GermanGPT2(LLM):
12
+ def __init__(self, max_new_tokens=128, temperature=0.7):
13
+ self.pipeline = pipeline("text-generation", model="dbmdz/german-gpt2")
14
  self.max_new_tokens = max_new_tokens
15
  self.temperature = temperature
16
 
 
20
 
21
  @property
22
  def _identifying_params(self):
23
+ return {"model": "dbmdz/german-gpt2"}
24
 
25
  @property
26
  def _llm_type(self):
27
+ return "custom_german_gpt2"
28
 
29
+ # Lade und verarbeite das PDF beim Start
30
  loader = PyPDFLoader("TrendingMedia_ChatbotBasis_FINAL.pdf")
31
  documents = loader.load()
32
  splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=100)
 
35
  db = FAISS.from_documents(texts, embeddings)
36
  retriever = db.as_retriever(search_kwargs={"k": 2})
37
 
38
+ # Verwende den neuen GermanGPT2-Wrapper als LLM
39
+ llm = GermanGPT2(max_new_tokens=128, temperature=0.7)
40
 
41
  qa_chain = RetrievalQA.from_chain_type(llm=llm, chain_type="stuff", retriever=retriever)
42