OttoYu commited on
Commit
efa6eba
·
verified ·
1 Parent(s): ff49c07

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -17
app.py CHANGED
@@ -3,34 +3,22 @@ import asyncio
3
  import json
4
  import os
5
  import pickle
6
- import torch
7
- from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline, GenerationConfig
8
  from langchain_huggingface import HuggingFaceEmbeddings
9
  from langchain_community.vectorstores import FAISS
10
  from langchain_core.prompts import PromptTemplate
11
  from langchain_community.document_loaders import PDFMinerLoader, CSVLoader, JSONLoader
12
  from langchain.text_splitter import SentenceTransformersTokenTextSplitter
 
13
 
14
- # Define the model and tokenizer
15
- MODEL_NAME = "TheBloke/Llama-2-13B-chat-GPTQ"
16
 
17
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=True)
18
- model = AutoModelForCausalLM.from_pretrained(
19
- MODEL_NAME, torch_dtype=torch.float16, trust_remote_code=True, device_map="auto"
20
- )
21
-
22
- generation_config = GenerationConfig.from_pretrained(MODEL_NAME)
23
- generation_config.max_new_tokens = 1024
24
- generation_config.temperature = 0.0001
25
- generation_config.top_p = 0.95
26
- generation_config.do_sample = True
27
- generation_config.repetition_penalty = 1.15
28
 
29
  text_pipeline = pipeline(
30
  "text-generation",
31
  model=model,
32
- tokenizer=tokenizer,
33
- generation_config=generation_config,
34
  )
35
 
36
  # Define the prompt template
 
3
  import json
4
  import os
5
  import pickle
 
 
6
  from langchain_huggingface import HuggingFaceEmbeddings
7
  from langchain_community.vectorstores import FAISS
8
  from langchain_core.prompts import PromptTemplate
9
  from langchain_community.document_loaders import PDFMinerLoader, CSVLoader, JSONLoader
10
  from langchain.text_splitter import SentenceTransformersTokenTextSplitter
11
+ from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
12
 
13
+ MODEL_NAME = "TheBloke/Llama-2-13B"
 
14
 
15
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
16
+ model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, device_map="cpu")
 
 
 
 
 
 
 
 
 
17
 
18
  text_pipeline = pipeline(
19
  "text-generation",
20
  model=model,
21
+ tokenizer=tokenizer
 
22
  )
23
 
24
  # Define the prompt template