Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,21 +28,28 @@ GROQ_API_KEY = os.environ["GROQ_API_KEY"]
|
|
| 28 |
HF_TOKEN = os.environ["HF_Token"]
|
| 29 |
|
| 30 |
|
| 31 |
-
VECTORSTORE_DIRECTORY =
|
| 32 |
-
CHUNK_SIZE =
|
| 33 |
-
CHUNK_OVERLAP =
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 46 |
|
| 47 |
|
| 48 |
login(HF_TOKEN)
|
|
|
|
| 28 |
HF_TOKEN = os.environ["HF_Token"]
|
| 29 |
|
| 30 |
|
| 31 |
+
VECTORSTORE_DIRECTORY = "vectorstore_directory"
|
| 32 |
+
CHUNK_SIZE = 1024
|
| 33 |
+
CHUNK_OVERLAP = 128
|
| 34 |
+
|
| 35 |
+
EMBEDDING_MODEL_NAME = "ibm-granite/granite-embedding-125m-english"
|
| 36 |
+
EMBEDDING_MODEL_VERSION = "latest"
|
| 37 |
+
|
| 38 |
+
LLM_MODEL_NAME = "llama-3.3-70b-versatile"
|
| 39 |
+
LLM_MODEL_TEMPERATURE = 0.0
|
| 40 |
+
|
| 41 |
+
GITLAB_API_URL = "${GITLAB_API_URL}"
|
| 42 |
+
GITLAB_PROJECT_ID = "iam-cms/kadi-apy"
|
| 43 |
+
GITLAB_PROJECT_VERSION = "v0.45.0"
|
| 44 |
+
|
| 45 |
+
DATA_DIR = "${DATA_DIR}"
|
| 46 |
+
|
| 47 |
+
HF_SPACE_NAME = "bupa1018/Kadi4Mat_Ai-Assistant-Bot"
|
| 48 |
+
|
| 49 |
+
#DOCS_FOLDER = config.get("usage", "docs", "folder")
|
| 50 |
+
#DOCS_FILE = config.get("usage", "docs", "file")
|
| 51 |
+
#KADI_APY_FOLDER = config.get("usage", "kadi_apy", "folder")
|
| 52 |
+
#KADI_APY_FILE = config.get("usage", "kadi_apy", "file")
|
| 53 |
|
| 54 |
|
| 55 |
login(HF_TOKEN)
|