Spaces:
Sleeping
Sleeping
Update config.py
Browse files
config.py
CHANGED
|
@@ -11,10 +11,6 @@
|
|
| 11 |
API_KEY = "此处填API密钥" # 可同时填写多个API-KEY,用英文逗号分割,例如API_KEY = "sk-openaikey1,sk-openaikey2,fkxxxx-api2dkey3,azure-apikey4"
|
| 12 |
|
| 13 |
|
| 14 |
-
# [step 1]>> API_KEY = "sk-123456789xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx123456789"。极少数情况下,还需要填写组织(格式如org-123456789abcdefghijklmno的),请向下翻,找 API_ORG 设置项
|
| 15 |
-
API_KEY = "此处填API密钥" # 可同时填写多个API-KEY,用英文逗号分割,例如API_KEY = "sk-openaikey1,sk-openaikey2,fkxxxx-api2dkey3,azure-apikey4"
|
| 16 |
-
|
| 17 |
-
|
| 18 |
# [step 2]>> 改为True应用代理,如果直接在海外服务器部署,此处不修改;如果使用本地或无地域限制的大模型时,此处也不需要修改
|
| 19 |
USE_PROXY = False
|
| 20 |
if USE_PROXY:
|
|
@@ -44,12 +40,12 @@ API_URL_REDIRECT = {}
|
|
| 44 |
|
| 45 |
# 多线程函数插件中,默认允许多少路线程同时访问OpenAI。Free trial users的限制是每分钟3次,Pay-as-you-go users的限制是每分钟3500次
|
| 46 |
# 一言以蔽之:免费(5刀)用户填3,OpenAI绑了信用卡的用户可以填 16 或者更高。提高限制请查询:https://platform.openai.com/docs/guides/rate-limits/overview
|
| 47 |
-
DEFAULT_WORKER_NUM =
|
| 48 |
|
| 49 |
|
| 50 |
# 色彩主题, 可选 ["Default", "Chuanhu-Small-and-Beautiful", "High-Contrast"]
|
| 51 |
# 更多主题, 请查阅Gradio主题商店: https://huggingface.co/spaces/gradio/theme-gallery 可选 ["Gstaff/Xkcd", "NoCrypt/Miku", ...]
|
| 52 |
-
THEME = "
|
| 53 |
AVAIL_THEMES = ["Default", "Chuanhu-Small-and-Beautiful", "High-Contrast", "Gstaff/Xkcd", "NoCrypt/Miku"]
|
| 54 |
|
| 55 |
|
|
@@ -70,11 +66,11 @@ LAYOUT = "LEFT-RIGHT" # "LEFT-RIGHT"(左右布局) # "TOP-DOWN"(上下
|
|
| 70 |
|
| 71 |
|
| 72 |
# 暗色模式 / 亮色模式
|
| 73 |
-
DARK_MODE =
|
| 74 |
|
| 75 |
|
| 76 |
# 发送请求到OpenAI后,等待多久判定为超时
|
| 77 |
-
TIMEOUT_SECONDS =
|
| 78 |
|
| 79 |
|
| 80 |
# 网页的端口, -1代表随机端口
|
|
@@ -84,35 +80,20 @@ WEB_PORT = -1
|
|
| 84 |
# 如果OpenAI不响应(网络卡顿、代理失败、KEY失效),重试的次数限制
|
| 85 |
MAX_RETRY = 2
|
| 86 |
|
| 87 |
-
# OpenAI模型选择是(gpt4现在只对申请成功的人开放)
|
| 88 |
-
LLM_MODEL = "gpt-3.5-turbo" # 可选 "chatglm"
|
| 89 |
-
AVAIL_LLM_MODELS = ["gpt-3.5-turbo", "gpt-4", "api2d-gpt-4", "api2d-gpt-3.5-turbo", "spark", "azure-gpt-3.5"]
|
| 90 |
|
| 91 |
# 插件分类默认选项
|
| 92 |
-
DEFAULT_FN_GROUPS = ['对话', '编程'
|
| 93 |
|
| 94 |
|
| 95 |
# 模型选择是 (注意: LLM_MODEL是默认选中的模型, 它*必须*被包含在AVAIL_LLM_MODELS列表中 )
|
| 96 |
LLM_MODEL = "gpt-3.5-turbo" # 可选 ↓↓↓
|
| 97 |
-
AVAIL_LLM_MODELS = ["gpt-3.5-turbo-1106","gpt-
|
| 98 |
-
|
| 99 |
-
|
| 100 |
-
|
| 101 |
-
|
| 102 |
-
# P.S. 其他可用的模型还包括 ["zhipuai", "qianfan", "deepseekcoder", "llama2", "qwen-local", "gpt-3.5-turbo-0613", "gpt-3.5-turbo-16k-0613", "gpt-3.5-random"
|
| 103 |
# "spark", "sparkv2", "sparkv3", "chatglm_onnx", "claude-1-100k", "claude-2", "internlm", "jittorllms_pangualpha", "jittorllms_llama"
|
| 104 |
-
#
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
# 定义界面上“询问多个GPT模型”插件应该使用哪些模型,请从AVAIL_LLM_MODELS中选择,并在不同模型之间用`&`间隔,例如"gpt-3.5-turbo&chatglm3&azure-gpt-4"
|
| 108 |
-
MULTI_QUERY_LLM_MODELS = "gpt-3.5-turbo&chatglm3"
|
| 109 |
-
|
| 110 |
-
|
| 111 |
-
# 选择本地模型变体(只有当AVAIL_LLM_MODELS包含了对应本地模型时,才会起作用)
|
| 112 |
-
# 如果你选择Qwen系列的模型,那么请在下面的QWEN_MODEL_SELECTION中指定具体的模型
|
| 113 |
-
# 也可以是具体的模型路径
|
| 114 |
-
QWEN_LOCAL_MODEL_SELECTION = "Qwen/Qwen-1_8B-Chat-Int8"
|
| 115 |
-
|
| 116 |
|
| 117 |
# 接入通义千问在线大模型 https://dashscope.console.aliyun.com/
|
| 118 |
DASHSCOPE_API_KEY = "" # 阿里灵积云API_KEY
|
|
@@ -128,9 +109,6 @@ BAIDU_CLOUD_QIANFAN_MODEL = 'ERNIE-Bot' # 可选 "ERNIE-Bot-4"(文心大模
|
|
| 128 |
CHATGLM_PTUNING_CHECKPOINT = "" # 例如"/home/hmp/ChatGLM2-6B/ptuning/output/6b-pt-128-1e-2/checkpoint-100"
|
| 129 |
|
| 130 |
|
| 131 |
-
# 本地LLM模型如ChatGLM的执行方式 CPU/GPU
|
| 132 |
-
LOCAL_MODEL_DEVICE = "cpu" # 可选 "cuda"
|
| 133 |
-
LOCAL_MODEL_QUANT = "FP16" # 默认 "FP16" "INT4" 启用量化INT4版本 "INT8" 启用量化INT8版本
|
| 134 |
|
| 135 |
# 设置gradio的并行线程数(不需要修改)
|
| 136 |
CONCURRENT_COUNT = 100
|
|
@@ -141,7 +119,7 @@ AUTO_CLEAR_TXT = False
|
|
| 141 |
|
| 142 |
|
| 143 |
# 加一个live2d装饰
|
| 144 |
-
ADD_WAIFU =
|
| 145 |
|
| 146 |
|
| 147 |
# 设置用户名和密码(不需要修改)(相关功能不稳定,与gradio版本和���络都相关,如果本地使用不建议加这个)
|
|
@@ -162,11 +140,6 @@ SSL_CERTFILE = ""
|
|
| 162 |
API_ORG = ""
|
| 163 |
|
| 164 |
|
| 165 |
-
# 如果需要使用Slack Claude,使用教程详情见 request_llms/README.md
|
| 166 |
-
SLACK_CLAUDE_BOT_ID = ''
|
| 167 |
-
SLACK_CLAUDE_USER_TOKEN = ''
|
| 168 |
-
|
| 169 |
-
|
| 170 |
# 如果需要使用AZURE(方法一:单个azure模型部署)详情请见额外文档 docs\use_azure.md
|
| 171 |
AZURE_ENDPOINT = "https://你亲手写的api名称.openai.azure.com/"
|
| 172 |
AZURE_API_KEY = "填入azure openai api的密钥" # 建议直接在API_KEY处填写,该选项即将被弃用
|
|
@@ -203,25 +176,17 @@ ZHIPUAI_API_KEY = ""
|
|
| 203 |
ZHIPUAI_MODEL = "chatglm_turbo"
|
| 204 |
|
| 205 |
|
| 206 |
-
# Claude API KEY
|
| 207 |
-
ANTHROPIC_API_KEY = ""
|
| 208 |
-
|
| 209 |
|
| 210 |
# 自定义API KEY格式
|
| 211 |
CUSTOM_API_KEY_PATTERN = ""
|
| 212 |
|
| 213 |
|
| 214 |
-
#
|
| 215 |
-
|
| 216 |
|
| 217 |
|
| 218 |
-
#
|
| 219 |
-
|
| 220 |
-
GROBID_URLS = [
|
| 221 |
-
"https://qingxu98-grobid.hf.space","https://qingxu98-grobid2.hf.space","https://qingxu98-grobid3.hf.space",
|
| 222 |
-
"https://qingxu98-grobid4.hf.space","https://qingxu98-grobid5.hf.space", "https://qingxu98-grobid6.hf.space",
|
| 223 |
-
"https://qingxu98-grobid7.hf.space", "https://qingxu98-grobid8.hf.space",
|
| 224 |
-
]
|
| 225 |
|
| 226 |
|
| 227 |
# 是否允许通过自然语言描述修改本页的配置,该功能具有一定的危险性,默认关闭
|
|
@@ -250,7 +215,7 @@ BLOCK_INVALID_APIKEY = False
|
|
| 250 |
|
| 251 |
|
| 252 |
# 启用插件热加载
|
| 253 |
-
PLUGIN_HOT_RELOAD =
|
| 254 |
|
| 255 |
|
| 256 |
# 自定义按钮的最大数量限制
|
|
@@ -299,25 +264,13 @@ NUM_CUSTOM_BASIC_BTN = 4
|
|
| 299 |
├── "qwen-turbo" 等通义千问大模型
|
| 300 |
│ └── DASHSCOPE_API_KEY
|
| 301 |
│
|
|
|
|
|
|
|
|
|
|
| 302 |
└── "newbing" Newbing接口不再稳定,不推荐使用
|
| 303 |
├── NEWBING_STYLE
|
| 304 |
└── NEWBING_COOKIES
|
| 305 |
|
| 306 |
-
|
| 307 |
-
本地大模型示意图
|
| 308 |
-
│
|
| 309 |
-
├── "chatglm3"
|
| 310 |
-
├── "chatglm"
|
| 311 |
-
├── "chatglm_onnx"
|
| 312 |
-
├── "chatglmft"
|
| 313 |
-
├── "internlm"
|
| 314 |
-
├── "moss"
|
| 315 |
-
├── "jittorllms_pangualpha"
|
| 316 |
-
├── "jittorllms_llama"
|
| 317 |
-
├── "deepseekcoder"
|
| 318 |
-
├── "qwen-local"
|
| 319 |
-
├── RWKV的支持见Wiki
|
| 320 |
-
└── "llama2"
|
| 321 |
|
| 322 |
|
| 323 |
用户图形界面布局依赖关系示意图
|
|
|
|
| 11 |
API_KEY = "此处填API密钥" # 可同时填写多个API-KEY,用英文逗号分割,例如API_KEY = "sk-openaikey1,sk-openaikey2,fkxxxx-api2dkey3,azure-apikey4"
|
| 12 |
|
| 13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
# [step 2]>> 改为True应用代理,如果直接在海外服务器部署,此处不修改;如果使用本地或无地域限制的大模型时,此处也不需要修改
|
| 15 |
USE_PROXY = False
|
| 16 |
if USE_PROXY:
|
|
|
|
| 40 |
|
| 41 |
# 多线程函数插件中,默认允许多少路线程同时访问OpenAI。Free trial users的限制是每分钟3次,Pay-as-you-go users的限制是每分钟3500次
|
| 42 |
# 一言以蔽之:免费(5刀)用户填3,OpenAI绑了信用卡的用户可以填 16 或者更高。提高限制请查询:https://platform.openai.com/docs/guides/rate-limits/overview
|
| 43 |
+
DEFAULT_WORKER_NUM = 10
|
| 44 |
|
| 45 |
|
| 46 |
# 色彩主题, 可选 ["Default", "Chuanhu-Small-and-Beautiful", "High-Contrast"]
|
| 47 |
# 更多主题, 请查阅Gradio主题商店: https://huggingface.co/spaces/gradio/theme-gallery 可选 ["Gstaff/Xkcd", "NoCrypt/Miku", ...]
|
| 48 |
+
THEME = "Default"
|
| 49 |
AVAIL_THEMES = ["Default", "Chuanhu-Small-and-Beautiful", "High-Contrast", "Gstaff/Xkcd", "NoCrypt/Miku"]
|
| 50 |
|
| 51 |
|
|
|
|
| 66 |
|
| 67 |
|
| 68 |
# 暗色模式 / 亮色模式
|
| 69 |
+
DARK_MODE = True
|
| 70 |
|
| 71 |
|
| 72 |
# 发送请求到OpenAI后,等待多久判定为超时
|
| 73 |
+
TIMEOUT_SECONDS = 300
|
| 74 |
|
| 75 |
|
| 76 |
# 网页的端口, -1代表随机端口
|
|
|
|
| 80 |
# 如果OpenAI不响应(网络卡顿、代理失败、KEY失效),重试的次数限制
|
| 81 |
MAX_RETRY = 2
|
| 82 |
|
|
|
|
|
|
|
|
|
|
| 83 |
|
| 84 |
# 插件分类默认选项
|
| 85 |
+
DEFAULT_FN_GROUPS = ['对话', '编程']
|
| 86 |
|
| 87 |
|
| 88 |
# 模型选择是 (注意: LLM_MODEL是默认选中的模型, 它*必须*被包含在AVAIL_LLM_MODELS列表中 )
|
| 89 |
LLM_MODEL = "gpt-3.5-turbo" # 可选 ↓↓↓
|
| 90 |
+
AVAIL_LLM_MODELS = ["gpt-3.5-turbo-1106","gpt-3.5-turbo-16k", "gpt-3.5-turbo","qwen-max","gemini-pro"]
|
| 91 |
+
# P.S. 其他可用的模型还包括 [
|
| 92 |
+
# "qwen-turbo", "qwen-plus", "qwen-max"
|
| 93 |
+
# "zhipuai", "qianfan", "deepseekcoder", "llama2", "qwen-local", "gpt-3.5-turbo-0613",
|
| 94 |
+
# "gpt-3.5-turbo-16k-0613", "gpt-3.5-random", "api2d-gpt-3.5-turbo", 'api2d-gpt-3.5-turbo-16k',
|
|
|
|
| 95 |
# "spark", "sparkv2", "sparkv3", "chatglm_onnx", "claude-1-100k", "claude-2", "internlm", "jittorllms_pangualpha", "jittorllms_llama"
|
| 96 |
+
# ]
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 97 |
|
| 98 |
# 接入通义千问在线大模型 https://dashscope.console.aliyun.com/
|
| 99 |
DASHSCOPE_API_KEY = "" # 阿里灵积云API_KEY
|
|
|
|
| 109 |
CHATGLM_PTUNING_CHECKPOINT = "" # 例如"/home/hmp/ChatGLM2-6B/ptuning/output/6b-pt-128-1e-2/checkpoint-100"
|
| 110 |
|
| 111 |
|
|
|
|
|
|
|
|
|
|
| 112 |
|
| 113 |
# 设置gradio的并行线程数(不需要修改)
|
| 114 |
CONCURRENT_COUNT = 100
|
|
|
|
| 119 |
|
| 120 |
|
| 121 |
# 加一个live2d装饰
|
| 122 |
+
ADD_WAIFU = False
|
| 123 |
|
| 124 |
|
| 125 |
# 设置用户名和密码(不需要修改)(相关功能不稳定,与gradio版本和���络都相关,如果本地使用不建议加这个)
|
|
|
|
| 140 |
API_ORG = ""
|
| 141 |
|
| 142 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 143 |
# 如果需要使用AZURE(方法一:单个azure模型部署)详情请见额外文档 docs\use_azure.md
|
| 144 |
AZURE_ENDPOINT = "https://你亲手写的api名称.openai.azure.com/"
|
| 145 |
AZURE_API_KEY = "填入azure openai api的密钥" # 建议直接在API_KEY处填写,该选项即将被弃用
|
|
|
|
| 176 |
ZHIPUAI_MODEL = "chatglm_turbo"
|
| 177 |
|
| 178 |
|
|
|
|
|
|
|
|
|
|
| 179 |
|
| 180 |
# 自定义API KEY格式
|
| 181 |
CUSTOM_API_KEY_PATTERN = ""
|
| 182 |
|
| 183 |
|
| 184 |
+
# Google Gemini API-Key
|
| 185 |
+
GEMINI_API_KEY = ''
|
| 186 |
|
| 187 |
|
| 188 |
+
# HUGGINGFACE的TOKEN,下载LLAMA时起作用 https://huggingface.co/docs/hub/security-tokens
|
| 189 |
+
HUGGINGFACE_ACCESS_TOKEN = "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"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 190 |
|
| 191 |
|
| 192 |
# 是否允许通过自然语言描述修改本页的配置,该功能具有一定的危险性,默认关闭
|
|
|
|
| 215 |
|
| 216 |
|
| 217 |
# 启用插件热加载
|
| 218 |
+
PLUGIN_HOT_RELOAD = True
|
| 219 |
|
| 220 |
|
| 221 |
# 自定义按钮的最大数量限制
|
|
|
|
| 264 |
├── "qwen-turbo" 等通义千问大模型
|
| 265 |
│ └── DASHSCOPE_API_KEY
|
| 266 |
│
|
| 267 |
+
├── "Gemini"
|
| 268 |
+
│ └── GEMINI_API_KEY
|
| 269 |
+
│
|
| 270 |
└── "newbing" Newbing接口不再稳定,不推荐使用
|
| 271 |
├── NEWBING_STYLE
|
| 272 |
└── NEWBING_COOKIES
|
| 273 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 274 |
|
| 275 |
|
| 276 |
用户图形界面布局依赖关系示意图
|