tastypear commited on
Commit
aafd1b6
·
verified ·
1 Parent(s): 5bca693

Upload 7 files

Browse files
Files changed (7) hide show
  1. .gitattributes +35 -35
  2. Dockerfile +24 -0
  3. README.md +11 -10
  4. main.py +163 -0
  5. requirements.txt +3 -0
  6. space_checker.py +13 -0
  7. start.sh +4 -0
.gitattributes CHANGED
@@ -1,35 +1,35 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
Dockerfile ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # read the doc: https://huggingface.co/docs/hub/spaces-sdks-docker
2
+ # you will also find guides on how best to write your Dockerfile
3
+
4
+ FROM python:3.9
5
+
6
+ WORKDIR /code
7
+
8
+ COPY ./main.py /code/main.py
9
+
10
+ COPY ./space_checker.py /code/space_checker.py
11
+
12
+ COPY ./requirements.txt /code/requirements.txt
13
+
14
+ COPY ./start.sh /code/start.sh
15
+
16
+ COPY . .
17
+
18
+ RUN pip install -r /code/requirements.txt
19
+
20
+ RUN chmod +x start.sh
21
+
22
+ RUN sed -i 's/\r$//' start.sh
23
+
24
+ CMD ["bash","start.sh"]
README.md CHANGED
@@ -1,10 +1,11 @@
1
- ---
2
- title: MiniMax M1 Chat
3
- emoji: 😻
4
- colorFrom: indigo
5
- colorTo: blue
6
- sdk: docker
7
- pinned: false
8
- ---
9
-
10
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
1
+ ---
2
+ title: Minimax Text Chat
3
+ emoji: 💬🗨️
4
+ colorFrom: red
5
+ colorTo: blue
6
+ sdk: docker
7
+ pinned: false
8
+ license: apache-2.0
9
+ ---
10
+
11
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
main.py ADDED
@@ -0,0 +1,163 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gevent.pywsgi
2
+ from gevent import monkey;monkey.patch_all()
3
+ from flask import Flask, request, Response, jsonify
4
+ import argparse
5
+ import requests
6
+ import random
7
+ import string
8
+ import time
9
+ import json
10
+ import os
11
+
12
+ app = Flask(__name__)
13
+ app.json.sort_keys = False
14
+
15
+ parser = argparse.ArgumentParser(description="An example of minimax demo with a similar API to OAI.")
16
+ parser.add_argument("--host", type=str, help="Set the ip address.(default: 0.0.0.0)", default='0.0.0.0')
17
+ parser.add_argument("--port", type=int, help="Set the port.(default: 7860)", default=7860)
18
+ args = parser.parse_args()
19
+
20
+ base_url = os.getenv('MODEL_BASE_URL')
21
+
22
+ @app.route('/api/v1/models', methods=["GET", "POST"])
23
+ @app.route('/v1/models', methods=["GET", "POST"])
24
+ def model_list():
25
+ time_now = int(time.time())
26
+ model_list = {
27
+ "object": "list",
28
+ "data": [
29
+ {
30
+ "id": "MiniMax-Text-01",
31
+ "object": "model",
32
+ "created": time_now,
33
+ "owned_by": "minimax"
34
+ },
35
+ {
36
+ "id": "gpt-3.5-turbo",
37
+ "object": "model",
38
+ "created": time_now,
39
+ "owned_by": "openai"
40
+ }
41
+ ]
42
+ }
43
+ return jsonify(model_list)
44
+
45
+ @app.route("/", methods=["GET"])
46
+ def index():
47
+ return Response(f'Minimax OpenAI Compatible API<br><br>'+
48
+ f'Set "{os.getenv("SPACE_URL")}/api" as proxy (or API Domain) in your Chatbot.<br><br>'+
49
+ f'The complete API is: {os.getenv("SPACE_URL")}/api/v1/chat/completions')
50
+
51
+ @app.route("/api/v1/chat/completions", methods=["POST", "OPTIONS"])
52
+ @app.route("/v1/chat/completions", methods=["POST", "OPTIONS"])
53
+ def chat_completions():
54
+
55
+ if request.method == "OPTIONS":
56
+ return Response(
57
+ headers={
58
+ "Access-Control-Allow-Origin": "*",
59
+ "Access-Control-Allow-Headers": "*",
60
+ }
61
+ )
62
+
63
+ data = request.get_json()
64
+
65
+ # reorganize data
66
+ system = None
67
+ chat_history = []
68
+ prompt = ""
69
+
70
+ if "messages" in data:
71
+ messages = data["messages"]
72
+ message_size = len(messages)
73
+
74
+ prompt = messages[-1].get("content")
75
+
76
+ # gen a random char(11) hash
77
+ chars = string.ascii_lowercase + string.digits
78
+ session_hash = "".join(random.choice(chars) for _ in range(10))
79
+
80
+ single_prompt_data = {
81
+ 'data': [
82
+ prompt,
83
+ None,
84
+ ],
85
+ 'event_data': None,
86
+ 'fn_index': 0,
87
+ 'trigger_id': 7,
88
+ 'session_hash': session_hash,
89
+ }
90
+ response = requests.post(f'{base_url}/gradio_api/run/predict', json=single_prompt_data)
91
+
92
+ context_data = {
93
+ 'data': [
94
+ None,
95
+ messages,
96
+ 1000000,
97
+ data['temperature'],
98
+ data['top_p'],
99
+ ],
100
+ 'event_data': None,
101
+ 'fn_index': 2,
102
+ 'trigger_id': 7,
103
+ 'session_hash': session_hash,
104
+ }
105
+ response = requests.post(f'{base_url}/gradio_api/queue/join', json=context_data)
106
+
107
+ def generate():
108
+
109
+ url = f"{base_url}/gradio_api/queue/data?session_hash={session_hash}"
110
+ data = requests.get(url, stream=True)
111
+
112
+ time_now = int(time.time())
113
+
114
+ for line in data.iter_lines():
115
+ if line:
116
+ decoded_line = line.decode("utf-8")
117
+ json_line = json.loads(decoded_line[6:])
118
+ if json_line["msg"] == "process_starts":
119
+ res_data = gen_res_data({}, time_now=time_now, start=True)
120
+ yield f"data: {json.dumps(res_data)}\n\n"
121
+ elif json_line["msg"] == "process_generating":
122
+ res_data = gen_res_data(json_line, time_now=time_now)
123
+ yield f"data: {json.dumps(res_data)}\n\n"
124
+ elif json_line["msg"] == "process_completed":
125
+ yield "data: [DONE]"
126
+
127
+ return Response(
128
+ generate(),
129
+ mimetype="text/event-stream",
130
+ headers={
131
+ "Access-Control-Allow-Origin": "*",
132
+ "Access-Control-Allow-Headers": "*",
133
+ },
134
+ )
135
+
136
+
137
+ def gen_res_data(data, time_now=0, start=False):
138
+ res_data = {
139
+ "id": "chatcmpl",
140
+ "object": "chat.completion.chunk",
141
+ "created": time_now,
142
+ "model": "MiniMax-Text-01",
143
+ "choices": [{"index": 0, "finish_reason": None}],
144
+ }
145
+
146
+ if start:
147
+ res_data["choices"][0]["delta"] = {"role": "assistant", "content": ""}
148
+ else:
149
+ chat_pair = data["output"]["data"][0]
150
+
151
+ if chat_pair == []:
152
+ res_data["choices"][0]["finish_reason"] = "stop"
153
+ else:
154
+ try:
155
+ res_data["choices"][0]["delta"] = {"content": chat_pair[-1][-1]}
156
+ except:
157
+ res_data["choices"][0]["delta"] = {"content": chat_pair[-1]["content"]}
158
+ return res_data
159
+
160
+
161
+ if __name__ == "__main__":
162
+ #app.run(host=args.host, port=args.port, debug=True)
163
+ gevent.pywsgi.WSGIServer((args.host, args.port), app).serve_forever()
requirements.txt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ flask
2
+ requests
3
+ gevent
space_checker.py ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import requests
2
+ import time
3
+ import os
4
+
5
+ space_url = os.getenv('SPACE_URL')
6
+
7
+ def fetch_url(url):
8
+ response = requests.get(url)
9
+ print(response.text)
10
+
11
+ while True:
12
+ fetch_url(space_url)
13
+ time.sleep(3600)
start.sh ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ #!/bin/bash
2
+ python main.py &
3
+ python space_checker.py &
4
+ wait