vikasdeep commited on
Commit
f9d2833
·
verified ·
1 Parent(s): e384a57

Upload 3 files

Browse files
Files changed (3) hide show
  1. .gitattributes +35 -35
  2. README.md +15 -14
  3. app.py +169 -0
.gitattributes CHANGED
@@ -1,35 +1,35 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ *.7z filter=lfs diff=lfs merge=lfs -text
2
+ *.arrow filter=lfs diff=lfs merge=lfs -text
3
+ *.bin filter=lfs diff=lfs merge=lfs -text
4
+ *.bz2 filter=lfs diff=lfs merge=lfs -text
5
+ *.ckpt filter=lfs diff=lfs merge=lfs -text
6
+ *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gz filter=lfs diff=lfs merge=lfs -text
8
+ *.h5 filter=lfs diff=lfs merge=lfs -text
9
+ *.joblib filter=lfs diff=lfs merge=lfs -text
10
+ *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
+ *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
+ *.model filter=lfs diff=lfs merge=lfs -text
13
+ *.msgpack filter=lfs diff=lfs merge=lfs -text
14
+ *.npy filter=lfs diff=lfs merge=lfs -text
15
+ *.npz filter=lfs diff=lfs merge=lfs -text
16
+ *.onnx filter=lfs diff=lfs merge=lfs -text
17
+ *.ot filter=lfs diff=lfs merge=lfs -text
18
+ *.parquet filter=lfs diff=lfs merge=lfs -text
19
+ *.pb filter=lfs diff=lfs merge=lfs -text
20
+ *.pickle filter=lfs diff=lfs merge=lfs -text
21
+ *.pkl filter=lfs diff=lfs merge=lfs -text
22
+ *.pt filter=lfs diff=lfs merge=lfs -text
23
+ *.pth filter=lfs diff=lfs merge=lfs -text
24
+ *.rar filter=lfs diff=lfs merge=lfs -text
25
+ *.safetensors filter=lfs diff=lfs merge=lfs -text
26
+ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
+ *.tar.* filter=lfs diff=lfs merge=lfs -text
28
+ *.tar filter=lfs diff=lfs merge=lfs -text
29
+ *.tflite filter=lfs diff=lfs merge=lfs -text
30
+ *.tgz filter=lfs diff=lfs merge=lfs -text
31
+ *.wasm filter=lfs diff=lfs merge=lfs -text
32
+ *.xz filter=lfs diff=lfs merge=lfs -text
33
+ *.zip filter=lfs diff=lfs merge=lfs -text
34
+ *.zst filter=lfs diff=lfs merge=lfs -text
35
+ *tfevents* filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,14 +1,15 @@
1
- ---
2
- title: Neuro Vision
3
- emoji: 📈
4
- colorFrom: yellow
5
- colorTo: pink
6
- sdk: gradio
7
- sdk_version: 5.31.0
8
- app_file: app.py
9
- pinned: false
10
- license: apache-2.0
11
- short_description: mathematical model
12
- ---
13
-
14
- Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
 
 
1
+ ---
2
+ title: Neuro Vision
3
+ emoji: 🧮
4
+ colorFrom: indigo
5
+ colorTo: indigo
6
+ sdk: gradio
7
+ sdk_version: 4.38.1
8
+ app_file: app.py
9
+ pinned: false
10
+ license: other
11
+ models:
12
+ - Qwen/Qwen2-Math-72B-Instruct
13
+ ---
14
+
15
+ Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
app.py ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import gradio as gr
2
+ import os
3
+
4
+ os.system('pip install dashscope -U')
5
+ import tempfile
6
+ from pathlib import Path
7
+ import secrets
8
+ import dashscope
9
+ from dashscope import MultiModalConversation, Generation
10
+ from PIL import Image
11
+
12
+
13
+ # API
14
+ YOUR_API_TOKEN = os.getenv('YOUR_API_TOKEN')
15
+ dashscope.api_key = YOUR_API_TOKEN
16
+ math_messages = []
17
+ def process_image(image, shouldConvert=False):
18
+
19
+ global math_messages
20
+ math_messages = [] # reset when upload image
21
+ uploaded_file_dir = os.environ.get("GRADIO_TEMP_DIR") or str(
22
+ Path(tempfile.gettempdir()) / "gradio"
23
+ )
24
+ os.makedirs(uploaded_file_dir, exist_ok=True)
25
+
26
+
27
+ name = f"tmp{secrets.token_hex(20)}.jpg"
28
+ filename = os.path.join(uploaded_file_dir, name)
29
+
30
+ if shouldConvert:
31
+ new_img = Image.new('RGB', size=(image.width, image.height), color=(255, 255, 255))
32
+ new_img.paste(image, (0, 0), mask=image)
33
+ image = new_img
34
+ image.save(filename)
35
+
36
+ # qwen-vl-max-0809
37
+ messages = [{
38
+ 'role': 'system',
39
+ 'content': [{'text': 'You are a helpful assistant.'}]
40
+ }, {
41
+ 'role': 'user',
42
+ 'content': [
43
+ {'image': f'file://{filename}'},
44
+ {'text': 'Please describe the math-related content in this image, ensuring that any LaTeX formulas are correctly transcribed. Non-mathematical details do not need to be described.'}
45
+ ]
46
+ }]
47
+
48
+ response = MultiModalConversation.call(model='qwen-vl-max-0809', messages=messages)
49
+
50
+
51
+ os.remove(filename)
52
+
53
+ return response.output.choices[0]["message"]["content"]
54
+
55
+ def get_math_response(image_description, user_question):
56
+ global math_messages
57
+ if not math_messages:
58
+ math_messages.append({'role': 'system', 'content': 'You are a helpful math assistant.'})
59
+ math_messages = math_messages[:1]
60
+ if image_description is not None:
61
+ content = f'Image description: {image_description}\n\n'
62
+ else:
63
+ content = ''
64
+ query = f"{content}User question: {user_question}"
65
+ math_messages.append({'role': 'user', 'content': query})
66
+ response = Generation.call(
67
+ model="qwen2.5-math-72b-instruct",
68
+ messages=math_messages,
69
+ result_format='message',
70
+ stream=True
71
+ )
72
+ answer = None
73
+ for resp in response:
74
+ if resp.output is None:
75
+ continue
76
+ answer = resp.output.choices[0].message.content
77
+ yield answer.replace("\\", "\\\\")
78
+ print(f'query: {query}\nanswer: {answer}')
79
+ if answer is None:
80
+ math_messages.pop()
81
+ else:
82
+ math_messages.append({'role': 'assistant', 'content': answer})
83
+
84
+
85
+ def math_chat_bot(image, sketchpad, question, state):
86
+ current_tab_index = state["tab_index"]
87
+ image_description = None
88
+ # Upload
89
+ if current_tab_index == 0:
90
+ if image is not None:
91
+ image_description = process_image(image)
92
+ # Sketch
93
+ elif current_tab_index == 1:
94
+ print(sketchpad)
95
+ if sketchpad and sketchpad["composite"]:
96
+ image_description = process_image(sketchpad["composite"], True)
97
+ yield from get_math_response(image_description, question)
98
+
99
+ css = """
100
+ #qwen-md .katex-display { display: inline; }
101
+ #qwen-md .katex-display>.katex { display: inline; }
102
+ #qwen-md .katex-display>.katex>.katex-html { display: inline; }
103
+ """
104
+
105
+ def tabs_select(e: gr.SelectData, _state):
106
+ _state["tab_index"] = e.index
107
+
108
+
109
+ # Gradio
110
+ with gr.Blocks(css=css) as demo:
111
+ gr.HTML("""\
112
+ <p align="center"><img src="https://modelscope.oss-cn-beijing.aliyuncs.com/resource/qwen.png" style="height: 60px"/><p>"""
113
+ """<center><font size=8>📖 Qwen2.5-Math Demo</center>"""
114
+ """\
115
+ <center><font size=3>This WebUI is based on Qwen2-VL for OCR and Qwen2.5-Math for mathematical reasoning. You can input either images or texts of mathematical or arithmetic problems.</center>"""
116
+ )
117
+ state = gr.State({"tab_index": 0})
118
+ with gr.Row():
119
+ with gr.Column():
120
+ with gr.Tabs() as input_tabs:
121
+ with gr.Tab("Upload"):
122
+ input_image = gr.Image(type="pil", label="Upload"),
123
+ with gr.Tab("Sketch"):
124
+ input_sketchpad = gr.Sketchpad(type="pil", label="Sketch", layers=False)
125
+ input_tabs.select(fn=tabs_select, inputs=[state])
126
+ input_text = gr.Textbox(label="input your question")
127
+ with gr.Row():
128
+ with gr.Column():
129
+ clear_btn = gr.ClearButton(
130
+ [*input_image, input_sketchpad, input_text])
131
+ with gr.Column():
132
+ submit_btn = gr.Button("Submit", variant="primary")
133
+ with gr.Column():
134
+ output_md = gr.Markdown(label="answer",
135
+ latex_delimiters=[{
136
+ "left": "\\(",
137
+ "right": "\\)",
138
+ "display": True
139
+ }, {
140
+ "left": "\\begin\{equation\}",
141
+ "right": "\\end\{equation\}",
142
+ "display": True
143
+ }, {
144
+ "left": "\\begin\{align\}",
145
+ "right": "\\end\{align\}",
146
+ "display": True
147
+ }, {
148
+ "left": "\\begin\{alignat\}",
149
+ "right": "\\end\{alignat\}",
150
+ "display": True
151
+ }, {
152
+ "left": "\\begin\{gather\}",
153
+ "right": "\\end\{gather\}",
154
+ "display": True
155
+ }, {
156
+ "left": "\\begin\{CD\}",
157
+ "right": "\\end\{CD\}",
158
+ "display": True
159
+ }, {
160
+ "left": "\\[",
161
+ "right": "\\]",
162
+ "display": True
163
+ }],
164
+ elem_id="qwen-md")
165
+ submit_btn.click(
166
+ fn=math_chat_bot,
167
+ inputs=[*input_image, input_sketchpad, input_text, state],
168
+ outputs=output_md)
169
+ demo.launch()