Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -10,36 +10,32 @@ model_name = "google/matcha-base"
|
|
| 10 |
model = Pix2StructForConditionalGeneration.from_pretrained(model_name)
|
| 11 |
processor = Pix2StructProcessor.from_pretrained(model_name)
|
| 12 |
|
| 13 |
-
# Move model to GPU if available
|
| 14 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 15 |
model.to(device)
|
|
|
|
| 16 |
|
| 17 |
def solve_math_problem(image):
|
| 18 |
-
# Preprocess the image and include a
|
| 19 |
-
# You can adjust the prompt to better match your task if needed.
|
| 20 |
inputs = processor(images=image, text="Solve the math problem:", return_tensors="pt")
|
| 21 |
-
#
|
| 22 |
inputs = {key: value.to(device) for key, value in inputs.items()}
|
| 23 |
|
| 24 |
-
# Generate the solution using beam search
|
| 25 |
-
|
| 26 |
-
|
| 27 |
-
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
num_beams=5,
|
| 34 |
-
early_stopping=True,
|
| 35 |
-
temperature=0.5
|
| 36 |
-
)
|
| 37 |
|
| 38 |
-
# Decode the
|
| 39 |
solution = processor.decode(predictions[0], skip_special_tokens=True)
|
| 40 |
return solution
|
| 41 |
|
| 42 |
-
# Set up
|
| 43 |
demo = gr.Interface(
|
| 44 |
fn=solve_math_problem,
|
| 45 |
inputs=gr.Image(type="pil", label="Upload Handwritten Math Problem"),
|
|
@@ -51,3 +47,4 @@ demo = gr.Interface(
|
|
| 51 |
|
| 52 |
if __name__ == "__main__":
|
| 53 |
demo.launch()
|
|
|
|
|
|
| 10 |
model = Pix2StructForConditionalGeneration.from_pretrained(model_name)
|
| 11 |
processor = Pix2StructProcessor.from_pretrained(model_name)
|
| 12 |
|
| 13 |
+
# Move model to GPU if available and set to evaluation mode
|
| 14 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
| 15 |
model.to(device)
|
| 16 |
+
model.eval()
|
| 17 |
|
| 18 |
def solve_math_problem(image):
|
| 19 |
+
# Preprocess the image and include a prompt.
|
|
|
|
| 20 |
inputs = processor(images=image, text="Solve the math problem:", return_tensors="pt")
|
| 21 |
+
# Move all tensors to the same device as the model
|
| 22 |
inputs = {key: value.to(device) for key, value in inputs.items()}
|
| 23 |
|
| 24 |
+
# Generate the solution using beam search within a no_grad context
|
| 25 |
+
with torch.no_grad():
|
| 26 |
+
predictions = model.generate(
|
| 27 |
+
**inputs,
|
| 28 |
+
max_new_tokens=150, # Increase this if longer answers are needed
|
| 29 |
+
num_beams=5, # Beam search for more stable outputs
|
| 30 |
+
early_stopping=True,
|
| 31 |
+
temperature=0.5 # Lower temperature for more deterministic output
|
| 32 |
+
)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
+
# Decode the generated tokens to a string, skipping special tokens
|
| 35 |
solution = processor.decode(predictions[0], skip_special_tokens=True)
|
| 36 |
return solution
|
| 37 |
|
| 38 |
+
# Set up the Gradio interface
|
| 39 |
demo = gr.Interface(
|
| 40 |
fn=solve_math_problem,
|
| 41 |
inputs=gr.Image(type="pil", label="Upload Handwritten Math Problem"),
|
|
|
|
| 47 |
|
| 48 |
if __name__ == "__main__":
|
| 49 |
demo.launch()
|
| 50 |
+
|