Spaces:
Runtime error
Runtime error
Commit ·
ab1a626
1
Parent(s): 9e1fa46
Update
Browse files- .ipynb_checkpoints/app-checkpoint.py +12 -3
- .ipynb_checkpoints/requirements-checkpoint.txt +3 -1
- app.py +12 -3
- requirements.txt +3 -1
.ipynb_checkpoints/app-checkpoint.py
CHANGED
|
@@ -3,10 +3,18 @@ import gradio as gr
|
|
| 3 |
import re
|
| 4 |
import json
|
| 5 |
import os
|
| 6 |
-
|
| 7 |
from transformers import DonutProcessor, VisionEncoderDecoderModel
|
| 8 |
from datasets import load_dataset
|
| 9 |
import torch
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
|
| 11 |
def get_attributes(input_img):
|
| 12 |
access_token = str(os.environ.get('key'))
|
|
@@ -17,8 +25,9 @@ def get_attributes(input_img):
|
|
| 17 |
|
| 18 |
model.eval()
|
| 19 |
model.to(device)
|
| 20 |
-
|
| 21 |
-
|
|
|
|
| 22 |
pixel_values = pixel_values.to(device)
|
| 23 |
print(pixel_values.size())
|
| 24 |
# prepare decoder inputs
|
|
|
|
| 3 |
import re
|
| 4 |
import json
|
| 5 |
import os
|
| 6 |
+
from pdf2image import convert_from_bytes
|
| 7 |
from transformers import DonutProcessor, VisionEncoderDecoderModel
|
| 8 |
from datasets import load_dataset
|
| 9 |
import torch
|
| 10 |
+
import imghdr
|
| 11 |
+
|
| 12 |
+
def check_image(image):
|
| 13 |
+
if imghdr.what(image) == 'jpeg' or imghdr.what(image) == 'png':
|
| 14 |
+
return image
|
| 15 |
+
else:
|
| 16 |
+
images = convert_from_bytes(image.read(), fmt="jpeg", size=(960,1280))
|
| 17 |
+
return images[0]
|
| 18 |
|
| 19 |
def get_attributes(input_img):
|
| 20 |
access_token = str(os.environ.get('key'))
|
|
|
|
| 25 |
|
| 26 |
model.eval()
|
| 27 |
model.to(device)
|
| 28 |
+
|
| 29 |
+
image = check_image(input_img)
|
| 30 |
+
pixel_values = processor(image, return_tensors="pt").pixel_values
|
| 31 |
pixel_values = pixel_values.to(device)
|
| 32 |
print(pixel_values.size())
|
| 33 |
# prepare decoder inputs
|
.ipynb_checkpoints/requirements-checkpoint.txt
CHANGED
|
@@ -1,4 +1,6 @@
|
|
| 1 |
torch
|
| 2 |
transformers
|
| 3 |
numpy
|
| 4 |
-
datasets
|
|
|
|
|
|
|
|
|
| 1 |
torch
|
| 2 |
transformers
|
| 3 |
numpy
|
| 4 |
+
datasets
|
| 5 |
+
pdf2image
|
| 6 |
+
imghdr
|
app.py
CHANGED
|
@@ -3,10 +3,18 @@ import gradio as gr
|
|
| 3 |
import re
|
| 4 |
import json
|
| 5 |
import os
|
| 6 |
-
|
| 7 |
from transformers import DonutProcessor, VisionEncoderDecoderModel
|
| 8 |
from datasets import load_dataset
|
| 9 |
import torch
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
|
| 11 |
def get_attributes(input_img):
|
| 12 |
access_token = str(os.environ.get('key'))
|
|
@@ -17,8 +25,9 @@ def get_attributes(input_img):
|
|
| 17 |
|
| 18 |
model.eval()
|
| 19 |
model.to(device)
|
| 20 |
-
|
| 21 |
-
|
|
|
|
| 22 |
pixel_values = pixel_values.to(device)
|
| 23 |
print(pixel_values.size())
|
| 24 |
# prepare decoder inputs
|
|
|
|
| 3 |
import re
|
| 4 |
import json
|
| 5 |
import os
|
| 6 |
+
from pdf2image import convert_from_bytes
|
| 7 |
from transformers import DonutProcessor, VisionEncoderDecoderModel
|
| 8 |
from datasets import load_dataset
|
| 9 |
import torch
|
| 10 |
+
import imghdr
|
| 11 |
+
|
| 12 |
+
def check_image(image):
|
| 13 |
+
if imghdr.what(image) == 'jpeg' or imghdr.what(image) == 'png':
|
| 14 |
+
return image
|
| 15 |
+
else:
|
| 16 |
+
images = convert_from_bytes(image.read(), fmt="jpeg", size=(960,1280))
|
| 17 |
+
return images[0]
|
| 18 |
|
| 19 |
def get_attributes(input_img):
|
| 20 |
access_token = str(os.environ.get('key'))
|
|
|
|
| 25 |
|
| 26 |
model.eval()
|
| 27 |
model.to(device)
|
| 28 |
+
|
| 29 |
+
image = check_image(input_img)
|
| 30 |
+
pixel_values = processor(image, return_tensors="pt").pixel_values
|
| 31 |
pixel_values = pixel_values.to(device)
|
| 32 |
print(pixel_values.size())
|
| 33 |
# prepare decoder inputs
|
requirements.txt
CHANGED
|
@@ -1,4 +1,6 @@
|
|
| 1 |
torch
|
| 2 |
transformers
|
| 3 |
numpy
|
| 4 |
-
datasets
|
|
|
|
|
|
|
|
|
| 1 |
torch
|
| 2 |
transformers
|
| 3 |
numpy
|
| 4 |
+
datasets
|
| 5 |
+
pdf2image
|
| 6 |
+
imghdr
|