JoseArcosO commited on
Commit
ab1a626
·
1 Parent(s): 9e1fa46
.ipynb_checkpoints/app-checkpoint.py CHANGED
@@ -3,10 +3,18 @@ import gradio as gr
3
  import re
4
  import json
5
  import os
6
-
7
  from transformers import DonutProcessor, VisionEncoderDecoderModel
8
  from datasets import load_dataset
9
  import torch
 
 
 
 
 
 
 
 
10
 
11
  def get_attributes(input_img):
12
  access_token = str(os.environ.get('key'))
@@ -17,8 +25,9 @@ def get_attributes(input_img):
17
 
18
  model.eval()
19
  model.to(device)
20
-
21
- pixel_values = processor(input_img, return_tensors="pt").pixel_values
 
22
  pixel_values = pixel_values.to(device)
23
  print(pixel_values.size())
24
  # prepare decoder inputs
 
3
  import re
4
  import json
5
  import os
6
+ from pdf2image import convert_from_bytes
7
  from transformers import DonutProcessor, VisionEncoderDecoderModel
8
  from datasets import load_dataset
9
  import torch
10
+ import imghdr
11
+
12
+ def check_image(image):
13
+ if imghdr.what(image) == 'jpeg' or imghdr.what(image) == 'png':
14
+ return image
15
+ else:
16
+ images = convert_from_bytes(image.read(), fmt="jpeg", size=(960,1280))
17
+ return images[0]
18
 
19
  def get_attributes(input_img):
20
  access_token = str(os.environ.get('key'))
 
25
 
26
  model.eval()
27
  model.to(device)
28
+
29
+ image = check_image(input_img)
30
+ pixel_values = processor(image, return_tensors="pt").pixel_values
31
  pixel_values = pixel_values.to(device)
32
  print(pixel_values.size())
33
  # prepare decoder inputs
.ipynb_checkpoints/requirements-checkpoint.txt CHANGED
@@ -1,4 +1,6 @@
1
  torch
2
  transformers
3
  numpy
4
- datasets
 
 
 
1
  torch
2
  transformers
3
  numpy
4
+ datasets
5
+ pdf2image
6
+ imghdr
app.py CHANGED
@@ -3,10 +3,18 @@ import gradio as gr
3
  import re
4
  import json
5
  import os
6
-
7
  from transformers import DonutProcessor, VisionEncoderDecoderModel
8
  from datasets import load_dataset
9
  import torch
 
 
 
 
 
 
 
 
10
 
11
  def get_attributes(input_img):
12
  access_token = str(os.environ.get('key'))
@@ -17,8 +25,9 @@ def get_attributes(input_img):
17
 
18
  model.eval()
19
  model.to(device)
20
-
21
- pixel_values = processor(input_img, return_tensors="pt").pixel_values
 
22
  pixel_values = pixel_values.to(device)
23
  print(pixel_values.size())
24
  # prepare decoder inputs
 
3
  import re
4
  import json
5
  import os
6
+ from pdf2image import convert_from_bytes
7
  from transformers import DonutProcessor, VisionEncoderDecoderModel
8
  from datasets import load_dataset
9
  import torch
10
+ import imghdr
11
+
12
+ def check_image(image):
13
+ if imghdr.what(image) == 'jpeg' or imghdr.what(image) == 'png':
14
+ return image
15
+ else:
16
+ images = convert_from_bytes(image.read(), fmt="jpeg", size=(960,1280))
17
+ return images[0]
18
 
19
  def get_attributes(input_img):
20
  access_token = str(os.environ.get('key'))
 
25
 
26
  model.eval()
27
  model.to(device)
28
+
29
+ image = check_image(input_img)
30
+ pixel_values = processor(image, return_tensors="pt").pixel_values
31
  pixel_values = pixel_values.to(device)
32
  print(pixel_values.size())
33
  # prepare decoder inputs
requirements.txt CHANGED
@@ -1,4 +1,6 @@
1
  torch
2
  transformers
3
  numpy
4
- datasets
 
 
 
1
  torch
2
  transformers
3
  numpy
4
+ datasets
5
+ pdf2image
6
+ imghdr