code
stringlengths
3
6.57k
sort_line(line)
sort(key=lambda x: x['boundingBox']['vertices'][0]['x'],reverse=False)
cell_ocr_word(lang, page_path, line,save_base_path,mode_height)
horzontal_merging(line['regions'])
coord_adjustment(page_path,line['regions'] ,save_base_path)
enumerate(dynamic_line)
correct_region(word)
get_coord(word)
len(coord)
abs(coord_crop[1,1]-coord_crop[2,1])
get_text(page_path, coord_crop, lang,mode_height,save_base_path,8)
conf_dicts.extend(conf_dict)
cell_text_ocr(lang, page_path, line,save_base_path,mode_height)
horzontal_merging(line['regions'])
enumerate(line['regions'])
correct_region(word)
get_coord(word)
len(coord)
abs(coord_crop[1,1]-coord_crop[2,1])
get_text(page_path, coord_crop, lang,mode_height,save_base_path,8)
cell_ocr(lang, page_path, line,save_base_path,mode_height,psm)
horzontal_merging(line['regions'])
coord_adjustment(page_path,updated_lines ,save_base_path)
correct_region(updated_line)
get_coord(corrected_line)
len(coord)
abs(coord_crop[1,1]-coord_crop[2,1])
get_text(page_path, coord_crop, lang,mode_height,save_base_path,psm)
conf_dicts.extend(conf_dict)
text_extraction(df,lang, page_path, regions,save_base_path)
enumerate(regions)
frequent_height(level['regions'])
enumerate(level['regions'])
coord_adjustment(page_path, line['regions'],save_base_path)
enumerate(line['regions'])
correct_region(word)
get_coord(word)
len(word_text)
len(coord)
abs(coord_crop[1,1]-coord_crop[2,1])
get_text(page_path, coord_crop, lang,mode_height,save_base_path,8)
text.split()
join(text_list)
seq_matcher(text,word['text'])
len(remove_space(word['text'])
word.keys()
pd.DataFrame([word],columns=columns)
df.append(df2, ignore_index=True)
len(word_text)
seq_matcher("",word['text'])
word.keys()
pd.DataFrame([word],columns=columns)
df.append(df2, ignore_index=True)
coord_adjustment(page_path, level['regions'],save_base_path)
enumerate(lines_adjusted)
copy.deepcopy(sort_line(line_org)
merger_text(line_sorted)
copy.deepcopy(correct_region(line_sorted)
print(line_text)
get_coord(line)
print("line text",line_text)
len(remove_space(line_text)
len(coord)
abs(coord_crop[1,1]-coord_crop[2,1])
line.keys()
cell_ocr(lang, page_path, line,save_base_path,mode_height,psm)
line.keys()
cell_ocr_word(lang, page_path, line,save_base_path,mode_height)
get_text(page_path, coord_crop, lang,mode_height,save_base_path,psm)
text.split()
join(text_list)
seq_matcher(text,line_text)
check_psm(page_path,coord_crop,lang,mode_height,save_base_path,psm,score,text,line_text,conf_dict)
len(remove_space(line_text)
line.keys()
pd.DataFrame([line],columns=columns)
df.append(df2, ignore_index=True)
len(remove_space(line_text)
seq_matcher("",line_text)
line.keys()
pd.DataFrame([line],columns=columns)
df.append(df2, ignore_index=True)
glob.glob(output_path+"/*/gv.json")
tesseract(json_files)
json_file.split('/')
split('.json')
json_file.split('/')
print("file name--------------------->>>>>>>>>>>>>>>>>>",pdf_name)
os.path.exists(base_path+pdf_name)
os.mkdir(base_path+pdf_name)
open(json_file,'r+')
json.load(f)
pd.DataFrame(columns=columns)
Draw(data,save_base_path,regions='regions')
len(data['outputs'][0]['pages'])
enumerate(data['outputs'][0]['pages'])
time.time()
print("processing started for page no. ",idx)
pd.DataFrame()
text_extraction(df,lang, page_path, regions,save_base_path)