code stringlengths 3 6.57k |
|---|
sort_line(line) |
sort(key=lambda x: x['boundingBox']['vertices'][0]['x'],reverse=False) |
cell_ocr_word(lang, page_path, line,save_base_path,mode_height) |
horzontal_merging(line['regions']) |
coord_adjustment(page_path,line['regions'] ,save_base_path) |
enumerate(dynamic_line) |
correct_region(word) |
get_coord(word) |
len(coord) |
abs(coord_crop[1,1]-coord_crop[2,1]) |
get_text(page_path, coord_crop, lang,mode_height,save_base_path,8) |
conf_dicts.extend(conf_dict) |
cell_text_ocr(lang, page_path, line,save_base_path,mode_height) |
horzontal_merging(line['regions']) |
enumerate(line['regions']) |
correct_region(word) |
get_coord(word) |
len(coord) |
abs(coord_crop[1,1]-coord_crop[2,1]) |
get_text(page_path, coord_crop, lang,mode_height,save_base_path,8) |
cell_ocr(lang, page_path, line,save_base_path,mode_height,psm) |
horzontal_merging(line['regions']) |
coord_adjustment(page_path,updated_lines ,save_base_path) |
correct_region(updated_line) |
get_coord(corrected_line) |
len(coord) |
abs(coord_crop[1,1]-coord_crop[2,1]) |
get_text(page_path, coord_crop, lang,mode_height,save_base_path,psm) |
conf_dicts.extend(conf_dict) |
text_extraction(df,lang, page_path, regions,save_base_path) |
enumerate(regions) |
frequent_height(level['regions']) |
enumerate(level['regions']) |
coord_adjustment(page_path, line['regions'],save_base_path) |
enumerate(line['regions']) |
correct_region(word) |
get_coord(word) |
len(word_text) |
len(coord) |
abs(coord_crop[1,1]-coord_crop[2,1]) |
get_text(page_path, coord_crop, lang,mode_height,save_base_path,8) |
text.split() |
join(text_list) |
seq_matcher(text,word['text']) |
len(remove_space(word['text']) |
word.keys() |
pd.DataFrame([word],columns=columns) |
df.append(df2, ignore_index=True) |
len(word_text) |
seq_matcher("",word['text']) |
word.keys() |
pd.DataFrame([word],columns=columns) |
df.append(df2, ignore_index=True) |
coord_adjustment(page_path, level['regions'],save_base_path) |
enumerate(lines_adjusted) |
copy.deepcopy(sort_line(line_org) |
merger_text(line_sorted) |
copy.deepcopy(correct_region(line_sorted) |
print(line_text) |
get_coord(line) |
print("line text",line_text) |
len(remove_space(line_text) |
len(coord) |
abs(coord_crop[1,1]-coord_crop[2,1]) |
line.keys() |
cell_ocr(lang, page_path, line,save_base_path,mode_height,psm) |
line.keys() |
cell_ocr_word(lang, page_path, line,save_base_path,mode_height) |
get_text(page_path, coord_crop, lang,mode_height,save_base_path,psm) |
text.split() |
join(text_list) |
seq_matcher(text,line_text) |
check_psm(page_path,coord_crop,lang,mode_height,save_base_path,psm,score,text,line_text,conf_dict) |
len(remove_space(line_text) |
line.keys() |
pd.DataFrame([line],columns=columns) |
df.append(df2, ignore_index=True) |
len(remove_space(line_text) |
seq_matcher("",line_text) |
line.keys() |
pd.DataFrame([line],columns=columns) |
df.append(df2, ignore_index=True) |
glob.glob(output_path+"/*/gv.json") |
tesseract(json_files) |
json_file.split('/') |
split('.json') |
json_file.split('/') |
print("file name--------------------->>>>>>>>>>>>>>>>>>",pdf_name) |
os.path.exists(base_path+pdf_name) |
os.mkdir(base_path+pdf_name) |
open(json_file,'r+') |
json.load(f) |
pd.DataFrame(columns=columns) |
Draw(data,save_base_path,regions='regions') |
len(data['outputs'][0]['pages']) |
enumerate(data['outputs'][0]['pages']) |
time.time() |
print("processing started for page no. ",idx) |
pd.DataFrame() |
text_extraction(df,lang, page_path, regions,save_base_path) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.