| import os |
| import time |
| import json |
| import argparse |
| from pathlib import Path |
|
|
| import unstructured_client |
| from unstructured_client.models import operations, shared |
|
|
| from utils import read_file_paths, validate_json_save_path, load_json_file |
|
|
|
|
| CATEGORY_MAP = { |
| "NarrativeText": "paragraph", |
| "ListItem": "paragraph", |
| "Title": "heading1", |
| "Address": "paragraph", |
| "Header": "header", |
| "Footer": "footer", |
| "UncategorizedText": "paragraph", |
| "Formula": "equation", |
| "FigureCaption": "caption", |
| "Table": "table", |
| "PageBreak": "paragraph", |
| "Image": "figure", |
| "PageNumber": "paragraph", |
| "CodeSnippet": "paragraph" |
| } |
|
|
|
|
| class UnstructuredInference: |
| def __init__( |
| self, |
| save_path, |
| input_formats=[".pdf", ".jpg", ".jpeg", ".png", ".bmp", ".tiff", ".heic"] |
| ): |
| """Initialize the UnstructuredInference class |
| Args: |
| save_path (str): the json path to save the results |
| input_formats (list, optional): the supported file formats. |
| """ |
| self.formats = input_formats |
|
|
| self.api_key = os.getenv("UNSTRUCTURED_API_KEY") or "" |
| self.url = os.getenv("UNSTRUCTURED_URL") or "" |
|
|
| if not self.api_key or not self.url: |
| raise ValueError("Please set the environment variables for Unstructured") |
|
|
| self.languages = ["eng", "kor"] |
| self.get_coordinates = True |
| self.infer_table_structure = True |
|
|
| |
| validate_json_save_path(save_path) |
| self.save_path = save_path |
| self.processed_data = load_json_file(save_path) |
|
|
| self.client = unstructured_client.UnstructuredClient( |
| api_key_auth=self.api_key, |
| server_url=self.url, |
| ) |
|
|
| def post_process(self, data): |
| processed_dict = {} |
| for input_key in data.keys(): |
| output_data = data[input_key] |
|
|
| processed_dict[input_key] = { |
| "elements": [] |
| } |
|
|
| id_counter = 0 |
| for elem in output_data: |
| transcription = elem["text"] |
| category = CATEGORY_MAP.get(elem["type"], "paragraph") |
| if elem["metadata"]["coordinates"] is None: |
| continue |
|
|
| xy_coord = [{"x": x, "y": y} for x, y in elem["metadata"]["coordinates"]["points"]] |
|
|
| if category == "table": |
| transcription = elem["metadata"]["text_as_html"] |
|
|
| data_dict = { |
| "coordinates": xy_coord, |
| "category": category, |
| "id": id_counter, |
| "content": { |
| "text": str(transcription) if category != "table" else "", |
| "html": transcription if category == "table" else "", |
| "markdown": "" |
| } |
| } |
| processed_dict[input_key]["elements"].append(data_dict) |
|
|
| id_counter += 1 |
|
|
| for key in self.processed_data: |
| processed_dict[key] = self.processed_data[key] |
|
|
| return processed_dict |
|
|
| def infer(self, file_path): |
| """Infer the layout of the documents in the given file path |
| Args: |
| file_path (str): the path to the file or directory containing the documents to process |
| """ |
| paths = read_file_paths(file_path, supported_formats=self.formats) |
|
|
| error_files = [] |
|
|
| result_dict = {} |
| for filepath in paths: |
| print("({}/{}) Processing {}".format(paths.index(filepath) + 1, len(paths), filepath)) |
| filename = filepath.name |
| if filename in self.processed_data.keys(): |
| print(f"'{filename}' is already in the loaded dictionary. Skipping this sample") |
| continue |
|
|
| with open(filepath, "rb") as f: |
| data = f.read() |
|
|
| req = operations.PartitionRequest( |
| partition_parameters=shared.PartitionParameters( |
| files=shared.Files( |
| content=data, |
| file_name=str(filepath), |
| ), |
| |
| strategy=shared.Strategy.HI_RES, |
| pdf_infer_table_structure=self.infer_table_structure, |
| coordinates=self.get_coordinates, |
| languages=self.languages, |
| ), |
| ) |
|
|
| try: |
| res = self.client.general.partition(request=req) |
| elements = res.elements |
| except Exception as e: |
| print(e) |
| print("Error processing document..") |
| error_files.append(filepath) |
| continue |
|
|
| result_dict[filename] = elements |
|
|
| result_dict = self.post_process(result_dict) |
|
|
| with open(self.save_path, "w") as f: |
| json.dump(result_dict, f) |
|
|
| for error_file in error_files: |
| print(f"Error processing file: {error_file}") |
|
|
| print("Finished processing all documents") |
| print("Results saved to: {}".format(self.save_path)) |
| print("Number of errors: {}".format(len(error_files))) |
|
|
|
|
| if __name__ == "__main__": |
| args = argparse.ArgumentParser() |
| args.add_argument( |
| "--data_path", |
| type=str, default="", required=True, |
| help="Path containing the documents to process" |
| ) |
| args.add_argument( |
| "--save_path", |
| type=str, default="", required=True, |
| help="Path to save the results" |
| ) |
| args.add_argument( |
| "--input_formats", |
| type=list, default=[ |
| ".pdf", ".jpg", ".jpeg", ".png", ".bmp", ".tiff", ".heic" |
| ], |
| help="Supported input file formats" |
| ) |
| args = args.parse_args() |
|
|
| unstructured_inference = UnstructuredInference( |
| args.save_path, |
| input_formats=args.input_formats |
| ) |
| unstructured_inference.infer(args.data_path) |
|
|
|
|