| import os |
| import json |
| import google |
| import argparse |
|
|
| from glob import glob |
| from typing import Optional, Sequence |
|
|
| from google.api_core.client_options import ClientOptions |
| from google.cloud import documentai |
|
|
| from utils import read_file_paths, validate_json_save_path, load_json_file |
|
|
| CATEGORY_MAP = { |
| "paragraph": "paragraph", |
| "footer": "footer", |
| "header": "header", |
| "heading-1": "heading1", |
| "heading-2": "heading1", |
| "heading-3": "heading1", |
| "table": "table", |
| "title": "heading1" |
| } |
|
|
|
|
| class GoogleInference: |
| def __init__( |
| self, |
| save_path, |
| input_formats=[".pdf", ".jpg", ".jpeg", ".png", ".bmp", ".tiff", ".heic"] |
| ): |
| """Initialize the GoogleInference class |
| Args: |
| save_path (str): the json path to save the results |
| input_formats (list, optional): the supported file formats. |
| """ |
| self.project_id = os.getenv("GOOGLE_PROJECT_ID") or "" |
| self.processor_id = os.getenv("GOOGLE_PROCESSOR_ID") or "" |
| self.location = os.getenv("GOOGLE_LOCATION") or "" |
| self.endpoint = os.getenv("GOOGLE_ENDPOINT") or "" |
|
|
| if not all([self.project_id, self.processor_id, self.location, self.endpoint]): |
| raise ValueError("Please set the environment variables for Google Cloud") |
|
|
| self.processor_version = "rc" |
|
|
| validate_json_save_path(save_path) |
| self.save_path = save_path |
| self.processed_data = load_json_file(save_path) |
|
|
| self.formats = input_formats |
|
|
| @staticmethod |
| def generate_html_table(table_data): |
| html = "<table border='1'>\n" |
|
|
| |
| for row in table_data["bodyRows"]: |
| html += " <tr>\n" |
| for cell in row["cells"]: |
| text = cell["blocks"][0]["textBlock"]["text"] if cell["blocks"] else "" |
| row_span = f" rowspan='{cell['rowSpan']}'" if cell["rowSpan"] > 1 else "" |
| col_span = f" colspan='{cell['colSpan']}'" if cell["colSpan"] > 1 else "" |
| html += f" <td{row_span}{col_span}>{text}</td>\n" |
| html += " </tr>\n" |
|
|
| html += "</table>" |
| return html |
|
|
| @staticmethod |
| def iterate_blocks(data): |
| block_sequence = [] |
|
|
| def recurse_blocks(blocks): |
| for block in blocks: |
| block_id = block.get("blockId", "") |
| block_type = block.get("textBlock", {}).get("type", "") |
| block_text = block.get("textBlock", {}).get("text", "") |
|
|
| if block_type: |
| |
| block_sequence.append((block_id, block_type, block_text)) |
|
|
| block_id = block.get("blockId", "") |
| block_table = block.get("tableBlock", {}) |
|
|
| if block_table: |
| block_table_html = GoogleInference.generate_html_table(block_table) |
| block_sequence.append((block_id, "table", block_table_html)) |
|
|
| |
| if block.get("textBlock", {}).get("blocks", []): |
| recurse_blocks(block["textBlock"]["blocks"]) |
|
|
| if "documentLayout" in data: |
| recurse_blocks(data["documentLayout"].get("blocks", [])) |
|
|
| return block_sequence |
|
|
| def post_process(self, data): |
|
|
| processed_dict = {} |
| for input_key in data.keys(): |
| output_data = data[input_key] |
|
|
| processed_dict[input_key] = { |
| "elements": [] |
| } |
|
|
| blocks = self.iterate_blocks(output_data) |
|
|
| id_counter = 0 |
| for _, category, transcription in blocks: |
| category = CATEGORY_MAP.get(category, "paragraph") |
|
|
| data_dict = { |
| "coordinates": [[0, 0], [0, 0], [0, 0], [0, 0]], |
| "category": category, |
| "id": id_counter, |
| "content": { |
| "text": transcription if category != "table" else "", |
| "html": transcription if category == "table" else "", |
| "markdown": "" |
| } |
| } |
| processed_dict[input_key]["elements"].append(data_dict) |
|
|
| id_counter += 1 |
|
|
| for key in self.processed_data: |
| processed_dict[key] = self.processed_data[key] |
|
|
| return processed_dict |
|
|
| def process_document_layout_sample(self, file_path, mime_type, chunk_size=1000) -> None: |
| process_options = documentai.ProcessOptions( |
| layout_config=documentai.ProcessOptions.LayoutConfig( |
| chunking_config=documentai.ProcessOptions.LayoutConfig.ChunkingConfig( |
| chunk_size=chunk_size, |
| include_ancestor_headings=True, |
| ) |
| ) |
| ) |
| document = self.process_document( |
| file_path, |
| mime_type, |
| process_options=process_options, |
| ) |
|
|
| document_dict = json.loads(google.cloud.documentai_v1.Document.to_json(document)) |
|
|
| return document_dict |
|
|
| def process_document( |
| self, file_path, |
| mime_type: str, |
| process_options: Optional[documentai.ProcessOptions] = None, |
| ) -> documentai.Document: |
| client = documentai.DocumentProcessorServiceClient( |
| client_options=ClientOptions( |
| api_endpoint=f"{self.endpoint}" |
| ) |
| ) |
|
|
| with open(file_path, "rb") as image: |
| image_content = image.read() |
|
|
| name = client.processor_version_path( |
| self.project_id, |
| self.location, |
| self.processor_id, |
| self.processor_version |
| ) |
| request = documentai.ProcessRequest( |
| name=name, |
| raw_document=documentai.RawDocument( |
| content=image_content, mime_type=mime_type |
| ), |
| process_options=process_options, |
| ) |
|
|
| result = client.process_document(request=request) |
|
|
| return result.document |
|
|
| def infer(self, file_path): |
| """Infer the layout of the documents in the given file path |
| Args: |
| file_path (str): the path to the file or directory containing the documents to process |
| """ |
| paths = read_file_paths(file_path, supported_formats=self.formats) |
|
|
| error_files = [] |
|
|
| result_dict = {} |
| for idx, filepath in enumerate(paths): |
| print("({}/{}) {}".format(idx+1, len(paths), filepath)) |
|
|
| if filepath.suffix == ".pdf": |
| mime_type = "application/pdf" |
| elif filepath.suffix == ".jpg" or filepath.suffix == ".jpeg": |
| mime_type = "image/jpeg" |
| elif filepath.suffix == ".png": |
| mime_type = "image/png" |
| else: |
| raise NotImplementedError |
|
|
| filename = filepath.name |
|
|
| if filename in self.processed_data.keys(): |
| print(f"'{filename}' is already in the loaded dictionary. Skipping this sample") |
| continue |
|
|
| try: |
| document_dict = self.process_document_layout_sample(filepath, mime_type) |
| except Exception as e: |
| print(e) |
| print("Error processing document..") |
| error_files.append(filepath) |
| continue |
|
|
| result_dict[filename] = document_dict |
|
|
| result_dict = self.post_process(result_dict) |
|
|
| with open(self.save_path, "w") as f: |
| json.dump(result_dict, f) |
|
|
| for error_file in error_files: |
| print(f"Error processing file: {error_file}") |
|
|
| print("Finished processing all documents") |
| print("Results saved to: {}".format(self.save_path)) |
| print("Number of errors: {}".format(len(error_files))) |
|
|
|
|
| if __name__ == "__main__": |
| args = argparse.ArgumentParser() |
| args.add_argument( |
| "--data_path", |
| type=str, default="", required=True, |
| help="Path containing the documents to process" |
| ) |
| args.add_argument( |
| "--save_path", |
| type=str, default="", required=True, |
| help="Path to save the results" |
| ) |
| args.add_argument( |
| "--input_formats", |
| type=list, default=[ |
| ".pdf", ".jpg", ".jpeg", ".png", ".bmp", ".tiff", ".heic" |
| ], |
| help="Supported input file formats" |
| ) |
| args = args.parse_args() |
|
|
| google_inference = GoogleInference( |
| args.save_path, |
| input_formats=args.input_formats |
| ) |
| google_inference.infer(args.data_path) |
|
|