--- license: other pretty_name: "Fintabnet-Logical" tags: - table-structure-recognition - table-understanding - document-ai - computer-vision - finance --- # Fintabnet-Logical ## Dataset Summary **Fintabnet-Logical** is a derivative of the original [FinTabNet](https://developer.adobe.com/document-services/docs/overview/pdf-extract-api/fintabnet-dataset/) dataset, specifically re-processed to create high-quality ground truth for logical table structure recognition (TSR). While the original dataset provides cell content and HTML structure, this version parses that HTML to generate precise **logical coordinates** for every cell, correctly handling complex tables with `rowspan` and `colspan`. Furthermore, it processes the source PDFs to group text into **line-level cells**, assigning each line the logical coordinates of its parent cell. The result is a clean, ready-to-use dataset for training models that predict not just the content of a table, but its fundamental logical grid structure. All table images are provided as high-resolution (144 DPI) crops for improved visual quality. ## Supported Tasks * **Table Structure Recognition**: This dataset is primarily designed for training and evaluating models that recognize the logical row and column structure of tables, including row and column spans. The line-level cells with logical coordinates are ideal for this task. ## Dataset Structure The dataset is organized into `train`, `val`, and `test` splits, mirroring the original FinTabNet. Each instance consists of a table image and a corresponding JSON annotation file. ### Data Instances A typical annotation file (`.json`) has the following structure: ```json { "fintabnet_annotations": { "... original fintabnet data ..." }, "fintabnet_cells": [ { "bbox": [187.0, 4.0, 261.0, 14.0], "tokens": ["...", "Practitioners", "..."], "logical_coords": [0, 0, 1, 5] } ], "word_cells": [ { "text": "Practitioners", "bbox": [187.0, 4.0, 261.0, 14.0], "logical_coords": [0, 0, 1, 5] } ], "line_cells": [ { "text": "General Practitioners", "bbox": [187.0, 4.0, 261.0, 14.0], "logical_coords": [0, 0, 1, 5] }, { "text": "1. Antipsychotic drug treatment", "bbox": [4.0, 58.0, 133.0, 86.0], "logical_coords": [2, 2, 0, 0] } ] } ``` ### Data Fields The most important key for training is `line_cells`: * `line_cells`: A list of dictionaries, where each entry represents a single line of text within a table cell. * `text` (`str`): The text content of the line. * `bbox` (`list[float]`): The bounding box of the text line, in `[x_min, y_min, x_max, y_max]` format relative to the cropped table image. * `logical_coords` (`list[int]`): The logical coordinates of the parent cell in `[row_start, row_end, col_start, col_end]` format. An unspanned cell at the top-left would be `[0, 0, 0, 0]`. A cell spanning the first two rows in the first column would be `[0, 1, 0, 0]`. ### Data Splits The dataset retains the original splits from FinTabNet: | Split | Number of Tables | |--------------|------------------| | **train** | 82,422 | | **validation** | 9,539 | | **test** | 9,599 | | **Total** | **101,560** | ## Dataset Creation ### Curation Rationale Many table recognition datasets provide only bounding boxes for cells, without the explicit logical row/column indices needed to understand the grid structure. This dataset was created to fill that gap. By parsing the HTML structure provided by FinTabNet, we generate a reliable ground truth for logical coordinates, which is invaluable for training and evaluating modern Table Structure Recognition models. ### Source Data This dataset is derived from the **FinTabNet** dataset, which consists of tables from the annual financial reports of S&P 500 companies. ### Annotations The annotation process is fully automated by a script that performs the following steps for each table: 1. **Parse HTML**: The `structure` tokens from the original annotations are parsed to build a virtual grid of the table. 2. **Calculate Logical Coordinates**: By traversing the virtual grid, the script calculates the `[row_start, row_end, col_start, col_end]` for every cell, accurately accounting for `rowspan` and `colspan` attributes. 3. **Extract Words**: The source PDF is processed to extract all words and their bounding boxes within the table region. 4. **Group into Lines**: Words are assigned to their parent cells based on spatial overlap. Within each cell, the words are grouped into lines based on reading order. 5. **Assign Coordinates to Lines**: Each generated line is assigned the logical coordinates of its parent cell, creating the final `line_cells` ground truth. --- ## Citation If you use this dataset, please cite the original FinTabNet paper: ```bibtex @article{zheng2021global, title={Global table extractor (gte): A framework for joint table identification and cell structure recognition using visual context}, author={Zheng, Xinyi and Burdick, Doug and Popa, Lucian and Sthankiya, Shachi and Teslee, Mitchell and Thomas, Bibin}, journal={arXiv preprint arXiv:2109.04946}, year={2021} } ```