| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """ |
| Convert bounding box formats in a Hugging Face object detection dataset. |
| |
| Mirrors panlabel's convert command. Converts between: |
| - COCO xywh: [x, y, width, height] in pixels |
| - XYXY: [xmin, ymin, xmax, ymax] in pixels |
| - VOC: [xmin, ymin, xmax, ymax] in pixels (alias for xyxy) |
| - YOLO: [center_x, center_y, width, height] normalized 0-1 |
| - TFOD: [xmin, ymin, xmax, ymax] normalized 0-1 |
| - Label Studio: [x, y, width, height] percentage 0-100 |
| |
| Reads from HF Hub, converts bboxes in-place, and pushes the result to a new |
| (or the same) dataset repo on HF Hub. |
| |
| Examples: |
| uv run convert-hf-dataset.py merve/coco-dataset merve/coco-xyxy --from coco_xywh --to xyxy |
| uv run convert-hf-dataset.py merve/yolo-dataset merve/yolo-coco --from yolo --to coco_xywh |
| uv run convert-hf-dataset.py merve/dataset merve/converted --from xyxy --to yolo --max-samples 100 |
| uv run convert-hf-dataset.py merve/dataset merve/converted --from tfod --to coco_xywh |
| uv run convert-hf-dataset.py merve/dataset merve/converted --from label_studio --to xyxy |
| """ |
|
|
| import argparse |
| import json |
| import logging |
| import os |
| import sys |
| import time |
| from datetime import datetime |
| from typing import Any |
|
|
| from datasets import load_dataset |
| from huggingface_hub import DatasetCard, login |
| from toolz import partition_all |
| from tqdm.auto import tqdm |
|
|
| logging.basicConfig(level=logging.INFO) |
| logger = logging.getLogger(__name__) |
|
|
| BBOX_FORMATS = ["coco_xywh", "xyxy", "voc", "yolo", "tfod", "label_studio"] |
|
|
|
|
| def convert_bbox( |
| bbox: list[float], |
| from_fmt: str, |
| to_fmt: str, |
| img_w: float = 1.0, |
| img_h: float = 1.0, |
| ) -> list[float]: |
| """Convert a single bbox between formats via XYXY pixel-space intermediate.""" |
| |
| if from_fmt == "coco_xywh": |
| x, y, w, h = bbox[:4] |
| xmin, ymin, xmax, ymax = x, y, x + w, y + h |
| elif from_fmt in ("xyxy", "voc"): |
| xmin, ymin, xmax, ymax = bbox[:4] |
| elif from_fmt == "yolo": |
| cx, cy, w, h = bbox[:4] |
| xmin = (cx - w / 2) * img_w |
| ymin = (cy - h / 2) * img_h |
| xmax = (cx + w / 2) * img_w |
| ymax = (cy + h / 2) * img_h |
| elif from_fmt == "tfod": |
| xmin_n, ymin_n, xmax_n, ymax_n = bbox[:4] |
| xmin = xmin_n * img_w |
| ymin = ymin_n * img_h |
| xmax = xmax_n * img_w |
| ymax = ymax_n * img_h |
| elif from_fmt == "label_studio": |
| x_pct, y_pct, w_pct, h_pct = bbox[:4] |
| xmin = x_pct / 100.0 * img_w |
| ymin = y_pct / 100.0 * img_h |
| xmax = (x_pct + w_pct) / 100.0 * img_w |
| ymax = (y_pct + h_pct) / 100.0 * img_h |
| else: |
| raise ValueError(f"Unknown source format: {from_fmt}") |
|
|
| |
| if to_fmt in ("xyxy", "voc"): |
| return [xmin, ymin, xmax, ymax] |
| elif to_fmt == "coco_xywh": |
| return [xmin, ymin, xmax - xmin, ymax - ymin] |
| elif to_fmt == "yolo": |
| if img_w <= 0 or img_h <= 0: |
| raise ValueError("YOLO format requires positive image dimensions") |
| w = xmax - xmin |
| h = ymax - ymin |
| cx = (xmin + w / 2) / img_w |
| cy = (ymin + h / 2) / img_h |
| return [cx, cy, w / img_w, h / img_h] |
| elif to_fmt == "tfod": |
| if img_w <= 0 or img_h <= 0: |
| raise ValueError("TFOD format requires positive image dimensions") |
| return [xmin / img_w, ymin / img_h, xmax / img_w, ymax / img_h] |
| elif to_fmt == "label_studio": |
| if img_w <= 0 or img_h <= 0: |
| raise ValueError("Label Studio format requires positive image dimensions") |
| x_pct = xmin / img_w * 100.0 |
| y_pct = ymin / img_h * 100.0 |
| w_pct = (xmax - xmin) / img_w * 100.0 |
| h_pct = (ymax - ymin) / img_h * 100.0 |
| return [x_pct, y_pct, w_pct, h_pct] |
| else: |
| raise ValueError(f"Unknown target format: {to_fmt}") |
|
|
|
|
| def convert_example( |
| example: dict[str, Any], |
| bbox_column: str, |
| from_fmt: str, |
| to_fmt: str, |
| width_column: str | None, |
| height_column: str | None, |
| ) -> dict[str, Any]: |
| """Convert bboxes in a single example.""" |
| objects = example.get("objects") |
| is_nested = objects is not None and isinstance(objects, dict) |
|
|
| if is_nested: |
| bboxes = objects.get(bbox_column, []) or [] |
| else: |
| bboxes = example.get(bbox_column, []) or [] |
|
|
| img_w = 1.0 |
| img_h = 1.0 |
| if width_column: |
| img_w = float(example.get(width_column, 1.0) or 1.0) |
| if height_column: |
| img_h = float(example.get(height_column, 1.0) or 1.0) |
|
|
| converted = [] |
| for bbox in bboxes: |
| if bbox is None or len(bbox) < 4: |
| converted.append(bbox) |
| continue |
| converted.append(convert_bbox(bbox, from_fmt, to_fmt, img_w, img_h)) |
|
|
| if is_nested: |
| new_objects = dict(objects) |
| new_objects[bbox_column] = converted |
| return {"objects": new_objects} |
| else: |
| return {bbox_column: converted} |
|
|
|
|
| def create_dataset_card( |
| source_dataset: str, |
| output_dataset: str, |
| from_fmt: str, |
| to_fmt: str, |
| num_samples: int, |
| processing_time: str, |
| split: str, |
| ) -> str: |
| return f"""--- |
| tags: |
| - object-detection |
| - bbox-conversion |
| - panlabel |
| - uv-script |
| - generated |
| --- |
| |
| # Bbox Format Conversion: {from_fmt} -> {to_fmt} |
| |
| Bounding boxes converted from `{from_fmt}` to `{to_fmt}` format. |
| |
| ## Processing Details |
| |
| - **Source Dataset**: [{source_dataset}](https://huggingface.co/datasets/{source_dataset}) |
| - **Conversion**: `{from_fmt}` -> `{to_fmt}` |
| - **Number of Samples**: {num_samples:,} |
| - **Processing Time**: {processing_time} |
| - **Processing Date**: {datetime.now().strftime("%Y-%m-%d %H:%M UTC")} |
| - **Split**: `{split}` |
| |
| ## Bbox Formats |
| |
| | Format | Description | |
| |--------|-------------| |
| | `coco_xywh` | `[x, y, width, height]` in pixels | |
| | `xyxy` | `[xmin, ymin, xmax, ymax]` in pixels | |
| | `voc` | `[xmin, ymin, xmax, ymax]` in pixels (alias for xyxy) | |
| | `yolo` | `[center_x, center_y, width, height]` normalized 0-1 | |
| | `tfod` | `[xmin, ymin, xmax, ymax]` normalized 0-1 | |
| | `label_studio` | `[x, y, width, height]` percentage 0-100 | |
| |
| ## Reproduction |
| |
| ```bash |
| uv run convert-hf-dataset.py {source_dataset} {output_dataset} --from {from_fmt} --to {to_fmt} |
| ``` |
| |
| Generated with panlabel-hf (convert-hf-dataset.py) |
| """ |
|
|
|
|
| def main( |
| input_dataset: str, |
| output_dataset: str, |
| from_fmt: str, |
| to_fmt: str, |
| bbox_column: str = "bbox", |
| width_column: str | None = "width", |
| height_column: str | None = "height", |
| split: str = "train", |
| max_samples: int | None = None, |
| batch_size: int = 1000, |
| hf_token: str | None = None, |
| private: bool = False, |
| create_pr: bool = False, |
| shuffle: bool = False, |
| seed: int = 42, |
| ): |
| """Convert bbox format in a HF dataset and push to Hub.""" |
|
|
| if from_fmt == to_fmt: |
| logger.error(f"Source and target formats are the same: {from_fmt}") |
| sys.exit(1) |
|
|
| start_time = datetime.now() |
|
|
| HF_TOKEN = hf_token or os.environ.get("HF_TOKEN") |
| if HF_TOKEN: |
| login(token=HF_TOKEN) |
|
|
| logger.info(f"Loading dataset: {input_dataset} (split={split})") |
| dataset = load_dataset(input_dataset, split=split) |
|
|
| if shuffle: |
| logger.info(f"Shuffling dataset with seed {seed}") |
| dataset = dataset.shuffle(seed=seed) |
|
|
| if max_samples: |
| dataset = dataset.select(range(min(max_samples, len(dataset)))) |
| logger.info(f"Limited to {len(dataset)} samples") |
|
|
| total_samples = len(dataset) |
| logger.info(f"Converting {total_samples:,} samples: {from_fmt} -> {to_fmt}") |
|
|
| |
| dataset = dataset.map( |
| lambda example: convert_example( |
| example, bbox_column, from_fmt, to_fmt, width_column, height_column |
| ), |
| desc=f"Converting {from_fmt} -> {to_fmt}", |
| ) |
|
|
| processing_duration = datetime.now() - start_time |
| processing_time_str = f"{processing_duration.total_seconds():.1f}s" |
|
|
| |
| conversion_info = json.dumps({ |
| "source_format": from_fmt, |
| "target_format": to_fmt, |
| "source_dataset": input_dataset, |
| "timestamp": datetime.now().isoformat(), |
| "script": "convert-hf-dataset.py", |
| }) |
|
|
| if "conversion_info" not in dataset.column_names: |
| dataset = dataset.add_column( |
| "conversion_info", [conversion_info] * len(dataset) |
| ) |
|
|
| |
| logger.info(f"Pushing to {output_dataset}") |
| max_retries = 3 |
| for attempt in range(1, max_retries + 1): |
| try: |
| if attempt > 1: |
| logger.warning("Disabling XET (fallback to HTTP upload)") |
| os.environ["HF_HUB_DISABLE_XET"] = "1" |
| dataset.push_to_hub( |
| output_dataset, |
| private=private, |
| token=HF_TOKEN, |
| max_shard_size="500MB", |
| create_pr=create_pr, |
| ) |
| break |
| except Exception as e: |
| logger.error(f"Upload attempt {attempt}/{max_retries} failed: {e}") |
| if attempt < max_retries: |
| delay = 30 * (2 ** (attempt - 1)) |
| logger.info(f"Retrying in {delay}s...") |
| time.sleep(delay) |
| else: |
| logger.error("All upload attempts failed.") |
| sys.exit(1) |
|
|
| |
| card_content = create_dataset_card( |
| source_dataset=input_dataset, |
| output_dataset=output_dataset, |
| from_fmt=from_fmt, |
| to_fmt=to_fmt, |
| num_samples=total_samples, |
| processing_time=processing_time_str, |
| split=split, |
| ) |
| card = DatasetCard(card_content) |
| card.push_to_hub(output_dataset, token=HF_TOKEN) |
|
|
| logger.info("Done!") |
| logger.info(f"Dataset: https://huggingface.co/datasets/{output_dataset}") |
| logger.info(f"Converted {total_samples:,} samples in {processing_time_str}") |
|
|
|
|
| if __name__ == "__main__": |
| parser = argparse.ArgumentParser( |
| description="Convert bbox formats in a HF object detection dataset", |
| formatter_class=argparse.RawDescriptionHelpFormatter, |
| epilog=""" |
| Bbox formats: |
| coco_xywh [x, y, width, height] in pixels |
| xyxy [xmin, ymin, xmax, ymax] in pixels |
| voc [xmin, ymin, xmax, ymax] in pixels (alias for xyxy) |
| yolo [cx, cy, w, h] normalized 0-1 |
| tfod [xmin, ymin, xmax, ymax] normalized 0-1 |
| label_studio [x, y, w, h] percentage 0-100 |
| |
| Examples: |
| uv run convert-hf-dataset.py merve/coco merve/coco-xyxy --from coco_xywh --to xyxy |
| uv run convert-hf-dataset.py merve/yolo merve/yolo-coco --from yolo --to coco_xywh |
| uv run convert-hf-dataset.py merve/tfod merve/tfod-coco --from tfod --to coco_xywh |
| """, |
| ) |
|
|
| parser.add_argument("input_dataset", help="Input dataset ID on HF Hub") |
| parser.add_argument("output_dataset", help="Output dataset ID on HF Hub") |
| parser.add_argument("--from", dest="from_fmt", required=True, choices=BBOX_FORMATS, help="Source bbox format") |
| parser.add_argument("--to", dest="to_fmt", required=True, choices=BBOX_FORMATS, help="Target bbox format") |
| parser.add_argument("--bbox-column", default="bbox", help="Column containing bboxes (default: bbox)") |
| parser.add_argument("--width-column", default="width", help="Column for image width (default: width)") |
| parser.add_argument("--height-column", default="height", help="Column for image height (default: height)") |
| parser.add_argument("--split", default="train", help="Dataset split (default: train)") |
| parser.add_argument("--max-samples", type=int, help="Max samples to process") |
| parser.add_argument("--batch-size", type=int, default=1000, help="Batch size for map (default: 1000)") |
| parser.add_argument("--hf-token", help="HF API token") |
| parser.add_argument("--private", action="store_true", help="Make output dataset private") |
| parser.add_argument("--create-pr", action="store_true", help="Create PR instead of direct push") |
| parser.add_argument("--shuffle", action="store_true", help="Shuffle dataset before processing") |
| parser.add_argument("--seed", type=int, default=42, help="Random seed (default: 42)") |
|
|
| args = parser.parse_args() |
|
|
| main( |
| input_dataset=args.input_dataset, |
| output_dataset=args.output_dataset, |
| from_fmt=args.from_fmt, |
| to_fmt=args.to_fmt, |
| bbox_column=args.bbox_column, |
| width_column=args.width_column, |
| height_column=args.height_column, |
| split=args.split, |
| max_samples=args.max_samples, |
| batch_size=args.batch_size, |
| hf_token=args.hf_token, |
| private=args.private, |
| create_pr=args.create_pr, |
| shuffle=args.shuffle, |
| seed=args.seed, |
| ) |
|
|