| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
| |
|
|
| """ |
| Detect objects in images using Meta's SAM3 (Segment Anything Model 3). |
| |
| This script processes images from a HuggingFace dataset and detects a single object |
| type based on a text prompt, outputting bounding boxes in HuggingFace object detection format. |
| |
| Examples: |
| # Detect photographs in historical newspapers |
| uv run detect-objects.py \\ |
| davanstrien/newspapers-with-images-after-photography \\ |
| my-username/newspapers-detected \\ |
| --class-name photograph |
| |
| # Detect animals in camera trap images |
| uv run detect-objects.py \\ |
| wildlife-images \\ |
| wildlife-detected \\ |
| --class-name animal \\ |
| --confidence-threshold 0.6 |
| |
| # Test on small subset |
| uv run detect-objects.py input output \\ |
| --class-name table \\ |
| --max-samples 10 |
| |
| # Run on HF Jobs with GPU |
| hf jobs uv run --flavor a100-large \\ |
| -s HF_TOKEN=HF_TOKEN \\ |
| https://huggingface.co/datasets/uv-scripts/sam3/raw/main/detect-objects.py \\ |
| input-dataset output-dataset \\ |
| --class-name photograph \\ |
| --confidence-threshold 0.5 |
| |
| Note: To detect multiple object types, run the script multiple times with different |
| --class-name values and merge the results. |
| """ |
|
|
| import argparse |
| import logging |
| import os |
| import sys |
| import time |
| from typing import Any, Dict, List |
|
|
| import torch |
| from datasets import ClassLabel, Dataset, Features, Sequence, Value, load_dataset |
| from datasets import Image as ImageFeature |
| from huggingface_hub import DatasetCard, HfApi, login |
| from PIL import Image |
| from tqdm.auto import tqdm |
| from transformers import Sam3Model, Sam3Processor |
|
|
| os.environ["HF_XET_HIGH_PERFORMANCE"] = "1" |
| |
| logging.basicConfig( |
| level=logging.INFO, |
| format="%(asctime)s - %(levelname)s - %(message)s", |
| datefmt="%H:%M:%S", |
| ) |
| logger = logging.getLogger(__name__) |
|
|
| |
| if not torch.cuda.is_available(): |
| logger.error("❌ CUDA is not available. This script requires a GPU.") |
| logger.error("For local testing, ensure you have a CUDA-capable GPU.") |
| logger.error("For cloud execution, use HF Jobs with --flavor l4x1 or similar.") |
| sys.exit(1) |
|
|
|
|
| def parse_args(): |
| """Parse command line arguments.""" |
| parser = argparse.ArgumentParser( |
| description="Detect objects in images using SAM3", |
| formatter_class=argparse.RawDescriptionHelpFormatter, |
| epilog=__doc__, |
| ) |
|
|
| |
| parser.add_argument( |
| "input_dataset", help="Input HuggingFace dataset ID (e.g., 'username/dataset')" |
| ) |
| parser.add_argument( |
| "output_dataset", help="Output HuggingFace dataset ID (e.g., 'username/output')" |
| ) |
|
|
| |
| parser.add_argument( |
| "--class-name", |
| required=True, |
| help="Object class to detect (e.g., 'photograph', 'animal', 'table')", |
| ) |
| parser.add_argument( |
| "--confidence-threshold", |
| type=float, |
| default=0.5, |
| help="Minimum confidence score for detections (default: 0.5)", |
| ) |
| parser.add_argument( |
| "--mask-threshold", |
| type=float, |
| default=0.5, |
| help="Threshold for mask generation (default: 0.5)", |
| ) |
|
|
| |
| parser.add_argument( |
| "--image-column", |
| default="image", |
| help="Name of the column containing images (default: 'image')", |
| ) |
| parser.add_argument( |
| "--split", default="train", help="Dataset split to process (default: 'train')" |
| ) |
| parser.add_argument( |
| "--max-samples", |
| type=int, |
| default=None, |
| help="Maximum number of samples to process (for testing)", |
| ) |
| parser.add_argument( |
| "--shuffle", action="store_true", help="Shuffle dataset before processing" |
| ) |
|
|
| |
| parser.add_argument( |
| "--batch-size", |
| type=int, |
| default=4, |
| help="Batch size for processing (default: 4)", |
| ) |
| parser.add_argument( |
| "--model", |
| default="facebook/sam3", |
| help="SAM3 model ID (default: 'facebook/sam3')", |
| ) |
| parser.add_argument( |
| "--dtype", |
| default="bfloat16", |
| choices=["float32", "float16", "bfloat16"], |
| help="Model precision (default: 'bfloat16')", |
| ) |
|
|
| |
| parser.add_argument( |
| "--private", action="store_true", help="Make output dataset private" |
| ) |
| parser.add_argument( |
| "--hf-token", |
| default=None, |
| help="HuggingFace token (default: uses HF_TOKEN env var or cached token)", |
| ) |
|
|
| return parser.parse_args() |
|
|
|
|
| def create_dataset_card( |
| source_dataset: str, |
| model: str, |
| class_name: str, |
| num_samples: int, |
| total_detections: int, |
| images_with_detections: int, |
| processing_time: str, |
| confidence_threshold: float, |
| mask_threshold: float, |
| batch_size: int, |
| dtype: str, |
| image_column: str = "image", |
| split: str = "train", |
| ) -> str: |
| """Create a dataset card documenting the object detection process.""" |
| from datetime import datetime |
|
|
| model_name = model.split("/")[-1] |
| avg_detections = total_detections / num_samples if num_samples > 0 else 0 |
| detection_rate = ( |
| (images_with_detections / num_samples * 100) if num_samples > 0 else 0 |
| ) |
|
|
| return f"""--- |
| tags: |
| - object-detection |
| - sam3 |
| - segment-anything |
| - bounding-boxes |
| - uv-script |
| - generated |
| --- |
| |
| # Object Detection: {class_name.title()} Detection using {model_name} |
| |
| This dataset contains object detection results (bounding boxes) for **{class_name}** detected in images from [{source_dataset}](https://huggingface.co/datasets/{source_dataset}) using Meta's SAM3 (Segment Anything Model 3). |
| |
| **Generated using**: [uv-scripts/sam3](https://huggingface.co/datasets/uv-scripts/sam3) detection script |
| |
| ## Detection Statistics |
| |
| - **Objects Detected**: {class_name} |
| - **Total Detections**: {total_detections:,} |
| - **Images with Detections**: {images_with_detections:,} / {num_samples:,} ({detection_rate:.1f}%) |
| - **Average Detections per Image**: {avg_detections:.2f} |
| |
| ## Processing Details |
| |
| - **Source Dataset**: [{source_dataset}](https://huggingface.co/datasets/{source_dataset}) |
| - **Model**: [{model}](https://huggingface.co/{model}) |
| - **Script Repository**: [uv-scripts/sam3](https://huggingface.co/datasets/uv-scripts/sam3) |
| - **Number of Samples Processed**: {num_samples:,} |
| - **Processing Time**: {processing_time} |
| - **Processing Date**: {datetime.now().strftime("%Y-%m-%d %H:%M UTC")} |
| |
| ### Configuration |
| |
| - **Image Column**: `{image_column}` |
| - **Dataset Split**: `{split}` |
| - **Class Name**: `{class_name}` |
| - **Confidence Threshold**: {confidence_threshold} |
| - **Mask Threshold**: {mask_threshold} |
| - **Batch Size**: {batch_size} |
| - **Model Dtype**: {dtype} |
| |
| ## Model Information |
| |
| SAM3 (Segment Anything Model 3) is Meta's state-of-the-art object detection and segmentation model that excels at: |
| - 🎯 **Zero-shot detection** - Detect objects using natural language prompts |
| - 📦 **Bounding boxes** - Accurate object localization |
| - 🎭 **Instance segmentation** - Pixel-perfect masks (not included in this dataset) |
| - 🖼️ **Any image domain** - Works on photos, documents, medical images, etc. |
| |
| This dataset uses SAM3 in text-prompted detection mode to find instances of "{class_name}" in the source images. |
| |
| ## Dataset Structure |
| |
| The dataset contains all original columns from the source dataset plus an `objects` column with detection results in HuggingFace object detection format (dict-of-lists): |
| |
| - **bbox**: List of bounding boxes in `[x, y, width, height]` format (pixel coordinates) |
| - **category**: List of category indices (always `0` for single-class detection) |
| - **score**: List of confidence scores (0.0 to 1.0) |
| |
| ### Schema |
| |
| ```python |
| {{ |
| "objects": {{ |
| "bbox": [[x, y, w, h], ...], # List of bounding boxes |
| "category": [0, 0, ...], # All same class |
| "score": [0.95, 0.87, ...] # Confidence scores |
| }} |
| }} |
| ``` |
| |
| ## Usage |
| |
| ```python |
| from datasets import load_dataset |
| |
| # Load the dataset |
| dataset = load_dataset("{{{{output_dataset_id}}}}", split="{split}") |
| |
| # Access detections for an image |
| example = dataset[0] |
| detections = example["objects"] |
| |
| # Iterate through all detected objects in this image |
| for bbox, category, score in zip( |
| detections["bbox"], |
| detections["category"], |
| detections["score"] |
| ): |
| x, y, w, h = bbox |
| print(f"Detected {class_name} at ({{x}}, {{y}}) with confidence {{score:.2f}}") |
| |
| # Filter high-confidence detections |
| high_conf_examples = [ |
| ex for ex in dataset |
| if any(score > 0.8 for score in ex["objects"]["score"]) |
| ] |
| |
| # Count total detections across dataset |
| total = sum(len(ex["objects"]["bbox"]) for ex in dataset) |
| print(f"Total detections: {{total}}") |
| ``` |
| |
| ## Visualization |
| |
| To visualize the detections, you can use the visualization script from the same repository: |
| |
| ```bash |
| # Visualize first sample with detections |
| uv run https://huggingface.co/datasets/uv-scripts/sam3/raw/main/visualize-detections.py \\ |
| {{{{output_dataset_id}}}} \\ |
| --first-with-detections |
| |
| # Visualize random samples |
| uv run https://huggingface.co/datasets/uv-scripts/sam3/raw/main/visualize-detections.py \\ |
| {{{{output_dataset_id}}}} \\ |
| --num-samples 5 |
| |
| # Save visualizations to files |
| uv run https://huggingface.co/datasets/uv-scripts/sam3/raw/main/visualize-detections.py \\ |
| {{{{output_dataset_id}}}} \\ |
| --num-samples 3 \\ |
| --output-dir ./visualizations |
| ``` |
| |
| ## Reproduction |
| |
| This dataset was generated using the [uv-scripts/sam3](https://huggingface.co/datasets/uv-scripts/sam3) object detection script: |
| |
| ```bash |
| uv run https://huggingface.co/datasets/uv-scripts/sam3/raw/main/detect-objects.py \\ |
| {source_dataset} \\ |
| <output-dataset> \\ |
| --class-name {class_name} \\ |
| --confidence-threshold {confidence_threshold} \\ |
| --mask-threshold {mask_threshold} \\ |
| --batch-size {batch_size} \\ |
| --dtype {dtype} |
| ``` |
| |
| ### Running on HuggingFace Jobs (GPU) |
| |
| This script requires a GPU. To run on HuggingFace infrastructure: |
| |
| ```bash |
| hf jobs uv run --flavor a100-large \\ |
| -s HF_TOKEN=HF_TOKEN \\ |
| https://huggingface.co/datasets/uv-scripts/sam3/raw/main/detect-objects.py \\ |
| {source_dataset} \\ |
| <output-dataset> \\ |
| --class-name {class_name} \\ |
| --confidence-threshold {confidence_threshold} |
| ``` |
| |
| ## Performance |
| |
| - **Processing Speed**: ~{num_samples / (float(processing_time.split()[0]) * 60) if processing_time.split()[0].replace(".", "").isdigit() else "N/A":.1f} images/second |
| - **GPU Configuration**: CUDA with {dtype} precision |
| |
| --- |
| |
| Generated with 🤖 [UV Scripts](https://huggingface.co/uv-scripts) |
| """ |
|
|
|
|
| def load_and_validate_dataset( |
| dataset_id: str, |
| split: str, |
| image_column: str, |
| max_samples: int = None, |
| shuffle: bool = False, |
| hf_token: str = None, |
| ) -> Dataset: |
| """Load dataset and validate it has the required image column.""" |
| logger.info(f"📂 Loading dataset: {dataset_id} (split: {split})") |
|
|
| try: |
| dataset = load_dataset(dataset_id, split=split, token=hf_token) |
| except Exception as e: |
| logger.error(f"Failed to load dataset '{dataset_id}': {e}") |
| sys.exit(1) |
|
|
| |
| if image_column not in dataset.column_names: |
| logger.error(f"Column '{image_column}' not found in dataset") |
| logger.error(f"Available columns: {dataset.column_names}") |
| sys.exit(1) |
|
|
| |
| if shuffle: |
| logger.info("🔀 Shuffling dataset") |
| dataset = dataset.shuffle() |
|
|
| |
| if max_samples is not None: |
| logger.info(f"🔢 Limiting to {max_samples} samples") |
| dataset = dataset.select(range(min(max_samples, len(dataset)))) |
|
|
| logger.info(f"✅ Loaded {len(dataset)} samples") |
| return dataset |
|
|
|
|
| def process_batch( |
| batch: Dict[str, List[Any]], |
| image_column: str, |
| class_name: str, |
| processor: Sam3Processor, |
| model: Sam3Model, |
| confidence_threshold: float, |
| mask_threshold: float, |
| ) -> Dict[str, List[List[Dict[str, Any]]]]: |
| """Process a batch of images and return detections for a single class.""" |
| images = batch[image_column] |
|
|
| |
| pil_images = [] |
| for img in images: |
| if isinstance(img, str): |
| img = Image.open(img) |
| if img.mode == "L" or img.mode != "RGB": |
| img = img.convert("RGB") |
| pil_images.append(img) |
|
|
| |
| try: |
| inputs = processor( |
| images=pil_images, |
| text=[class_name] * len(pil_images), |
| return_tensors="pt", |
| ).to(model.device, dtype=model.dtype) |
|
|
| with torch.no_grad(): |
| outputs = model(**inputs) |
|
|
| |
| results = processor.post_process_instance_segmentation( |
| outputs, |
| threshold=confidence_threshold, |
| mask_threshold=mask_threshold, |
| target_sizes=inputs.get("original_sizes").tolist(), |
| ) |
|
|
| except Exception as e: |
| logger.warning(f"⚠️ Failed to process batch: {e}") |
| |
| return { |
| "objects": [ |
| {"bbox": [], "category": [], "score": []} |
| for _ in range(len(pil_images)) |
| ] |
| } |
|
|
| |
| batch_objects = [] |
|
|
| for result in results: |
| boxes = result.get("boxes", torch.tensor([])) |
| scores = result.get("scores", torch.tensor([])) |
|
|
| |
| if len(boxes) == 0: |
| batch_objects.append({"bbox": [], "category": [], "score": []}) |
| continue |
|
|
| |
| image_bboxes = [] |
| image_categories = [] |
| image_scores = [] |
|
|
| |
| boxes_np = boxes.cpu().float().numpy() |
| scores_np = scores.cpu().float().numpy() |
|
|
| for box, score in zip(boxes_np, scores_np): |
| x1, y1, x2, y2 = box |
| width = x2 - x1 |
| height = y2 - y1 |
|
|
| image_bboxes.append([float(x1), float(y1), float(width), float(height)]) |
| image_categories.append(0) |
| image_scores.append(float(score)) |
|
|
| batch_objects.append( |
| { |
| "bbox": image_bboxes, |
| "category": image_categories, |
| "score": image_scores, |
| } |
| ) |
|
|
| return {"objects": batch_objects} |
|
|
|
|
| def main(): |
| args = parse_args() |
|
|
| class_name = args.class_name.strip() |
| if not class_name: |
| logger.error("❌ Invalid --class-name argument. Provide a class name.") |
| sys.exit(1) |
|
|
| logger.info("🚀 SAM3 Object Detection") |
| logger.info(f" Input: {args.input_dataset}") |
| logger.info(f" Output: {args.output_dataset}") |
| logger.info(f" Class: {class_name}") |
| logger.info(f" Confidence threshold: {args.confidence_threshold}") |
| logger.info(f" Batch size: {args.batch_size}") |
|
|
| |
| if args.hf_token: |
| login(token=args.hf_token) |
| elif os.getenv("HF_TOKEN"): |
| login(token=os.getenv("HF_TOKEN")) |
|
|
| |
| dataset = load_and_validate_dataset( |
| args.input_dataset, |
| args.split, |
| args.image_column, |
| args.max_samples, |
| args.shuffle, |
| args.hf_token, |
| ) |
|
|
| |
| logger.info(f"🤖 Loading SAM3 model: {args.model}") |
| try: |
| processor = Sam3Processor.from_pretrained(args.model) |
| model = Sam3Model.from_pretrained( |
| args.model, torch_dtype=getattr(torch, args.dtype), device_map="auto" |
| ) |
| logger.info(f"✅ Model loaded on {model.device}") |
| except Exception as e: |
| logger.error(f"❌ Failed to load model: {e}") |
| logger.error("Ensure the model exists and you have access permissions") |
| sys.exit(1) |
|
|
| |
| logger.info("📊 Creating output schema...") |
| new_features = dataset.features.copy() |
| new_features["objects"] = { |
| "bbox": Sequence(Sequence(Value("float32"), length=4)), |
| "category": Sequence(ClassLabel(names=[class_name])), |
| "score": Sequence(Value("float32")), |
| } |
|
|
| |
| logger.info("🔍 Processing images...") |
| start_time = time.time() |
| processed_dataset = dataset.map( |
| lambda batch: process_batch( |
| batch, |
| args.image_column, |
| class_name, |
| processor, |
| model, |
| args.confidence_threshold, |
| args.mask_threshold, |
| ), |
| batched=True, |
| batch_size=args.batch_size, |
| features=new_features, |
| desc="Detecting objects", |
| ) |
| end_time = time.time() |
| processing_time_seconds = end_time - start_time |
| processing_time_str = f"{processing_time_seconds / 60:.1f} minutes" |
|
|
| |
| total_detections = sum(len(objs) for objs in processed_dataset["objects"]) |
| images_with_detections = sum(len(objs) > 0 for objs in processed_dataset["objects"]) |
|
|
| logger.info("✅ Detection complete!") |
| logger.info(f" Total detections: {total_detections}") |
| logger.info( |
| f" Images with detections: {images_with_detections}/{len(processed_dataset)}" |
| ) |
| logger.info( |
| f" Average detections per image: {total_detections / len(processed_dataset):.2f}" |
| ) |
|
|
| |
| logger.info(f"📤 Pushing to HuggingFace Hub: {args.output_dataset}") |
| try: |
| processed_dataset.push_to_hub(args.output_dataset, private=args.private) |
| logger.info( |
| f"✅ Dataset available at: https://huggingface.co/datasets/{args.output_dataset}" |
| ) |
| except Exception as e: |
| logger.error(f"❌ Failed to push to hub: {e}") |
| logger.info("💾 Saving locally as backup...") |
| processed_dataset.save_to_disk("./output_dataset") |
| logger.info("✅ Saved to ./output_dataset") |
| sys.exit(1) |
|
|
| |
| logger.info("📝 Creating dataset card...") |
| card_content = create_dataset_card( |
| source_dataset=args.input_dataset, |
| model=args.model, |
| class_name=class_name, |
| num_samples=len(processed_dataset), |
| total_detections=total_detections, |
| images_with_detections=images_with_detections, |
| processing_time=processing_time_str, |
| confidence_threshold=args.confidence_threshold, |
| mask_threshold=args.mask_threshold, |
| batch_size=args.batch_size, |
| dtype=args.dtype, |
| image_column=args.image_column, |
| split=args.split, |
| ) |
|
|
| card = DatasetCard(card_content) |
| card.push_to_hub(args.output_dataset, token=args.hf_token or os.getenv("HF_TOKEN")) |
| logger.info("✅ Dataset card created and pushed!") |
|
|
|
|
| if __name__ == "__main__": |
| main() |
|
|