Detect Objects (HF) avatar

Detect Objects (HF)

1 version
Open in App

Detect objects using HuggingFace models with automatic backend selection

Use This When

  • Building detection pipelines that need to swap models without changing downstream components
  • Leveraging pre-trained HuggingFace models like DETR, YOLOv8, or RT-DETR for quick prototyping
  • Comparing different detection architectures while maintaining consistent output format
  • Deploying detection systems where model choice depends on accuracy/speed requirements

What It Does

  • Auto-detects optimal backend based on model repository tags and configuration
  • Supports Transformers pipeline, PyTorch models, ONNX Runtime, and Ultralytics YOLO
  • Returns standardized bounding boxes with class IDs and confidence scores
  • Optionally filters returned detections by class ID
  • Handles model loading, preprocessing, inference, and postprocessing transparently

Works Best With

  • Video inputs → this component → track-objects-yolo for persistent tracking across frames
  • Detections → segment-image-hf or detect-landmarks for detailed region analysis
  • Integration with visualize or check-object-distance for spatial monitoring
  • Feeding detect-zone-transition or evaluate-expression for rule-based alerting

Caveats

  • Class ID mapping varies by model; verify downstream components use correct label mappings
  • ONNX and YOLO backends require GPU for real-time performance on high resolutions
  • Pipeline backend adds convenience but slightly more overhead than direct model inference
  • Threshold filtering happens post-NMS; adjust threshold based on precision/recall needs

Versions

  • 4bc889dalatestdefaultlinux/amd64

    Automated release