⚠️

This is a Dataset, not a Model

The following metrics do not apply: FNI Score, Deployment Options, Model Architecture

πŸ“Š

physicalai-smartspaces

FNI 20
by nvidia Dataset

"--- license: cc-by-4.0 --- !Demo of MTMC_Tracking_2025 Comprehensive, annotated dataset for multi-camera tracking and 2D/3D object detection. This dataset is synthetically generated with Omniverse. This dataset consists of over 250 hours of video from across nearly 1,500 cameras from indoor scenes i..."

Best Scenarios

✨ Data Science

Technical Constraints

Generic Use
- Size
- Rows
Parquet Format
59 Likes

Capabilities

  • βœ… Data Science

πŸ”¬Deep Dive

Expand Details [+]

πŸ› οΈ Technical Profile

⚑ Hardware & Scale

Size
-
Total Rows
-
Files
3193

🧠 Training & Env

Format
Parquet
Cleaning
Raw

🌐 Cloud & Rights

Source
huggingface
License
CC-BY-4.0

πŸ‘οΈ Data Preview

feature label split
example_text_1 0 train
example_text_2 1 train
example_text_3 0 test
example_text_4 1 validation
example_text_5 0 train
Showing 5 sample rows. Real-time preview requires login.

🧬 Schema & Configs

Fields

feature: string
label: int64
split: string

Dataset Card

Physical AI Smart Spaces Dataset

!Demo of MTMC_Tracking_2025

Overview

Comprehensive, annotated dataset for multi-camera tracking and 2D/3D object detection. This dataset is synthetically generated with Omniverse.

This dataset consists of over 250 hours of video from across nearly 1,500 cameras from indoor scenes in warehouses, hospitals, retail, and more. The dataset is time synchronized for tracking humans across multiple cameras using feature representation and no personal data.

Dataset Description

Dataset Owner(s)

NVIDIA

Dataset Creation Date

We started to create this dataset in December, 2023. First version was completed and released as part of 8th AI City Challenge in conjunction with CVPR 2024.

Dataset Characterization

  • Data Collection Method: Synthetic
  • Labeling Method: Automatic with IsaacSim

Video Format

  • Video Standard: MP4 (H.264)
  • Video Resolution: 1080p
  • Video Frame rate: 30 FPS

Ground Truth Format (MOTChallenge) for MTMC_Tracking_2024

Annotations are provided in the following text format per line:

code
<camera_id> <obj_id> <frame_id> <xmin> <ymin> <width> <height> <xworld> <yworld>

  • : Numeric identifier for the camera.
  • : Consistent numeric identifier for each object across cameras.
  • : Frame index starting from 0.
  • : Axis-aligned bounding box coordinates in pixels (top-left origin).
  • : Global coordinates (projected bottom points of objects) based on provided camera matrices.
The video file and calibration (camera matrix and homography) are provided for each camera view.

Calibration and ground truth files in the updated 2025 JSON format are now also included for each scene.

Notes:

  • Some calibration fieldsβ€”such as camera coordinates, camera directions, and scale factorsβ€”are not be available for the 2024 dataset due to original data limitations.
  • Please be aware that the video identified as scene_071/camera_0649 has encountered corruption issues. We advise you to exclude this video from your submissions.
  • In the sequences ranging from scenes 071 to 080, you'll find a storage room distinct from the primary retail space. Although separate, individuals can access this area. We have ensured that these sequences are synchronized effectively, treating them as part of the same continuous space for analysis purposes.

Directory Structure for MTMC_Tracking_2025

  • videos/: Video files.
  • depth_maps/: Depth maps stored as PNG images and compressed within HDF5 files. These files are exceedingly large; you may choose to use RGB videos only if preferred.
  • ground_truth.json: Detailed ground truth annotations (see below).
  • calibration.json: Camera calibration and metadata.
  • map.png: Visualization map in top-down view.

Ground Truth Format (JSON) for MTMC_Tracking_2025

Annotations per frame:

```json { "": [ { "object_type": "",

9,524 characters total