Search is not available for this dataset
video
video

Dataset Card for WebUOT-238-Test

This is a FiftyOne dataset with 238 samples.

Installation

If you haven't already, install FiftyOne:

pip install -U fiftyone

Usage

import fiftyone as fo
from fiftyone.utils.huggingface import load_from_hub

# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = load_from_hub("Voxel51/WebUOT-238-Test")

# Launch the App
session = fo.launch_app(dataset)

Dataset Description

WebUOT-1M is the largest million-scale benchmark for underwater object tracking (UOT), designed to address limitations in existing datasets by providing diverse underwater scenarios, rich annotations, and language prompts. It comprises 1.1 million frames across 1,500 underwater videos, covering 408 target categories categorized into 12 superclasses (e.g., fish, molluscs, inanimate objects). The dataset includes high-quality bounding box annotations, 23 tracking attributes (e.g., illumination variation, camouflage), and language descriptions for multimodal tracking research.

Note: This dataset, which has been parsed into FiftyOne format, comprises 238 randomly selected videos from the WebUOT-1M test set for a total of 192,000+ frames.

Dataset Details

  • Curated by:
    Chunhui Zhang (Shanghai Jiao Tong University), Li Liu (HKUST-Guangzhou), Guanjie Huang (HKUST-Guangzhou), Hao Wen (CloudWalk), Xi Zhou (CloudWalk), Yanfeng Wang (Shanghai Jiao Tong University).
  • Funded by:
    National Natural Science Foundation of China (No. 62101351), Key R&D Program of Chongqing (cstc2021jscx-gksbX0032).
  • Language(s):
    English (annotations and language prompts).
  • License:
    Creative Commons (intended for academic research).
  • Shared by: Harpreet Sahota, Hacker-in-Residence @ Voxel51

Dataset Sources

Uses

Direct Use

  • Training/evaluating UOT algorithms.
  • Multimodal tracking (vision + language prompts).
  • Studying domain adaptation (underwater vs. open-air environments).
  • Marine conservation, underwater robotics, and search/rescue applications.

Out-of-Scope Use

  • Non-underwater tracking tasks (e.g., aerial/terrestrial tracking).
  • Commercial applications without proper licensing.
  • Non-visual tasks (e.g., audio analysis).

Dataset Structure

  • Fields:
    • Videos: 1,500 clips (1,020 train / 480 test).
    • Annotations: Bounding boxes, absent labels, 23 attributes (e.g., low visibility, similar distractors).
    • Language Prompts: Text descriptions of targets (e.g., "red clownfish in yellow coral").
    • Metadata: Object categories (408), superclasses (12), resolution, duration.
  • Splits:
    Train/Test sets divided by videos, ensuring no overlap in categories or scenarios.

Dataset Creation

Curation Rationale

To bridge the gap in UOT research caused by small-scale datasets, WebUOT-1M was created to enable robust model training/evaluation, domain adaptation, and multimodal tracking in complex underwater environments.

Source Data

Data Collection and Processing

  • Sources: YouTube, Bilibili (filtered for diversity).
  • Processing:
    • Manual selection of moving targets.
    • Semi-supervised enhancement for blurry/low-visibility frames.
    • Professional annotation team for bounding boxes and attributes.
    • Final verification by authors.

Who are the source data producers?

Videos were captured by divers, underwater robots, and hobbyists using varied devices (cameras, phones).

Annotations

Annotation Process

  • Tools: In-house annotation tools; enhanced frames for challenging cases.
  • Guidelines: Focus on target motion, bounding box accuracy, and attribute labeling (23 attributes).
  • Validation: Multiple rounds of correction by authors.

Who are the annotators?

A professional labeling team and the authors performed verification.

Citation

BibTeX:

@article{zhang2024webuot,
  title={WebUOT-1M: Advancing Deep Underwater Object Tracking with A Million-Scale Benchmark},
  author={Zhang, Chunhui and Liu, Li and Huang, Guanjie and Wen, Hao and Zhou, Xi and Wang, Yanfeng},
  journal={arXiv preprint arXiv:2405.19818},
  year={2024}
}

Glossary

The following glossary details the attributes of each video.

Here's the content parsed as a markdown table:

Attribute Definition
01. LR If the size of the bounding box of the target in one frame is less than 400 pixels.
02. FM The center position of the target in two consecutive frames exceeds 20 pixels.
03. SV The ratio of the target bounding box is not within the range [0.5, 2].
04. ARV The aspect ratio of the target bounding box is not in the range [0.5, 2].
05. CM There is severe camera movement in the video frame.
06. VC Viewpoint changes significantly affect the appearance of the target.
07. PO If the target appears partially occluded in one frame.
08. FO As long as the target is completely occluded in one frame.
09. OV There is one frame where the target completely leaves the video frame.
10. ROT The target rotates in the video frame.
11. DEF The target appears deformation in the video frame.
12. SD Similarity interference appears around the target.
13. IV The illumination of the target area changes significantly.
14. MB The target area becomes blurred due to target motion or camera motion.
15. PTI In the initial frame only partial information about the target is visible.
16. NAO The target belongs to a natural or artificial object.
17. CAM The target is camouflaging in the video frame.
18. UV The underwater visibility of the target area (low, medium, or high visibility).
19. WCV The color of the water of the target area.
20. US Different underwater scenarios where the target is located.
21. SP Different shooting perspectives (underwater, outside-water, and fish-eye views).
22. SIZ The size s = √(w × h) of the video is small (s < √(640 × 480)), medium (√(640 × 480) ≤ s < √(1280 × 720)), or large (s ≥ √(1280 × 720)).
23. LEN The length l of the video is short (l ≤ 600 frames), medium (600 frames < l ≤ 1800 frames), or long (l > 1800 frames).
Downloads last month
55