|
|
---
|
|
|
license: openrail
|
|
|
tags:
|
|
|
- robotics
|
|
|
- trajectory-prediction
|
|
|
- manipulation
|
|
|
- computer-vision
|
|
|
- time-series
|
|
|
pretty_name: Codatta Robotic Manipulation Trajectory
|
|
|
configs:
|
|
|
- config_name: default
|
|
|
data_files:
|
|
|
- split: train
|
|
|
path: data/train-*
|
|
|
dataset_info:
|
|
|
features:
|
|
|
- name: id
|
|
|
dtype: string
|
|
|
- name: total_frames
|
|
|
dtype: int32
|
|
|
- name: annotations
|
|
|
dtype: string
|
|
|
- name: trajectory_image
|
|
|
dtype: image
|
|
|
- name: video_path
|
|
|
dtype: string
|
|
|
splits:
|
|
|
- name: train
|
|
|
num_bytes: 39054025
|
|
|
num_examples: 50
|
|
|
download_size: 38738419
|
|
|
dataset_size: 39054025
|
|
|
language:
|
|
|
- en
|
|
|
size_categories:
|
|
|
- n<1K
|
|
|
---
|
|
|
|
|
|
# Codatta Robotic Manipulation Trajectory (Sample) |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
This dataset contains high-quality annotated trajectories of robotic gripper manipulations. It is designed to train models for fine-grained control, trajectory prediction, and object interaction tasks. |
|
|
|
|
|
Produced by **Codatta**, this dataset focuses on third-person views of robotic arms performing pick-and-place or manipulation tasks. Each sample includes the raw video, a visualization of the trajectory, and a rigorous JSON annotation of keyframes and coordinate points. |
|
|
|
|
|
**Note:** This is a sample dataset containing **50 annotated examples**. |
|
|
|
|
|
## Supported Tasks |
|
|
|
|
|
* **Trajectory Prediction:** Predicting the path of a gripper based on visual context. |
|
|
* **Keyframe Extraction:** Identifying critical moments in a manipulation task (e.g., contact, velocity change). |
|
|
* **Robotic Control:** Imitation learning from human-demonstrated or teleoperated data. |
|
|
|
|
|
## Dataset Structure |
|
|
|
|
|
### Data Fields |
|
|
|
|
|
* **`id`** (string): Unique identifier for the trajectory sequence. |
|
|
* **`total_frames`** (int32): Total number of frames in the video sequence. |
|
|
* **`video_path`** (string): Path to the source MP4 video file recording the manipulation action. |
|
|
* **`trajectory_image`** (image): A JPEG preview showing the overlaid trajectory path or keyframe visualization. |
|
|
* **`annotations`** (string): A JSON-formatted string containing the detailed coordinate data. |
|
|
* *Structure:* Contains lists of keyframes, timestamp, and the 5-point coordinates for the gripper in each annotated frame. |
|
|
|
|
|
### Data Preview |
|
|
*(Hugging Face's viewer will automatically render the `trajectory_image` here)* |
|
|
|
|
|
## Annotation Standards |
|
|
|
|
|
The data was annotated following a strict protocol to ensure precision and consistency. |
|
|
|
|
|
### 1. Viewpoint Scope |
|
|
* **Included:** Third-person views (fixed camera recording the robot). |
|
|
* [cite_start]**Excluded:** First-person views (Eye-in-Hand) are explicitly excluded to ensure consistent coordinate mapping[cite: 5, 15]. |
|
|
|
|
|
### 2. Keyframe Selection |
|
|
Annotations are not dense (every frame) but sparse, focusing on **Keyframes** that define the motion logic. [cite_start]A Keyframe is defined by the following events [cite: 20-25]: |
|
|
1. [cite_start]**Start Frame:** The gripper first appears in the screen[cite: 21]. |
|
|
2. [cite_start]**End Frame:** The gripper leaves the screen[cite: 22]. |
|
|
3. [cite_start]**Velocity Change:** Frames where the speed direction suddenly changes (marking the minimum speed point)[cite: 23]. |
|
|
4. [cite_start]**State Change:** Frames where the gripper opens or closes[cite: 24]. |
|
|
5. [cite_start]**Contact:** The precise moment the gripper touches the object[cite: 25]. |
|
|
|
|
|
### 3. The 5-Point Annotation Method |
|
|
[cite_start]For every annotated keyframe, the gripper is labeled with **5 specific coordinate points** to capture its pose and state accurately[cite: 27]: |
|
|
|
|
|
| Point ID | Description | Location Detail | |
|
|
| :--- | :--- | :--- | |
|
|
| **Point 1 & 2** | **Fingertips** | [cite_start]Center of the bottom edge of the gripper tips[cite: 28, 29]. | |
|
|
| **Point 3 & 4** | **Gripper Ends** | [cite_start]The rearmost points of the closing area (indicating the finger direction)[cite: 31]. | |
|
|
| **Point 5** | **Tiger's Mouth** | [cite_start]The center of the crossbeam (base of the gripper)[cite: 32]. | |
|
|
|
|
|
### 4. Quality Control |
|
|
* [cite_start]**Accuracy:** All datasets passed a rigorous quality assurance process with a minimum **95% accuracy rate**[cite: 78]. |
|
|
* **Occlusion Handling:** If the gripper is partially occluded, points are estimated based on object geometry. [cite_start]Sequences where the gripper is fully occluded or only shows a side profile without clear features are discarded[cite: 58, 63]. |
|
|
|
|
|
## Usage Example |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
import json |
|
|
|
|
|
# Load the dataset |
|
|
ds = load_dataset("Codatta/robotic-manipulation-trajectory", split="train") |
|
|
|
|
|
# Access a sample |
|
|
sample = ds[0] |
|
|
|
|
|
# View the image |
|
|
print(f"Trajectory ID: {sample['id']}") |
|
|
sample['trajectory_image'].show() |
|
|
|
|
|
# Parse annotations |
|
|
annotations = json.loads(sample['annotations']) |
|
|
print(f"Keyframes count: {len(annotations)}") |