File size: 4,779 Bytes
26e0670
 
9c153c8
 
 
 
 
 
 
c577e00
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9c153c8
 
 
 
26e0670
9c153c8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
---

license: openrail
tags:
- robotics
- trajectory-prediction
- manipulation
- computer-vision
- time-series
pretty_name: Codatta Robotic Manipulation Trajectory
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
dataset_info:
  features:
  - name: id
    dtype: string
  - name: total_frames
    dtype: int32
  - name: annotations
    dtype: string
  - name: trajectory_image
    dtype: image
  - name: video_path
    dtype: string
  splits:
  - name: train
    num_bytes: 39054025
    num_examples: 50
  download_size: 38738419
  dataset_size: 39054025
language:
- en
size_categories:
- n<1K
---


# Codatta Robotic Manipulation Trajectory (Sample)

## Dataset Summary

This dataset contains high-quality annotated trajectories of robotic gripper manipulations. It is designed to train models for fine-grained control, trajectory prediction, and object interaction tasks.

Produced by **Codatta**, this dataset focuses on third-person views of robotic arms performing pick-and-place or manipulation tasks. Each sample includes the raw video, a visualization of the trajectory, and a rigorous JSON annotation of keyframes and coordinate points.

**Note:** This is a sample dataset containing **50 annotated examples**.

## Supported Tasks

* **Trajectory Prediction:** Predicting the path of a gripper based on visual context.
* **Keyframe Extraction:** Identifying critical moments in a manipulation task (e.g., contact, velocity change).
* **Robotic Control:** Imitation learning from human-demonstrated or teleoperated data.

## Dataset Structure

### Data Fields

* **`id`** (string): Unique identifier for the trajectory sequence.
* **`total_frames`** (int32): Total number of frames in the video sequence.
* **`video_path`** (string): Path to the source MP4 video file recording the manipulation action.
* **`trajectory_image`** (image): A JPEG preview showing the overlaid trajectory path or keyframe visualization.
* **`annotations`** (string): A JSON-formatted string containing the detailed coordinate data. 
    * *Structure:* Contains lists of keyframes, timestamp, and the 5-point coordinates for the gripper in each annotated frame.

### Data Preview
*(Hugging Face's viewer will automatically render the `trajectory_image` here)*

## Annotation Standards

The data was annotated following a strict protocol to ensure precision and consistency.

### 1. Viewpoint Scope
* **Included:** Third-person views (fixed camera recording the robot).
* [cite_start]**Excluded:** First-person views (Eye-in-Hand) are explicitly excluded to ensure consistent coordinate mapping[cite: 5, 15].

### 2. Keyframe Selection
Annotations are not dense (every frame) but sparse, focusing on **Keyframes** that define the motion logic. [cite_start]A Keyframe is defined by the following events [cite: 20-25]:
1.  [cite_start]**Start Frame:** The gripper first appears in the screen[cite: 21].
2.  [cite_start]**End Frame:** The gripper leaves the screen[cite: 22].
3.  [cite_start]**Velocity Change:** Frames where the speed direction suddenly changes (marking the minimum speed point)[cite: 23].
4.  [cite_start]**State Change:** Frames where the gripper opens or closes[cite: 24].
5.  [cite_start]**Contact:** The precise moment the gripper touches the object[cite: 25].

### 3. The 5-Point Annotation Method
[cite_start]For every annotated keyframe, the gripper is labeled with **5 specific coordinate points** to capture its pose and state accurately[cite: 27]:

| Point ID | Description | Location Detail |
| :--- | :--- | :--- |
| **Point 1 & 2** | **Fingertips** | [cite_start]Center of the bottom edge of the gripper tips[cite: 28, 29]. |
| **Point 3 & 4** | **Gripper Ends** | [cite_start]The rearmost points of the closing area (indicating the finger direction)[cite: 31]. |
| **Point 5** | **Tiger's Mouth** | [cite_start]The center of the crossbeam (base of the gripper)[cite: 32]. |

### 4. Quality Control
* [cite_start]**Accuracy:** All datasets passed a rigorous quality assurance process with a minimum **95% accuracy rate**[cite: 78].
* **Occlusion Handling:** If the gripper is partially occluded, points are estimated based on object geometry. [cite_start]Sequences where the gripper is fully occluded or only shows a side profile without clear features are discarded[cite: 58, 63].

## Usage Example

```python
from datasets import load_dataset
import json

# Load the dataset
ds = load_dataset("Codatta/robotic-manipulation-trajectory", split="train")

# Access a sample
sample = ds[0]

# View the image
print(f"Trajectory ID: {sample['id']}")
sample['trajectory_image'].show()

# Parse annotations
annotations = json.loads(sample['annotations'])
print(f"Keyframes count: {len(annotations)}")