Datasets:
Update dataset card for VITA: correct links, add code, sample usage, abstract, and citation
#6
by
nielsr
HF Staff
- opened
README.md
CHANGED
|
@@ -1,9 +1,10 @@
|
|
| 1 |
---
|
| 2 |
-
license:
|
| 3 |
task_categories:
|
| 4 |
- robotics
|
| 5 |
tags:
|
| 6 |
- LeRobot
|
|
|
|
| 7 |
configs:
|
| 8 |
- config_name: default
|
| 9 |
data_files: data/*/*.parquet
|
|
@@ -13,11 +14,70 @@ This dataset was created using [LeRobot](https://github.com/huggingface/lerobot)
|
|
| 13 |
|
| 14 |
## Dataset Description
|
| 15 |
|
|
|
|
| 16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
|
| 22 |
## Dataset Structure
|
| 23 |
|
|
@@ -143,16 +203,15 @@ This dataset was created using [LeRobot](https://github.com/huggingface/lerobot)
|
|
| 143 |
}
|
| 144 |
```
|
| 145 |
|
| 146 |
-
|
| 147 |
## Citation
|
| 148 |
|
| 149 |
**BibTeX:**
|
| 150 |
|
| 151 |
```bibtex
|
| 152 |
-
@article{
|
| 153 |
-
|
| 154 |
-
|
| 155 |
-
|
| 156 |
-
|
| 157 |
}
|
| 158 |
```
|
|
|
|
| 1 |
---
|
| 2 |
+
license: apache-2.0
|
| 3 |
task_categories:
|
| 4 |
- robotics
|
| 5 |
tags:
|
| 6 |
- LeRobot
|
| 7 |
+
- flow-matching
|
| 8 |
configs:
|
| 9 |
- config_name: default
|
| 10 |
data_files: data/*/*.parquet
|
|
|
|
| 14 |
|
| 15 |
## Dataset Description
|
| 16 |
|
| 17 |
+
This dataset contains demonstration data for robotic tasks, utilized by **VITA: Vision-to-Action Flow Matching Policy**. VITA is a noise-free and conditioning-free policy learning framework that directly maps visual representations to latent actions using flow matching. It treats latent visual representations as the source of the flow, eliminating the need for conditioning. VITA is evaluated on 8 simulation and 2 real-world tasks from ALOHA and Robomimic, outperforming or matching state-of-the-art generative policies while achieving faster inference.
|
| 18 |
|
| 19 |
+
- **Homepage:** https://ucd-dare.github.io/VITA/
|
| 20 |
+
- **Paper:** https://huggingface.co/papers/2507.13231
|
| 21 |
+
- **Code:** https://github.com/ucd-dare/VITA
|
| 22 |
+
- **License:** Apache 2.0
|
| 23 |
|
| 24 |
+
## Sample Usage
|
| 25 |
+
|
| 26 |
+
To get started with VITA and use this dataset, follow these steps to set up the environment, preprocess datasets, and train a policy, as described in the [VITA GitHub repository](https://github.com/ucd-dare/VITA).
|
| 27 |
+
|
| 28 |
+
First, clone the repository and set up the conda environment:
|
| 29 |
+
```bash
|
| 30 |
+
git clone [email protected]:ucd-dare/VITA.git
|
| 31 |
+
cd VITA
|
| 32 |
+
conda create --name vita python==3.10
|
| 33 |
+
conda activate vita
|
| 34 |
+
conda install cmake
|
| 35 |
+
pip install -e .
|
| 36 |
+
pip install -r requirements.txt
|
| 37 |
+
# Install LeRobot dependencies
|
| 38 |
+
cd lerobot
|
| 39 |
+
pip install -e .
|
| 40 |
+
# Install ffmpeg for dataset processing
|
| 41 |
+
conda install -c conda-forge ffmpeg
|
| 42 |
+
```
|
| 43 |
+
|
| 44 |
+
Set the dataset storage path (replace `<PATH_TO_VITA>` with the absolute path to your cloned VITA directory):
|
| 45 |
+
```bash
|
| 46 |
+
echo 'export FLARE_DATASETS_DIR=<PATH_TO_VITA>/gym-av-aloha/outputs' >> ~/.bashrc
|
| 47 |
+
source ~/.bashrc
|
| 48 |
+
conda activate vita
|
| 49 |
+
```
|
| 50 |
+
|
| 51 |
+
Install benchmark dependencies for AV-ALOHA and/or Robomimic as needed:
|
| 52 |
+
```bash
|
| 53 |
+
# For AV-ALOHA
|
| 54 |
+
cd gym-av-aloha
|
| 55 |
+
pip install -e .
|
| 56 |
+
# For Robomimic
|
| 57 |
+
cd gym-robomimic
|
| 58 |
+
pip install -e .
|
| 59 |
+
```
|
| 60 |
+
|
| 61 |
+
To convert a HuggingFace dataset to offline zarr format (e.g., `iantc104/av_aloha_sim_hook_package`):
|
| 62 |
+
```bash
|
| 63 |
+
cd gym-av-aloha/scripts
|
| 64 |
+
python convert.py -r iantc104/av_aloha_sim_hook_package
|
| 65 |
+
```
|
| 66 |
+
The converted datasets will be stored in `./gym-av-aloha/outputs`.
|
| 67 |
+
|
| 68 |
+
To train a VITA policy, use the `flare/train.py` script:
|
| 69 |
+
```bash
|
| 70 |
+
python flare/train.py policy=vita task=hook_package session=test
|
| 71 |
+
```
|
| 72 |
+
You can customize training with various flags, for example:
|
| 73 |
+
```bash
|
| 74 |
+
# Use a specific GPU
|
| 75 |
+
python flare/train.py policy=vita task=hook_package session=test device=cuda:2
|
| 76 |
+
|
| 77 |
+
# Change online validation frequency and episodes
|
| 78 |
+
python flare/train.py policy=vita task=hook_package session=test \
|
| 79 |
+
val.val_online_freq=2000 val.eval_n_episodes=10
|
| 80 |
+
```
|
| 81 |
|
| 82 |
## Dataset Structure
|
| 83 |
|
|
|
|
| 203 |
}
|
| 204 |
```
|
| 205 |
|
|
|
|
| 206 |
## Citation
|
| 207 |
|
| 208 |
**BibTeX:**
|
| 209 |
|
| 210 |
```bibtex
|
| 211 |
+
@article{gao2025vita,
|
| 212 |
+
title={VITA: Vision-to-Action Flow Matching Policy},
|
| 213 |
+
author={Gao, Dechen and Zhao, Boqi and Lee, Andrew and Chuang, Ian and Zhou, Hanchu and Wang, Hang and Zhao, Zhe and Zhang, Junshan and Soltani, Iman},
|
| 214 |
+
journal={arXiv preprint arXiv:2507.13231},
|
| 215 |
+
year={2025}
|
| 216 |
}
|
| 217 |
```
|