The dataset viewer is not available for this split.
Error code: StreamingRowsError
Exception: ValueError
Message: Dataset 'ee_pose' has length 91 but expected 3
Traceback: Traceback (most recent call last):
File "/src/services/worker/src/worker/utils.py", line 99, in get_rows_or_raise
return get_rows(
^^^^^^^^^
File "/src/libs/libcommon/src/libcommon/utils.py", line 272, in decorator
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/src/services/worker/src/worker/utils.py", line 77, in get_rows
rows_plus_one = list(itertools.islice(ds, rows_max_number + 1))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2543, in __iter__
for key, example in ex_iterable:
^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2060, in __iter__
for key, pa_table in self._iter_arrow():
^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 2083, in _iter_arrow
for key, pa_table in self.ex_iterable._iter_arrow():
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 544, in _iter_arrow
for key, pa_table in iterator:
^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/iterable_dataset.py", line 383, in _iter_arrow
for key, pa_table in self.generate_tables_fn(**gen_kwags):
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 80, in _generate_tables
num_rows = _check_dataset_lengths(h5, self.info.features)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/local/lib/python3.12/site-packages/datasets/packaged_modules/hdf5/hdf5.py", line 359, in _check_dataset_lengths
raise ValueError(f"Dataset '{path}' has length {dset.shape[0]} but expected {num_rows}")
ValueError: Dataset 'ee_pose' has length 91 but expected 3Need help to make the dataset viewer work? Make sure to review how to configure the dataset viewer, and open a discussion for direct support.
YAML Metadata Warning: empty or missing yaml metadata in repo card
Check out the documentation for more information.
Seeing to Act, Prompting to Specify:
A Bayesian Factorization of Vision Language Action Policy
Kechun Xu · Zhenjie Zhu · Anzhe Chen · Shuqi Zhao · Qing Huang · Yifei Yang
Haojian Lu · Rong Xiong · Masayoshi Tomizuka · Yue Wang
TL; DR: BayesVLA decomposes the policy into a vision-action prior and a language-conditioned likelihood. The vision-action prior leverages visual information for action generation (seeing to act), while the language-conditioned likelihood aligns these action priors with the language instruction (prompting to specify).
🏆 Highlights
🔍 Key Findings: modality imbalance in VLA data encourage "visual shortcut", degrading generalization in language condition
✨ Key Insights:
- Bayesian factorization to VA prior and VLA likelihood to structurally address the data imbalance during fine-tuning. The prior focuses on action modeling and generation, while the likelihood focuses on language grounding and alignment.
Information-theoretical analysis reveals that the small conditional entropy $H(\ell\mid\mathbf{v})$ is the key reason for shortcut learning on visual cues, motivating self-bulit benchmarks demonstrating diverse language conditions.
Contact-aware architecture implementation: unified prior-likelihood formulation for both pre-contact and post-contact phases.
🧩 Overview
https://github.com/user-attachments/assets/6455fd89-311d-414e-a112-3eb96878c46d
Given VLA datasets with modality imbalance, BayesVLA models the VLA policy using a prior and a likelihood, trained with two stage procedure: For stage 1, we train a prior model that takes in visual input to generate multimodal action distribution. Based on the prior, for stage 2, we train the likelihood model to further align the action priors with the language instruction.
📘 Usage
Data Preparation
Droid
We use the processed data from cadence/droid_1.0.1 as it has camera extrinsic attached. Download it to anywhere you like, and make a symbolic link to it as
./data_raw/droid_1.0.1.bash scripts/data_preprocessing/process_libero.shLIBERO
Download the LIBERO dataset and make a symbolink to
./data_raw/libero.bash scripts/data_preprocessing/process_libero.shSelf-collected Datasets
We upload the processed datasets, including Pick-Place collected in IssacSim and Articulated Object Manipulation collected in IssacLab, and ALOHA collected in the real-world.
Pre-training (Optional)
Note that pretraining is only for post-contact phase.
bash scripts/postcontact/pretrain.sh
Post-training
- Pre-contact Phase
bash scripts/precontact/finetune.sh --config finetune_pp_arti - Post-contact Phase
# stage 0: va finetuning bash scripts/postcontact/finetune.sh --stage 0 --config finetune_pp_arti --va-name YOUR_VA_NAME # stage 1: vla finetuning bash scripts/postcontact/finetune.sh --stage 1 --config finetune_pp_arti --va-name YOUR_VA_NAME --vla-name YOUR_VLA_NAME
Evaluation
Launch the pyro4 naming server (something like roscore).
pyro4-nsBy default the naming server runs on
localhost:9090.Launch remote service of your fine-tuned model:
python -m infer_utils.remote_service \ --precontact_ckpt PRECONTACT_CKPT_PATH \ --postcontact_ckpt POSTCONTACT_CKPT_PATH \ --uri CUSTOM_URI_NAME
🤝 Acknowledgements
This projects builds upon OpenPi and E2VLA. We thank these teams for their open-source contributions.
📚 Citation
If you find this work useful, please consider citing:
@article{xu2025bayesvla,
title={Seeing to Act, Prompting to Specify: A Bayesian Factorization of Vision Language Action Policy},
author={Xu, Kechun and Zhu, Zhenjie and Chen, Anzhe and Zhao, Shuqi and Huang, Qing and Yang, Yifei and Lu, Haojian and Xiong, Rong and Tomizuka, Masayoshi and Wang, Yue},
journal={arXiv preprint arXiv:2512.11218},
year={2025}
}
- Downloads last month
- 9,737
