Being-H0.5: Scaling Human-Centric Robot Learning for Cross-Embodiment Generalization
- Hao LuoYe Wang Zongqing Lu
- 19 January 2026
Computer Science, Engineering
This work introduces Being-H0.5, a foundational Vision-Language-Action model designed for robust cross-embodiment generalization across diverse robotic platforms, and designs a unified sequential modeling and multi-task pre-training paradigm to bridge human demonstrations and robotic execution.
Rethinking Visual-Language-Action Model Scaling: Alignment, Mixture, and Regularization
- Ye WangSipeng Zheng Qin Jin
- 10 February 2026
Computer Science, Engineering
A systematic, controlled study of VLA scaling that revisits core training choices for pretraining across diverse robots and introduces a Grouped Blind Ensemble protocol that blinds operators to model identity and separates policy execution from outcome judgment, reducing experimenter bias.
Joint-Aligned Latent Action: Towards Scalable VLA Pretraining in the Wild
- Hao LuoYe Wang Zongqing Lu
- 25 February 2026
Computer Science
JALA is presented, a pretraining framework that learns Jointly-Aligned Latent Actions, a transition-aware, behavior-centric latent space for learning from heterogeneous human data that generates more realistic hand motions in both controlled and unconstrained scenarios.
Conservative Offline Robot Policy Learning via Posterior-Transition Reweighting
- Wanpeng ZhangHao Luo Zongqing Lu
- 17 March 2026
Computer Science
PTR reallocates credit according to how attributable each sample's post-action consequence is under the current representation, improving conservative offline adaptation to heterogeneous robot data.