From a142c365dd6cb0c019d5031d612cd6da98e8562f Mon Sep 17 00:00:00 2001 From: Pepijn <138571049+pkooij@users.noreply.github.com> Date: Tue, 23 Dec 2025 14:13:32 +0100 Subject: [PATCH] use syslink for wall-x readme (#2708) * use syslink for wall-x readme * remove whitespace --- docs/source/policy_walloss_README.md | 35 ++++++++++++++++++++++++++ src/lerobot/policies/wall_x/README.md | 36 +-------------------------- 2 files changed, 36 insertions(+), 35 deletions(-) create mode 100644 docs/source/policy_walloss_README.md mode change 100644 => 120000 src/lerobot/policies/wall_x/README.md diff --git a/docs/source/policy_walloss_README.md b/docs/source/policy_walloss_README.md new file mode 100644 index 000000000..78548bd8d --- /dev/null +++ b/docs/source/policy_walloss_README.md @@ -0,0 +1,35 @@ +# WALL-OSS + +This repository contains the Hugging Face port of **WALL-OSS**, a Vision-Language-Action model for cross-embodiment robotic control based on Qwen2.5-VL with flow matching/FAST action prediction. + +--- + +## Model Overview + +| Feature | Description | +| ------------------ | ----------------------------------------------------- | --- | +| Base Model | Qwen2.5-VL (Vision-Language Model) | +| Action Prediction | Flow Matching (diffusion) or FAST (discrete tokens) | +| Architecture | Mixture of Experts (MoE) with action-specific routing | | +| Multi-Modal Inputs | Vision (images/videos), Language, Proprioception | + +--- + +## Citation + +If you use this work, please cite: + +```bibtex +@article{zhai2025igniting, + title = {Igniting VLMs Toward the Embodied Space}, + author = {Zhai, Andy and Liu, Brae and Fang, Bruno and Cai, Chalse and Ma, Ellie and Yin, Ethan and Wang, Hao and Zhou, Hugo and Wang, James and Shi, Lights and Liang, Lucy and Wang, Make and Wang, Qian and Gan, Roy and Yu, Ryan and Li, Shalfun and Liu, Starrick and Chen, Sylas and Chen, Vincent and Xu, Zach}, + journal = {arXiv preprint arXiv:2509.11766}, + year = {2025} +} +``` + +--- + +## License + +This port follows the **Apache 2.0 License**. diff --git a/src/lerobot/policies/wall_x/README.md b/src/lerobot/policies/wall_x/README.md deleted file mode 100644 index 78548bd8d..000000000 --- a/src/lerobot/policies/wall_x/README.md +++ /dev/null @@ -1,35 +0,0 @@ -# WALL-OSS - -This repository contains the Hugging Face port of **WALL-OSS**, a Vision-Language-Action model for cross-embodiment robotic control based on Qwen2.5-VL with flow matching/FAST action prediction. - ---- - -## Model Overview - -| Feature | Description | -| ------------------ | ----------------------------------------------------- | --- | -| Base Model | Qwen2.5-VL (Vision-Language Model) | -| Action Prediction | Flow Matching (diffusion) or FAST (discrete tokens) | -| Architecture | Mixture of Experts (MoE) with action-specific routing | | -| Multi-Modal Inputs | Vision (images/videos), Language, Proprioception | - ---- - -## Citation - -If you use this work, please cite: - -```bibtex -@article{zhai2025igniting, - title = {Igniting VLMs Toward the Embodied Space}, - author = {Zhai, Andy and Liu, Brae and Fang, Bruno and Cai, Chalse and Ma, Ellie and Yin, Ethan and Wang, Hao and Zhou, Hugo and Wang, James and Shi, Lights and Liang, Lucy and Wang, Make and Wang, Qian and Gan, Roy and Yu, Ryan and Li, Shalfun and Liu, Starrick and Chen, Sylas and Chen, Vincent and Xu, Zach}, - journal = {arXiv preprint arXiv:2509.11766}, - year = {2025} -} -``` - ---- - -## License - -This port follows the **Apache 2.0 License**. diff --git a/src/lerobot/policies/wall_x/README.md b/src/lerobot/policies/wall_x/README.md new file mode 120000 index 000000000..4426967de --- /dev/null +++ b/src/lerobot/policies/wall_x/README.md @@ -0,0 +1 @@ +../../../../docs/source/policy_walloss_README.md \ No newline at end of file