This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

MoRE: Unlocking Scalability in Reinforcement Learning for Quadruped Vision-Language-Action Models

Han Zhao‡,1,2, Wenxuan Song‡,3,4, Donglin Wang‡,2*, Xinyang Tong2,
Pengxiang Ding1,2, Xuelian Cheng3, Zongyuan Ge3
1Zhejiang University, China 2MiLAB, Westlake University, China
3AIM Lab, Faculty of IT, Monash University, Australia 4HKUST(GZ), China
Abstract

Developing versatile quadruped robots that can smoothly perform various actions and tasks in real-world environments remains a significant challenge. This paper introduces a novel vision-language-action (VLA) model, mixture of robotic experts (MoRE), for quadruped robots that aim to introduce reinforcement learning (RL) for fine-tuning large-scale VLA models with a large amount of mixed-quality data. MoRE integrates multiple low-rank adaptation modules as distinct experts within a dense multi-modal large language model (MLLM), forming a sparse-activated mixture-of-experts model. This design enables the model to effectively adapt to a wide array of downstream tasks. Moreover, we employ a reinforcement learning-based training objective to train our model as a Q-function after deeply exploring the structural properties of our tasks. Effective learning from automatically collected mixed-quality data enhances data efficiency and model performance. Extensive experiments demonstrate that MoRE outperforms all baselines across six different skills and exhibits superior generalization capabilities in out-of-distribution scenarios. We further validate our method in real-world scenarios, confirming the practicality of our approach and laying a solid foundation for future research on multi-task learning in quadruped robots.

footnotetext: ‡ Equal contributionfootnotetext: * Corresponding author. E-mail: wangdonglin@westlake.edu.cn

I INTRODUCTION

In the field of robotics research, enhancing the versatility and generalization of robots has remained a popular topic. These capabilities can effectively enhance the performance in following various language instructions to execute multiple tasks in complex and ever-changing open-world scenarios. With the rise of large language models [1, 2, 3] and visual perception research [4, 5, 6, 7], there are numerous studies that leverage multi-modal large language models (MLLMs) as a controller for robots to handle different tasks, complex language instructions and visual scenes [8, 9]. MLLMs [10, 11, 12, 13] learn extensive prior knowledge about images and natural language through large-scale pre-training on Internet-scale datasets, which have achieved significant development and application in numerous fields [14, 15, 16].

Therefore, to combine the powerful vision-language understanding capabilities of MLLMs with the data-driven paradigm of robot learning [17, 18], many studies have been conducted to obtain vision-language-action (VLA) models by training these foundational models on data from the execution of various downstream tasks [19, 20, 21, 22], enabling these large models to directly receive visual signals and natural language commands and output executable action for robots in an end-to-end manner. However, most end-to-end large-scale VLA approaches primarily follow the following training and inference paradigms:

Refer to caption
Figure 1: Visualization of MoRE on multi tasks. MoRE has been verified to be robust across various tasks, commands, and scenarios, in both simulation environments and real-world deployments.

1) These approaches typically involve directly fine-tuning on existing MLLM architectures, without delving deeper into whether such architectures are suitable for the diverse range of downstream robotic tasks.

Refer to caption
Figure 2: Overview of MoRE as applied to our multi-task quadruped vision-language-action task. The overview consists of four key components: (1) broad sub-optimal data combined with narrow expert data, (2) the MLLM backbone to generate action tokens from image and text embedding, (3) the Mixture of LoRA Experts finetuned to adapt to different tasks, and (4) the RL objectives used for training.

2) Models based on the imitation learning (IL) paradigm are typically fine-tuned exclusively on expert data using a supervised training objective. However, collecting such expert data often requires substantial human supervision or complex automated data collection systems with data filtering [23]. Consequently, these methods are unable to leverage more easily gathered sub-optimal data, which includes many trajectories that exhibit poor execution performance or even fail to complete tasks.

To address the aforementioned issues, we introduce Mixture of Robotic Experts (MoRE), a novel quadruped VLA model. MoRE simultaneously receives RGB images and language instructions as inputs, and directly outputs control commands for quadruped robots. By incorporating multiple Low Rank Adaptation (LoRA) modules as different experts into a dense VLA model and fine-tuning it into a sparse-activated Mixture of LoRA Experts model, MoRE achieves a balance for a wide variety of downstream quadruped robotic tasks. Additionally, it is widely understood that offline reinforcement learning (RL) is able to extract good policies from highly sub-optimal data, a scenario where imitation learning finds sub-optimal solutions that do not improve over the demonstrator that generated the dataset. Thus, MoRE employs a RL-based training objective to train our model as a Q-function after deeply exploring the structural properties of our task, effectively utilizing a large amount of automatically-collected mixed-quality data.

As a product of increased data diversity, new model components, and tailored training process, MoRE outperforms all baselines by a clear margin across 6 skills. Its high success rate in out-of-distribution experiments demonstrates superior generalization capabilities. Furthermore, we deployed our method in real-world scenarios, achieving promising trajectories. This confirms the practicality of our approach and lays a solid foundation for future research on multi-task learning in quadruped robots.

The main contributions in this paper include:

  • We propose MoRE, a novel VLA model for quadruped robots. To the best of our knowledge, this is the first work to explore the application of the MoE architecture in large-scale end-to-end VLA models. It shows high success rate and generalization in multi-task settings.

  • By introducing a training objective based on RL, we have included sub-optimal trajectories collected automatically, thus effectively enhancing both data efficiency and model performance for large-scale VLA models.

  • We conducted extensive experiments in both simulation and the real world to study the performance of MoRE in various settings.

II RELATED WORKS

II-A Vision-Language-Action Models

Vision-language-action (VLA) models typically accept visual feedback from robots and operation instructions in natural language form as input, and directly outputs operation instructions that the robots can execute. Several large-scale VLA models [19, 21, 24, 25, 20, 26, 27, 28], based on a pre-trained multi-modal large model capable of processing visual question answering (VQA), are fine-tuned on large-scale robot datasets. However, these models usually follow the structure of existing large language models, which fails to account for the unique characteristics of VLA tasks.

Additionally, the training of these models is conducted through imitation learning on expert data where execution is successful, and they are unable to learn from failure data collected automatically by robots. Although some approaches [29, 30] have attempted to use reinforcement learning paradigms on datasets that include both successful and failed cases, they are still limited to smaller-scale models and have not applied these methods to multi-modal large language models.

Refer to caption
Figure 3: The network architecture of MoRE. This figure illustrates the architecture of MoRE, which uses a decoder-only transformer (Fuyu 8B [12]) integrated with a Mixture of LoRA Experts. Tokens from different tasks such as locomotion, navigation, and manipulation are routed through a shared feed-forward network (FFN) with each expert dynamically selected by the router to provide the most relevant token-specific adaptation. The mixture-of-experts approaches allows for flexible token adaption within a single model.

II-B Mixture-of-Experts in Multi-modal Large Language Models

In the field of multi-modal large language models (MLLMs), there has been significant work attempting to incorporate the mixture-of-experts (MoE) architecture into the backbone of pre-trained language models to better accommodate tasks across multiple domains. These models can adaptive select different experts for processing based on the input data through a routing module, effectively preventing task conflicts that can arise during mixed-task training.

MoE-LLaVA [31] duplicates the Feed Forward Network (FFN) from a dense MLLM multiple times to serve as the initialization for the MoE layers and only fine-tunes the MoE portion. Despite having fewer activated parameters than many larger-scale models, it achieves better performance on multiple VQA tasks. LLaVA-MoLE [32] effectively combines the advantages of parameter-efficient fine-tuning and the MoE architecture by incorporating multiple low-rank matrices into the FFN. These models have achieved success in various domain-specific VQA tasks; however, there has been limited research delving into the advantages of the MoE architecture in robotic tasks.

III METHOD

In this section, we first provide an introduction to some preliminaries, then elaborate on the structure of the proposed model, and finally present the training objective based on reinforcement learning (RL) that we utilize.

III-A Preliminaries

RL. The tasks that the robot executes can be viewed as a Markov decision process (MDP). At each time step, the robot is in a state ss and can take an action aa. The transition dynamics is determined by the state transition function T(s|s,a)T(s^{\prime}|s,a), which is related to the environment, and a reward function R(s,a)R(s,a) is obtained. The goal of RL is to learn a policy π\pi that selects actions aa to maximize the cumulative reward, where the dimensionality of a is d𝒜d_{\mathcal{A}}. Value-based methods estimate the Q-function Q(s,a)Q(s,a) to evaluate the future discounted return of state-action pairs, expressed as tγtR(st,at)\sum_{t}\gamma^{t}R(s_{t},a_{t}), and adopt π(a|s)=argmaxaQ(s,a)\pi(a|s)={\operatorname{argmax}_{a}}Q(s,a) as the policy. The Q-function is updated through the iterative application of the Bellman operator

Q(st,at)=R(st,at)+γmaxat+1Q(st+1,at+1),\displaystyle\mathcal{B}^{*}Q(s_{t},a_{t})=R(s_{t},a_{t})+\gamma\max_{a_{t+1}}Q(s_{t+1},a_{t+1}),\vspace{-3pt} (1)

which is approximated via function approximation.

Mixture-of-experts (MoE). Unlike dense networks, the MoE architecture processes input by activating one or several submodules in a sparse layer. The selection of the activation modules is typically based on a learnable router layer. Considering an input xx for a network, the output yy of an MoE layer can be represented as

y=k=1NG(x)kEk(x),y=\sum_{k=1}^{N}G(x)_{k}E_{k}(x), (2)

where NN is the total number of experts in the MoE layer, Ek(x)E_{k}(x) is the output of the kk-th expert. G(x)kG(x)_{k} represents the kk-th element of the routing logits G(x)G(x), which are typically obtained by applying a linear transformation to the input and then selecting the Top-KK elements with the highest values:

G(x)=Softmax(TopK(Wgx))G(x)={\rm Softmax(Top}K(W_{g}x)) (3)

In the MoE layer, the degree of sparsity of the network depends on the choice of KK. When KK equals NN, it signifies that all parameters in the layer are activated, whereas when KK is set to 1, only a single expert in the entire layer is activated to process the input. In large language models, MoE layers are often applied to the feed-forward network (FFN) portion of the model. Through this sparsity, the model can reduce computation while maintaining capabilities comparable to those of dense models of the same scale.

III-B Model Architecture

Overall structure. In Figure 2, MoRE enables training high-capacity transformer architectures on mixed quality data, consisting of narrow expert data and broad sub-optimal data. The image IRGBI_{\rm RGB} and instruction TInstT_{\rm Inst} are separately processed into image and text tokens. The input tokens are concatenated and processed by the Fuyu 8B [12], a decoder-only transformer that supports inputs of arbitrary resolution and number of images, as the MLLM backbone to generate action tokens. It contains a sequence of transformer blocks (32 layers), each containing an MoE layer, preceded by normalization layers and self-attention layers with LoRA. Specifically, a Mixture of LoRA Experts module is designed to fine-tune the model. The probability of outputting an action can be expressed as:

PLM(at|st)=i=0d𝒜PLM(ati|IRGB;TInst;at1:i1),P_{\rm LM}(a_{t}|s_{t})=\prod_{i=0}^{d_{\mathcal{A}}}P_{\rm LM}(a_{t}^{i}|I_{\rm RGB};T_{\rm Inst};a_{t}^{1:i-1}), (4)

where tt is a given time-step, atia_{t}^{i} represents the ii-th action token output by the language model, and at1:ia_{t}^{1:i} denotes a sequence composed of action tokens from the first dimension to the ii-th dimension. Eventually, ata_{t} is de-tokenized into 12-dimensional discretized robot commands:

[vx,vy,ωz,θ1,θ2,θ3,f,hz,ϕ,sy,hzf,T]\displaystyle\left[v_{x},v_{y},\omega_{z},\theta_{1},\theta_{2},\theta_{3},f,h_{z},\phi,s_{y},h_{z}^{f},T\right] (5)

Here, vxv_{x}, vyv_{y}, and ωz\omega_{z} represent the velocities along the x-axis, y-axis, and z-axis respectively. θ1\theta_{1}, θ2\theta_{2}, and θ3\theta_{3} indicate the gait pattern, ff denotes the frequency, hzh_{z} represents the height of the robot, ϕ\phi denotes the pitch angle, sys_{y} corresponds to the foot width, hzfh_{z}^{f} represents the foot height, and TT indicates the termination signal of the action.

Mixture of LoRA experts. For the decoder layers in the language model, we employed the MixLoRA [33] approach for parameter-efficient fine-tuning and adjusted the FFN layers into MoE layers during the fine-tuning process. As shown in Figure 3, specifically during the fine-tuning process, each expert in the MoE layer contains an FFN with parameters shared among experts, and LoRA adapters are applied to each linear layer of the FFN. Therefore, each expert can be represented as:

Ek(x)=(Wdown+WdownLoRAk)f((Wup+WupLoRAk)x)E_{k}(x)=(W_{down}+W_{down}^{{\rm LoRA}_{k}})f((W_{up}+W_{up}^{{\rm LoRA}_{k}})x) (6)

where f()f(\cdot) is the activation function and WlayerLoRAkW_{layer}^{{\rm LoRA}_{k}} is the kk-th LoRA adapter of a particular layer.

We also incorporate a single LoRA adapter into the attention module of each decoder layer. The shared FFN is initialized from the original parameters of the MLLM and the whole backbone model remains frozen. Only the LoRA adapters are fine-tuned during training.

III-C Training Objective

Refer to caption
Figure 4: The analysis of the structure of our task. This figure illustrates the whole structure and intuition behind the critical points.

As illustrated in Figure 4, we identify 4 specific structural properties of the MDP in our task: (1) Horizon-independent returns. This property arises from our reward structure, where we assign a value of 1 solely upon successful task completion, with all other states yielding a reward of 0. (2) Limited critical points. The return of any trajectory can mostly be explained by the actions taken at a small fraction of critical states (e.g., crouching to avoid obstacles, dumping a load into a container). The lasting large proportion of states in a trajectory is not costly to recover after deviating from the optimal trajectory, this could be simply because there exists a large volume of near-optimal trajectories, or because at all but a few states, the volume of “good-enough” actions is large and can be identified using reward information. (3) Long-horizon data. The trajectories in our task are generally long. (4) Distribution shifts between the collected dataset and the evaluation phase. These characteristics allow the policy to explore more freely and make sub-optimal decisions in non-critical states without compromising overall performance. In this scenario, offline RL algorithms can rapidly converge to good-enough actions in non-critical states, a capability that IL methods lack due to their inherent limitation of merely imitating the data distribution [34].

In this paper, we leverage the training loss function proposed in [29] and [30] to fully exploit the model’s capacity as a Q-function under the MDP of our task, and adapt it to better suit the fine-tuning process for large models. Under the setup of auto-regressive discrete Q-learning, where the policy treats each dimension of the action as a token discretized from continuous values and outputs them auto-regressively, the Bellman operator defined in (1) can be obtained with the following form:

Q(st,at1:i1,ati)\displaystyle Q(s_{t},a_{t}^{1:i-1},a_{t}^{i}) =\displaystyle=
{maxati+1Q(st,at1:i,ati+1)if i<d𝒜R(st,at)+γmaxat+11Q(st+1,at+11)if i=d𝒜\displaystyle\begin{aligned} \begin{cases}\underset{a_{t}^{i+1}}{\max}~{}Q(s_{t},a_{t}^{1:i},a_{t}^{i+1})&\text{if }i<d_{\mathcal{A}}\\ R(s_{t},a_{t})+\gamma~{}\underset{a_{t+1}^{1}}{\max}~{}Q(s_{t+1},a_{t+1}^{1})&\text{if }i=d_{\mathcal{A}}\end{cases}\end{aligned} (7)
TABLE I: Overall performance. We evaluate 25 episodes per task for the final checkpoint on 6 challenging tasks from QUARD and report the success rates. The tasks are divided into “Easy”, “Medium” and “Hard” according to their difficulty.
Easy Medium Hard Average
Total Params Distinguish Go to Go avoid Go through Crawl Unload
CLIP[35] 86M 0.44 0.43 0.45 0.19 0 0 0.25
VC-1[36] 307M 0.46 0.43 0.45 0.31 0 0 0.28
QUART[24] 8B 0.66 0.60 0.53 0.41 0.32 0.12 0.44
MoRE (ours) 9.82B 0.82 0.80 0.59 0.57 0.49 0.33 0.60
TABLE II: Ablation study. We ablate 3 important designs in MoRE.
Easy Medium Hard Average
S-Data Distinguish Go to Go avoid Go through Crawl Unload
QUART[24] N 0.66 0.60 0.53 0.41 0.32 0.12 0.44
w/o RL N 0.73 0.67 0.58 0.47 0.34 0.24 0.51
w/o MoE Y 0.70 0.63 0.55 0.45 0.37 0.18 0.48
w/o S-Data N 0.78 0.69 0.61 0.53 0.45 0.28 0.56
MoRE (ours) Y 0.82 0.80 0.59 0.57 0.49 0.33 0.60

We train on an offline dataset pre-collected by other behavior policies. To ensure training stability, it is necessary to incorporate some conservative terms in the training objective to constrain the policy to train on state-action pairs within the dataset scope, preventing excessive overestimation of Q-values outside the training set. Additionally, we aim to minimize the probability of invalid action tokens, making the RL loss function defined as:

RL\displaystyle\mathcal{L}_{RL} =12𝔼s𝒟,aπβ(a|s)[(Q(s,a)Qk(s,a))2]\displaystyle=~{}\frac{1}{2}~{}\mathbb{E}_{s\sim\mathcal{D},a\sim\pi_{\beta}(a|s)}\left[\left(Q(s,a)-\mathcal{B}^{*}Q^{k}(s,a)\right)^{2}\right]
+α12𝔼s𝒟,aπ~β(a|s)[(Q(s,a)0)2],\displaystyle\quad+\alpha\cdot\frac{1}{2}\mathbb{E}_{s\sim\mathcal{D},a\sim\tilde{\pi}_{\beta}(a|s)}\left[(Q(s,a)-0)^{2}\right], (8)

where Q(s,a)=σ(PLM(a|s))Q(s,a)=\sigma(P_{\rm LM}(a|s)) and σ\sigma is the sigmoid activation, which ensures that the values of the Q-function fall within the range (0,1)(0,1). πβ(a|s)\pi_{\beta}(a|s) is the behavior policy distribution of the action during data collection, α\alpha is a factor that modulates the strength of the conservative regularization. π~β(a|s)=1Z(s)(1.0πβ(a|s))\tilde{\pi}_{\beta}(a|s)=\frac{1}{Z(s)}(1.0-\pi_{\beta}(a|s)) is the distribution over all actions which have a very low density under πβ(a|s)\pi_{\beta}(a|s), where Z(s){Z(s)} is a normalizing term within the distribution. Since the output PLMP_{\rm LM} of the large model denotes the distribution over the entire vocabulary of the language model, it also includes tokens that do not represent a valid action.

Additionally, in order to prevent certain experts in the MoE layer from being frequently selected, leading to an imbalance in workload, we adopt an auxiliary loss used in [37] to encourage the workload balance across experts:

MoE=1Nk=1NfkPk,\mathcal{L}_{MoE}=\frac{1}{N}\sum_{k=1}^{N}f_{k}P_{k}, (9)

where fk=1Tx𝒟1{argmaxp(x)=k}f_{k}=\frac{1}{T}\sum_{x\in\mathcal{D}}1\{\arg\max p(x)=k\} denotes the fraction of tokens dispatched to the kk-th expert, and Pk=1Tx𝒟pk(x)P_{k}=\frac{1}{T}\sum_{x\in\mathcal{D}}p_{k}(x) is the fraction of the router probability allocated for the kk-th expert.

The final loss function in the training process is set as:

=RL+βMoE,\mathcal{L}=\mathcal{L}_{RL}+\beta\mathcal{L}_{MoE}, (10)

and we choose the coefficient as α=0.5\alpha=0.5 and β=0.002\beta=0.002.

IV EXPERIMENTS

We concentrated experiments on evaluating the multi-task ability, generalization capabilities, and real-world deployment and aim to answer the following questions: Q1. Why is MoRE preferable than existing VLA methods? Q2. Why were the design choices? Q3. Does MoRE learn a superior policy that handles real robot tasks in 6 different scenes with limited real-world data?

IV-A Experiments Setup

Simulation environment. We use NVIDIA Isaac Gym [38] to collect the trajectories of the robot executing different tasks in parallel.

Real-world setting. The robot used to collect the trajectory data is Unitree Go2, which is a quadruped robot with 12 joints. The command output is sent to the low-level command tracking controller, which will be discussed below, to generate the actual joint action of the robot. The perception data is provided by a RealSense d435 camera installed in the front of the robot. The real-world experiment was executed by manually manipulating the quadruped robot.

Mixed-quality Datasets. The offline dataset used in our experiment includes two categories: demonstrations and sub-optimal data. Demonstrations correspond to successful tasks, which consist of 6 types of tasks (“Distinguish”, “Go to”, “Go through”, “Go avoid”, “Unload”, “Crawl”) with a total of 1,822,405 vision-language-action sets, all sourced from human demonstration data in QUARD [24]. This part of sub-optimal data comprises 4 types of tasks (“Distinguish”, “Go to”, “Go through”, “Go avoid”) and includes a total of 440,732 vision-language-action sets with all data sourced from QUARD-Auto [29].

Low-level controller. We use the RL policy trained in walk-these-ways [39] as the low-level controller for our robot. In walk-these-ways, a controller capable of controlling various gaits and body instructions was trained using RL methods with a mix of different reward functions related to diverse gait and body state commands. The policy was trained in parallel across a large number of simulation environments and different environmental parameters. To enhance the ability to execute a wide range of commands, a variety of commands are randomized within a specific range, allowing for comprehensive training across different operations. By adjusting these command parameters, complex behaviors can be generalized during inference.

Baseline. We compared our method with three previous state-of-the-art approaches on the quadruped robot VLA task: QUART [24], VC-1 [36], and CLIP [35]. To ensure a comprehensive and thorough evaluation, we maintained the same implementation as in [24], which enabled these methods to work as a VLA model.

Refer to caption
Figure 5: Real-world experiments. These images showcase the robot successfully performing various tasks, involving navigation (“Go to”), adjusting body posture during locomotion (“Crawl”), and whole body manipulation (“Unload”).

Training details. For the training on expert data, our model was trained using 8 NVIDIA A100 GPUs over 3 epochs, which required approximately 100 hours. The batch size per device was set to 16. For the mixed-quality training, we maintained the same hyper parameter settings, with the total training time amounting to approximately 125 hours.

To ensure a fair comparison between IL and RL methods, we exclusively used failure data from the QUARD-Auto dataset to construct the mixed-quality training set, without incorporating any successful data.

Evaluation Details. We conducted a comprehensive and robust series of experiments. We define a comprehensive set of evaluation tasks that covers various axes of generalization, such as visual (unseen objects, distractor objects), motion (positions of the target objects and the robot), locomotion (different gaits/ speed), and semantic (unseen instructions) generalization. To rigorously test the overall performance and generalization capabilities of MoRE, these out-of-distribution scenes and instructions were systematically generated together with in-distribution scenes.

IV-B Main Results in Simulation

Table I shows that MoRE consistently outperforms all baseline methods across the majority of tasks, achieving improvements of up to 20%. The exceptional success rate of MoRE in the “distinguish” and “go to” tasks highlights the strong foundational perception capabilities of MLLM. Its progress in tasks like “go through” and “go avoid” further reflects an understanding of spatial relationships. Additionally, the success in more complex tasks such as “crawl” and “unload” demonstrates the ability of whole-body manipulation, which is enabled by the implicit learning of interdependencies between different action dimensions through the use of a transformer. The consistently high success rate across all tasks is attributed to the mixture of robot experts, where each expert specializes in particular action dimensions and skills, allowing adaptation to various tasks. Moreover, training on tasks with sub-optimal data reveals potential for even greater success rate gains. Learning the ability to identify good-enough choice from the reward information is easy, and luckily our task setting allows a large volume of good-enough action, as we discussed in section III-C. Given that the evaluation includes out-of-distribution scenes, such robust performance further confirms the generalization capabilities of MoRE, which can be attributed to the common sense of pretrained MLLM.

IV-C Ablation Study

Table II presents a detailed summary of the ablation studies performed on three key components of MoRE. The findings are as follows: 1. The ablation study of the RL loss demonstrates a significant performance boost with its inclusion. Notably, even in scenarios with only expert data, the success rate of the offline RL algorithm surpasses that of IL methods, which also benefit from RL choice as mentioned before. 2. The ablation study of the MoE modules shows that the MoE modules balance computational cost and performance by activating only a portion of the parameters during inference. While the number of activated parameters remains comparable to that of a standard model, the multiple experts adapt to different task tokens, thereby improving the success rate in multi-task learning. 3. The ablation study of sub-optimal data shows that the inclusion of sub-optimal data improves the performance of MoRE through RL. However, the overly sparse reward function introduces learning challenges, which leads to performance degradation in “Go avoid” task.

IV-D Real-world Results

Furthermore, we tested the performance of MoRE in real-world scenarios. We further fine-tuned our model on a small real-world dataset and obtained an instance that can be deployed in real-world settings. We conducted experiments in several scenarios and achieved promising results. As shown in Figure 5, MoRE successfully completed tasks such as “go to the computer”, “crawl under the bar”, and “unload the yellow ball” in real-world scenarios, despite our real-world dataset only containing tasks like “go to the colored ball” with several different colors or “distinguish the letter” with different letters on the box. This demonstrates the strong generalization capability of our model to real-world scenarios after training with simulation data.

V CONCLUSIONS

In this paper, we present MoRE, a large-scale VLA model to unlock scalability in reinforcement learning. By integrating a mixture of robot experts into the VLA model, MoRE demonstrates strong effectiveness in multi-task learning. Furthermore, utilizing the VLA model as a Q-function enables MoRE to adopt an optimal training pipeline. Lastly, through the incorporation of mixed-quality data, MoRE gains the ability to learn from sub-optimal data, further enhancing its performance. MoRE achieves high success rate across tasks of different levels of difficulties and further behaves generalization skills on out-of-distribution scenes. The results are validated in both simulation and real-world robot settings.

ACKNOWLEDGMENTS

This work was supported by the National Science and Technology Innovation 2030 - Major Project (Grant No. 2022ZD0208800), and NSFC General Program (Grant No. 62176215).

References

  • [1] OpenAI, J. Achiam, S. Adler, et al., “Gpt-4 technical report,” 2024. [Online]. Available: https://arxiv.org/abs/2303.08774
  • [2] H. Touvron, T. Lavril, G. Izacard, et al., “Llama: Open and efficient foundation language models,” 2023. [Online]. Available: https://arxiv.org/abs/2302.13971
  • [3] H. Touvron, L. Martin, K. Stone, et al., “Llama 2: Open foundation and fine-tuned chat models,” 2023. [Online]. Available: https://arxiv.org/abs/2307.09288
  • [4] A. Radford, J. W. Kim, C. Hallacy, et al., “Learning transferable visual models from natural language supervision,” 2021. [Online]. Available: https://arxiv.org/abs/2103.00020
  • [5] A. Kirillov, E. Mintun, N. Ravi, et al., “Segment anything,” 2023. [Online]. Available: https://arxiv.org/abs/2304.02643
  • [6] J. Xu, X. Liu, Y. Bai, J. Jiang, K. Wang, X. Chen, and X. Ji, “Multi-camera collaborative depth prediction via consistent structure estimation,” in Proceedings of the 30th ACM International Conference on Multimedia, 2022, pp. 2730–2738.
  • [7] J. Xu, W. Yin, D. Gong, J. Jiang, and X. Liu, “Sdge: Stereo guided depth estimation for 360° camera sets,” in 2024 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2024, pp. 11 179–11 186.
  • [8] D. Driess, F. Xia, M. S. M. Sajjadi, et al., “Palm-e: An embodied multimodal language model,” in International Conference on Machine Learning, 2023. [Online]. Available: https://api.semanticscholar.org/CorpusID:257364842
  • [9] M. Ahn, A. Brohan, N. Brown, et al., “Do as i can, not as i say: Grounding language in robotic affordances,” in Conference on Robot Learning, 2022. [Online]. Available: https://api.semanticscholar.org/CorpusID:247939706
  • [10] H. Liu, C. Li, Q. Wu, et al., “Visual instruction tuning,” 2023. [Online]. Available: https://arxiv.org/abs/2304.08485
  • [11] J. Li, D. Li, C. Xiong, et al., “Blip: Bootstrapping language-image pre-training for unified vision-language understanding and generation,” 2022. [Online]. Available: https://arxiv.org/abs/2201.12086
  • [12] R. Bavishi, E. Elsen, C. Hawthorne, et al., “Introducing our multimodal models,” 2023. [Online]. Available: https://www.adept.ai/blog/fuyu-8b
  • [13] H. Zhao, M. Zhang, W. Zhao, et al., “Cobra: Extending mamba to multi-modal large language model for efficient inference,” 2024. [Online]. Available: https://arxiv.org/abs/2403.14520
  • [14] X. Lai, Z. Tian, Y. Chen, et al., “Lisa: Reasoning segmentation via large language model,” 2024. [Online]. Available: https://arxiv.org/abs/2308.00692
  • [15] H. Ye, D.-A. Huang, Y. Lu, et al., “X-vila: Cross-modality alignment for large language model,” 2024. [Online]. Available: https://arxiv.org/abs/2405.19335
  • [16] Y. Liu, P. Ding, S. Huang, et al., “Pite: Pixel-temporal alignment for large video-language model,” 2024. [Online]. Available: https://arxiv.org/abs/2409.07239
  • [17] P. Sermanet, T. Ding, J. Zhao, et al., “Robovqa: Multimodal long-horizon reasoning for robotics,” in 2024 IEEE International Conference on Robotics and Automation (ICRA), 2024, pp. 645–652.
  • [18] A. O’Neill, A. Rehman, A. Maddukuri, et al., “Open x-embodiment: Robotic learning datasets and rt-x models : Open x-embodiment collaboration0,” in 2024 IEEE International Conference on Robotics and Automation (ICRA), 2024, pp. 6892–6903.
  • [19] A. Brohan, N. Brown, J. Carbajal, et al., “Rt-2: Vision-language-action models transfer web knowledge to robotic control,” 2023. [Online]. Available: https://arxiv.org/abs/2307.15818
  • [20] X. Li, M. Liu, H. Zhang, et al., “Vision-language foundation models as effective robot imitators,” 2024. [Online]. Available: https://arxiv.org/abs/2311.01378
  • [21] M. J. Kim, K. Pertsch, S. Karamcheti, et al., “Openvla: An open-source vision-language-action model,” 2024. [Online]. Available: https://arxiv.org/abs/2406.09246
  • [22] W. Song, J. Chen, P. Ding, H. Zhao, W. Zhao, Z. Zhong, Z. Ge, J. Ma, and H. Li, “Accelerating vision-language-action model integrated with action chunking via parallel decoding,” 2025. [Online]. Available: https://arxiv.org/abs/2503.02310
  • [23] M. Ahn, D. Dwibedi, C. Finn, et al., “Autort: Embodied foundation models for large scale orchestration of robotic agents,” ArXiv, vol. abs/2401.12963, 2024. [Online]. Available: https://api.semanticscholar.org/CorpusID:266906759
  • [24] P. Ding, H. Zhao, W. Song, et al., “Quar-vla: Vision-language-action model for quadruped robots,” 2024. [Online]. Available: https://arxiv.org/abs/2312.14457
  • [25] Y. Cai, S. Yang, M. Li, X. Chen, Y. Mao, X. Yi, and W. Yang, “Task2morph: Differentiable task-inspired framework for contact-aware robot design,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).   IEEE, 2023, pp. 452–459.
  • [26] Y. Shentu, P. Wu, A. Rajeswaran, et al., “From llms to actions: Latent codes as bridges in hierarchical robot control,” 2024. [Online]. Available: https://arxiv.org/abs/2405.04798
  • [27] Q. Yu, S. Huang, X. Yuan, Z. Jiang, C. Hao, X. Li, H. Chang, J. Wang, L. Liu, H. Li, et al., “Uniaff: A unified representation of affordances for tool usage and articulation with vision-language models,” arXiv preprint arXiv:2409.20551, 2024.
  • [28] S. Li, J. Wang, R. Dai, W. Ma, W. Y. Ng, Y. Hu, and Z. Li, “Robonurse-vla: Robotic scrub nurse system based on vision-language-action model,” arXiv preprint arXiv:2409.19590, 2024.
  • [29] W. Song, H. Zhao, P. Ding, et al., “Germ: A generalist robotic model with mixture-of-experts for quadruped robot,” 2024. [Online]. Available: https://arxiv.org/abs/2403.13358
  • [30] Y. Chebotar, Q. Vuong, A. Irpan, et al., “Q-transformer: Scalable offline reinforcement learning via autoregressive q-functions,” 2023. [Online]. Available: https://arxiv.org/abs/2309.10150
  • [31] B. Lin, Z. Tang, Y. Ye, et al., “Moe-llava: Mixture of experts for large vision-language models,” 2024. [Online]. Available: https://arxiv.org/abs/2401.15947
  • [32] S. Chen, Z. Jie, and L. Ma, “Llava-mole: Sparse mixture of lora experts for mitigating data conflicts in instruction finetuning mllms,” 2024. [Online]. Available: https://arxiv.org/abs/2401.16160
  • [33] D. Li, Y. Ma, N. Wang, et al., “Mixlora: Enhancing large language models fine-tuning with lora-based mixture of experts,” 2024. [Online]. Available: https://arxiv.org/abs/2404.15159
  • [34] A. Kumar, J. Hong, A. Singh, et al., “When should we prefer offline reinforcement learning over behavioral cloning?” ArXiv, vol. abs/2204.05618, 2022. [Online]. Available: https://api.semanticscholar.org/CorpusID:248118795
  • [35] A. Radford, J. W. Kim, C. Hallacy, et al., “Learning transferable visual models from natural language supervision,” 2021.
  • [36] A. Majumdar, K. Yadav, S. Arnaud, et al., “Where are we in the search for an artificial visual cortex for embodied intelligence?” 2023.
  • [37] W. Fedus, B. Zoph, and N. Shazeer, “Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity,” 2022. [Online]. Available: https://arxiv.org/abs/2101.03961
  • [38] V. Makoviychuk et al., “Isaac gym: High performance gpu-based physics simulation for robot learning,” 2021.
  • [39] G. B. Margolis and P. Agrawal, “Walk these ways: Tuning robot control for generalization with multiplicity of behavior,” in Proceedings of The 6th Conference on Robot Learning, ser. Proceedings of Machine Learning Research, K. Liu, D. Kulic, and J. Ichnowski, Eds., vol. 205.   PMLR, 14–18 Dec 2023, pp. 22–31.