This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Play to the Score: Stage-Guided Dynamic Multi-Sensory Fusion for Robotic Manipulation

Ruoxuan Feng1     Di Hu1,*     Wenke Ma2     Xuelong Li3
1Gaoling School of Artificial Intelligence, Renmin University of China
2Shenzhen Taobotics Co., Ltd.
3Institute of Artificial Intelligence (TeleAI), China Telecom
{fengruoxuan, dihu}@ruc.edu.cn, wenke@taobotics.com, xuelong_li@ieee.org
Abstract

Humans possess a remarkable talent for flexibly alternating to different senses when interacting with the environment. Picture a chef skillfully gauging the timing of ingredient additions and controlling the heat according to the colors, sounds, and aromas, seamlessly navigating through every stage of the complex cooking process. This ability is founded upon a thorough comprehension of task stages, as achieving the sub-goal within each stage can necessitate the utilization of different senses. In order to endow robots with similar ability, we incorporate the task stages divided by sub-goals into the imitation learning process to accordingly guide dynamic multi-sensory fusion. We propose MS-Bot, a stage-guided dynamic multi-sensory fusion method with coarse-to-fine stage understanding, which dynamically adjusts the priority of modalities based on the fine-grained state within the predicted current stage. We train a robot system equipped with visual, auditory, and tactile sensors to accomplish challenging robotic manipulation tasks: pouring and peg insertion with keyway. Experimental results indicate that our approach enables more effective and explainable dynamic fusion, aligning more closely with the human fusion process than existing methods. The code and demos can be found at https://gewu-lab.github.io/MS-Bot/.

footnotetext: * Corresponding author. “Play to the score” is slang for adjusting actions to fit current circumstances.

Keywords: Multi-Sensory, Robotic Manipulation, Multi-Stage

1 Introduction

Humans are blessed with the ability to flexibly use various sensors to perceive and interact with the world. Over the years, endowing robots with this potent capability has remained a dream for humanity. Fortunately, with the advancement of computing devices and sensors [1, 2], building a multi-sensory robot system to assist humans has gradually become achievable. We have witnessed the emergence of some exciting multi-sensory robots, such as Optimus [3] and Figure 01 [4].

One challenge in tasks that robots assist humans with is the complex object manipulation task, which requires achieving a series of sub-goals. These sub-goals naturally divide a complex task into multiple stages. Many efforts attempt to make models comprehend these task stages, whether through hierarchical learning [5, 6, 7] or employing LLMs [8, 9, 10]. However, when incorporated into multi-sensory robots, this challenge becomes more profound. We need to not only understand the stages themselves but also rethink multi-sensory fusion from the fresh perspective of task stages.

Completing sub-goals within a complex task may require different senses, with varying modality importance at each stage. Thus, stage transitions may entail changes in modality importance. To illustrate this, we build a multi-sensory robot system (Fig. 1, left). We then use imitation learning to train a MULSA [11] model with self-attention fusion to complete pouring task and record the testing confidence (Fig. 1, right). We observe correlations between the confidence and task stages after masking one modality. For example, vision is most important in the aligning stage, and masking it causes a notable confidence drop. We also note minor modality confidence changes inside stages. These indicate that modality importance undergoes both inter-stage changes and intra-stage adjustments. We summarize this as a challenge in multi-sensory imitation learning: Modality Temporality.

The key to addressing this challenge is understanding the current task stage and its concrete internal state, i.e., coarse-to-fine stage comprehension. Multi-sensory features should be dynamically fused to meet the needs of the coarse-grained stage and adjust to the fine-grained internal state. However, most of the existing works that utilize multiple sensors [12, 13, 14, 15] essentially use static fusion like concatenating. They lack the adaptability to dynamic changes of modality importance in complex manipulation tasks. Some recent works [16, 11] attempt to use attention for dynamic fusion, but they still fail to break out the paradigm of fusion based solely on current observations. They neglect the importance of stage understanding in multi-sensory fusion.

In light of this key insight, we incorporate the stages divided by sub-goals into multi-sensory fusion. To perform coarse-to-fine stage understanding, we first encode current observations and historical actions into a state token and identify the current stage based on it. We then inject the stage information into the state token by weighting learnable stage tokens. For fusion, we propose a dynamic Multi-Sensory fusion method for roBot (MS-Bot) to allocate modality weights based on the state within current stage. Using cross-attention [17] with stage-injected state token as query, we dynamically select multi-sensory features of interest based on the fine-grained internal state. The dynamic allocation of attention score to feature tokens ensures effective fusion aligned with requirements.

To evaluate our MS-Bot method in the real world, we build a robot system that can perceive the environment with audio, visual and tactile sensors. We test it on two challenging manipulation tasks: pouring, where the robot needs to pour tiny beads of specific quality, and peg insertion with keyway, where robot needs to rotate the peg with a key to align the keyway during the insertion.

Refer to caption
Figure 1: An illustration for our multi-sensory robot system and a visualization of Modality Temporality in a multi-stage task: pouring. We show confidence in action prediction (maximum softmax score) when using the inputs of all modalities and selectively masking uni-modal features. Due to the changing importance of modalities, both evident inter-stage (coarse-grained) and minor intra-stage (fine-grained) changes in confidence are observed when masking uni-model features. The low confidence fluctuations near stage boundaries also reflect insufficient task stage understanding.

To summarize, our contributions are as follows:

  • We summarize Modality Temporality as a key challenge and identify the coarse-to-fine stage comprehension as a crucial point to handle the changing modality importance in complex tasks.

  • Based on the above insight, we propose a dynamic multi-sensory fusion method guided by the prediction of fine-grained state within the identified task stage.

  • We build a multi-sensory robot system and evaluate our method on two challenging robotic manipulation tasks: pouring and peg insertion with keyway.

Experimental results show that MS-Bot better comprehends the fine-grained state in current task stage and dynamically adjusts modality attention as needed, outperforming prior methods in both tasks. We hope this work opens up a new perspective on multi-sensor fusion, inspiring future works.

2 Related Work

Multi-Sensory Robot Learning. Recent works have extensively integrated visual, auditory, and tactile sensors into robot systems. The two common visual modalities, RGB and depth, are widely combined and applied in grasping [18, 19, 20, 21, 22], navigation [23, 24], and pose estimation [25, 26, 27] tasks. In recent years, the development of tactile sensors has sparked interest in the study of tactile modality. Fine-grained tactile data can assist in tasks such as grasping [28, 29, 12, 13], shape reconstruction [30, 31] and peg insertion [32, 15]. Audio modalities are typically applied in tasks such as pouring [11, 14], object tracking [33, 34], and navigation [35, 36, 37]. Some works even attempt to jointly use visual, audio, and tactile modalities to manipulate objects [38, 39, 11]. Based on these efforts, we further explore effective multi-sensory fusion in complex manipulation tasks.

Robot Learning for Multi-Stage Tasks. Complex multi-stage tasks have always been a focal point of research in robot manipulation. Previous works hierarchically decompose the multi-stage tasks, where upper-level models predict the next sub-goal, and lower-level networks predict specific actions based on the sub-goal. Following this paradigm, many hierarchical reinforcement learning [7, 40, 41, 42, 43, 44] and imitation learning [5, 6, 45, 46, 47] methods have demonstrated their effectiveness across various tasks.  [8, 9, 10] use LLMs or VLMs to decompose long-horizon tasks. Recently, [48] introduces Chain-of-Thought into robot control. While these works primarily focus on uni-sensory scenarios, we aim to investigate the impact of stage changes on multi-sensory fusion.

Dynamic Multi-Modal Fusion. Dynamic multi-modal fusion essentially adapts the fusion process based on the input data by assigning different weights to the uni-modal features. A simple and effective way to assign weights is to train a gate network that scores based on the quality of uni-modal features [49, 50, 51, 52]. Another approach to assessing the quality of modalities and dynamically fusing them is uncertainty-based fusion [53, 54, 55]. With the introduction of the transformer [17], self-attention provides a natural mechanism to connect multi-modal signals [56]. Hence, attention-based fusion methods rapidly become a focal point of research [11, 57, 58, 59]. However, these methods primarily perform dynamic fusion based solely on the inputs themselves, while we design dynamic fusion from the perspective of task stages in robotic scenarios.

3 Method

In this section, we discuss the challenges in multi-sensory imitation learning from the perspective of task stages and propose the dynamic multi-sensory fusion framework MS-Bot, based on coarse-to-fine task stage understanding, to overcome these challenges. Specifically, the fine-grained state refers to a concept that describes the current situation of the environment, while the coarse-grained stage, divided by sub-goals of the task, is a collection composed of many states, as follows:

𝐄=(X1State 1,a1),(X2State 2,a2),,(Xt1Statet1,at1)Stage 1,(Xt1+1Statet1+1,at1+1),,(Xt2Statet2,at2)Stage 2,,\begin{gathered}\mathbf{E}=\underbrace{(\underbrace{\textbf{X}_{1}}_{State\ 1},a_{1}),(\underbrace{\textbf{X}_{2}}_{State\ 2},a_{2}),...,(\underbrace{\textbf{X}_{t_{1}}}_{State\ t_{1}},a_{t_{1}})}_{Stage\ 1},\underbrace{(\underbrace{\textbf{X}_{t_{1}+1}}_{State\ t_{1}+1},a_{t_{1}+1}),...,(\underbrace{\textbf{X}_{t_{2}}}_{State\ t_{2}},a_{t_{2}})}_{Stage\ 2},...,\end{gathered} (1)

where 𝐄\mathbf{E} is an episode, Xt\textbf{X}_{t} is the multi-sensory observation at timestep tt and ata_{t} is the action.

3.1 Challenges in Multi-Sensory Imitation learning

Imitation learning is a data-efficient robot learning algorithm where a robot learns a task by mimicking the behavior demonstrated by an expert. However, directly applying it to complex tasks with multi-sensory data could encounter issues, such as low confidence fluctuations and changing modality importance shown in Fig. 1. We attribute the causes of these issues to two key challenges:

Non-Markovity. Training data for imitation learning is typically derived from human demonstrations, usually collected through teleoperation devices. However, when humans manipulate robots, their actions are not solely based on data observed by robot sensors, but are also influenced by memory. This becomes particularly significant in more complex multi-stage tasks, as memory can provide cues about the task stages. Identifying the current stage solely based on current multi-sensory observations can be much more challenging. This challenge is not brought by the introduction of multiple sensors. Some works have already relaxed the Markov assumption and utilized action history as an additional input or leveraged sequence models to enhance performance [60, 61]. We identify the action history as a crucial cue for understanding the current stage and the fine-grained state within the stage, and use it to predict stages rather than directly predicting the next action.

Modality Temporality. In a complex manipulation task, the importance of various uni-modal features could change over stages. At timesteps from different stages, a particular modality may contribute significantly to the prediction, or serve as a supplementary role to the primary modality, or provide little useful information. Moreover, different states within a stage, such as the beginning and end, may also exhibit minor changes in modality importance. We distinguish them as coarse-grained and fine-grained importance change. However, previous works typically treated each modality equally and did not consider the issue of changing modality importance in method design. This could lead to sub-optimal outcomes as uni-modal features with low quality could disrupt the overall prediction. On the contrary, we utilize task stage information to guide multi-sensory fusion, dynamically assigning weights of modalities based on the fine-grained state in the current stage.

3.2 Stage-Guided Dynamic Multi-Sensory Fusion for Robot (MS-Bot)

Refer to caption
Figure 2: The pipeline of our method MS-Bot. It consists of four parts: feature extractor, state tokenizer, stage comprehension module, and dynamic fusion module.

In this section, we introduce how our method MS-Bot addresses the above challenges. Considering Non-Markovity, we use action history as input to assist the model in assessing the current task state. For Modality Temporality, we establish a coarse-to-fine stage comprehension, decomposing the task state into coarse-grained stages and fine-grained states within a stage. Dynamic fusion guided by the fine-grained state within the current stage is employed to handle the changing modality importance.

To introduce the understanding of stages into the process of imitation learning, we incorporate task stages divided by sub-goals into the training samples. Given a sample (Xt,at)(\textbf{X}_{t},a_{t}) at timestep tt in a trajectory where Xt={Xt1,Xt2,,XtM}\textbf{X}_{t}=\{X_{t}^{1},X_{t}^{2},...,X_{t}^{M}\} is the observation of MM modalities and ata_{t} is the corresponding action, a stage label st{1,2,,S}s_{t}\in\{1,2,...,S\} is added, where S is the number of stages in the task divided by different sub-goals. The sample at timestep tt is then represented as (Xt,at,st)(\textbf{X}_{t},a_{t},s_{t}). Based on the sample with stage division, we then introduce the four main components of our method MS-Bot:

Feature Extractor. This component consists of several uni-modal encoders. Each encoder takes a brief history of observations XtmT×Hm×Wm×CmX_{t}^{m}\in\mathbb{R}^{T\times H_{m}\times W_{m}\times C_{m}} of the modality mm as input, where TT is the timestep number of the brief history and Hm,Wm,CmH_{m},W_{m},C_{m} indicates the input shape of modality mm. These observations are then encoded into feature tokens 𝐟tM×T×d\mathbf{f}_{t}\in\mathbb{R}^{M\times T\times d} where dd denotes dimension.

State Tokenizer. This component aims to encode the observations and action history (a1,a2,,at1)(a_{1},a_{2},...,a_{t-1}) into a token that can represent the current state. Action history is similar to human memory and can help to indicate the current state within the whole task. We input the action history as a one-hot sequence into an LSTM [62], then concatenate the output with the feature tokens and encode them into a state token ztstatez^{state}_{t} through a Multi-Layer Perceptron (MLP).

Stage Comprehension Module. Merely using the state token alone is insufficient for achieving a comprehensive understanding of task stages. This module aims to perform coarse-to-grained stage understanding by injecting stage information into the state token. For a task with SS stages, we use SS learnable stage tokens [stage1],,[stageS][stage_{1}],...,[stage_{S}] to represent each stage. After warmup training, each stage token is initialized with the mean of all state tokens on samples within the stage. Next, we use a gate network (MLP) to predict the current stage, i.e., coarse-grained stage comprehension. The SS-dimensional output scores 𝐠t\mathbf{g}_{t} of the gate are softmax-normalized and multiplied by each of the SS stage tokens, followed by summing up the results to obtain the stage token ztstagez^{stage}_{t} at timestep tt via:

𝐠t=(gt1,,gtS)=softmax(MLP(ztstate)),ztstage=1Sj=1S(gtj[stagej]).\begin{gathered}\mathbf{g}_{t}=(g^{1}_{t},...,g^{S}_{t})=softmax(MLP(z^{state}_{t})),\\ z^{stage}_{t}=\frac{1}{S}\sum_{j=1}^{S}(g^{j}_{t}\cdot[stage_{j}]).\end{gathered} (2)

We use a softmax score instead of one-hot encoding because the transition of stages is a continuous and gradual process. Finally, we compute the weighted sum of the state token ztstatez^{state}_{t} and the current stage token ztstagez^{stage}_{t} using a weight β\beta to obtain the stage-injected state token ztz^{*}_{t}:

zt=βztstate+(1β)ztstage.z^{*}_{t}=\beta\cdot z^{state}_{t}+(1-\beta)\cdot z_{t}^{stage}. (3)

Different from the old state token ztstatez^{state}_{t}, the new state token ztz^{*}_{t} represents the fine-grained state within a stage. In this case, ztstagez^{stage}_{t} is regarded as an anchor stage, while ztstatez^{state}_{t} can indicate the shift inside the stage, thereby achieving coarse-to-fine stage comprehension.

During the training process, we utilize stage labels to supervise the stage scores output by the gate net. Specifically, we penalize scores that do not correspond to the current stage. Additionally, for samples within a range near the stage boundaries, we do not penalize its score on the nearest stage, achieving a soft constraint effect. The loss for the gate net gate\mathcal{L}_{gate} on the ii-th sample is as follows:

gate,i=j=1S(wijgij),j{1,2,,S},wij={0,(si=j)or(k,|ki|γ,sisk),1,otherwise,\begin{gathered}\mathcal{L}_{gate,i}=\sum_{j=1}^{S}(w_{i}^{j}\cdot g_{i}^{j}),\ j\in\{1,2,...,S\},\\ w_{i}^{j}=\left\{\begin{array}[]{ll}0,&(s_{i}=j)\ or\ (\exists k,\ |k-i|\leq\gamma,\ s_{i}\neq s_{k}),\\ 1,&otherwise,\end{array}\right.\\ \end{gathered} (4)

where kk indicates a nearby sample in the same trajectory, sis_{i} and sks_{k} represent stage labels and γ\gamma is a hyper-parameter used to determine the range near the stage boundaries. This soft constraint allows for smoother variations in the predicted stage scores.

Dynamic Fusion Module. Due to variations in modality importance across different stages and the minor importance changes within a stage, we aim for the model to dynamically select the modalities of interest based on the fine-grained state within the current stage. We use the state token with stage information ztz^{*}_{t} as query, and the feature tokens 𝐟t\mathbf{f}_{t} as key and value for cross-attention [17]. This mechanism dynamically allocates attention scores to feature tokens based on the state token ztz^{*}_{t}, thereby achieving dynamic multi-sensory fusion over time. It also contributes to a more stable multi-sensory fusion, as the anchor ztstagez^{stage}_{t} changes minimally within a stage. The features from all modalities are integrated into a fusion token ztfusz^{fus}_{t} based on the current stage’s requirements. Finally, the fusion token ztfusz^{fus}_{t} is fed into an MLP to predict the next action ata_{t}.

In order to prevent the model from simply memorizing the actions corresponding to attention score patterns, we also introduce random attention blur mechanism. For each input, we replace the attention scores on feature tokens with the same average value 1M×T\frac{1}{M\times T} with a probability pp. We encourage the model to simultaneously learn both stage information and feature information in the fusion token.

The loss during training consists of two components: the classification loss for action prediction cls\mathcal{L}_{cls} and the penalty on the scores of the gate network gate\mathcal{L}_{gate}. The total training loss \mathcal{L} is as follows:

=cls+λgate,\begin{gathered}\mathcal{L}=\mathcal{L}_{cls}+\lambda\mathcal{L}_{gate},\end{gathered} (5)

where λ\lambda is a hyper-parameter to control the intensity of the score penalty.

4 Experiments

In this section, we explore the answers to the following questions through experiments: (Q1) Compared to previous methods, how much performance improvement does stage-guided dynamic multi-sensory fusion provide? (Q2) Why is stage-based dynamic fusion superior to fusion based solely on the input itself? How does stage comprehension help? (Q3) Can stage-guided dynamic fusion maintain its advantage over static fusion in out-of-distribution settings?

Refer to caption
Figure 3: An illustration of the task setup for peg insertion with keyway and pouring task.

4.1 Task Setup

Pouring. For the pouring task, the robot needs to pour tiny steel beads of specific quality from the cylinder in the hand into another cylinder. This task consists of four stages [44]: 1) Aligning, where the robot needs to move a graduated cylinder containing beads and align it with the target cylinder, 2) Start Pouring, where the robot rotates the end effector to pour out the beads at an appropriate speed, 3) Holding Still, where the robot maintains its posture to continue pouring out beads steadily, and 4) End Pouring, where the robot rotates the end effector at the appropriate time to stop the flow of beads. The initial mass of the beads is 90g/120g, while the target mass for pouring out is 40g/60g, both indicated by prompts. We use audio, vision (RGB) and touch modalities in this task.

Peg Insertion with Keyway. This task is an upgraded version of peg insertion, where the robot needs to align a peg with a key to the keyway on the base by rotating and then insert the peg fully. The alignment primarily relies on tactile feedback, as the camera cannot observe inside the hole. This task consists of three stages: 1) First Insertion, where the robot aligns the peg with the hole and inserts it until the key collides with the base, 2) Rotating, where the robot aligns the key with the keyway on the base by rotating the peg based on tactile feedback, and 3) Second Insertion, where the robot further inserts the peg to the bottom. We use RGB, depth and touch modalities in this task.

Fig. 3 briefly shows the two tasks. See Appendix A for a more detailed task setup for both tasks.

4.2 Physical Setup and Experimental Details

Robot Setup and Data Collection. We use a 6-DoF UFACTORY xArm 6 robotic arm equipped with a Robotiq 2F-140 gripper in all experiments. For both tasks, we generate Cartesian space displacement commands at a policy frequency of 5 Hz. We use an Intel RealSense D435i camera to record visual data (RGB and depth) with a resolution of 640×480640\times 480 at a frequency of 30Hz. We use a desktop microphone to record audio at a sampling rate of 44.1kHz. Tactile data are recorded by a GelSight [1] mini sensor with a resolution of 400×300400\times 300 and a frequency of 15Hz.

Baselines. We compare our method with three baselines in both tasks: 1) the concat model which directly concatenates all the uni-modal features [12, 13, 14, 15], 2) Du et al. [63] that uses LSTM to fuse the uni-modal features and the additional proprioceptive information, and 3) MULSA [11] which fuses the uni-modal features via self-attention. We also compare our method with two variants in each task: MS-Bot (w/o A/D) and MS-Bot (w/o T/R) where one modality in the task is removed (audio and touch for pouring while depth and RGB for peg insertion).

Method Pouring-Initial Mass (g) Pouring-Target Mass (g) Insertion
90 120 40 60 Success Rate
Concat 4.80 ±\pm 1.14 8.72 ±\pm 2.39 8.40 ±\pm 2.21 6.54 ±\pm 2.14 5/10
Du et al. [63] 4.32 ±\pm 1.22 7.79 ±\pm 2.11 8.54 ±\pm 2.04 6.26 ±\pm 2.01 5/10
MULSA [11] 3.05 ±\pm 1.01 6.42 ±\pm 1.98 7.12 ±\pm 1.66 4.19 ±\pm 1.24 6/10
MS-Bot (w/o A/D) 10.55 ±\pm 2.25 15.76 ±\pm 3.18 15.64 ±\pm 3.25 13.02 ±\pm 3.13 5/10
MS-Bot (w/o T/R) 8.32 ±\pm 1.74 12.99 ±\pm 3.04 13.02 ±\pm 3.13 9.77 ±\pm 1.65 5/10
MS-Bot 1.60 ±\pm 1.10 5.58 ±\pm 1.79 6.48 ±\pm 1.55 1.80 ±\pm 0.95 8/10
Table 1: Comparison of performance on pouring (mean ± standard deviation) and peg insertion with keyway (Q1). A: Audio, D: Depth, T: Touch, R: RGB.
Refer to caption
Figure 4: Visualization of the aggregated attention scores for each modality and stage scores in the pouring task (Q2). At each timestep, we average the attention scores on all feature tokens of each modality separately. The stage score is the output of the gate network after softmax normalization.

4.3 How much improvement does stage-guided dynamic multi-sensory fusion provide?

To evaluate how much improvement MS-Bot brings compared to previous methods, we conduct comparative experiments on the two tasks. For each task (considering changing initial weight and target weight as two tasks), we collect 30 human demonstrations and use 20 of them as the training set, leaving the remaining 10 as the validation set. For both tasks, we conduct 10 real-world tests.

Tab. 1 shows the performance of baselines and our method on two tasks. Our MS-Bot outperforms all baselines in both the pouring task and peg insertion task, demonstrating the superiority of stage-guided dynamic multi-sensory fusion. Despite using proprioceptive information, the improvement of Du et al. [63] over the concat model is minimal, indicating that the inadequate understanding of the coarse-grained and fine-grained modality importance change is the critical barrier. Moreover, these two baselines exhibit delayed responses to stage changes, causing excessive pouring in the pouring task and misalignment in the insertion task. MULSA employs self-attention for dynamic fusion, but its lack of a thorough stage understanding prevents it from fully leveraging the advantages of dynamic fusion. Our MS-Bot achieves the best performance by better allocating modality weights during fusion based on the understanding of the fine-grained state within the current stage (Q1).

The notable performance drop when one modality is removed indicates that each modality contributes significantly. In the pouring task, removing audio and tactile modalities degrades the model’s ability to assess the quality of poured beads. The impact is more significant when the audio modality is removed, corresponding to the notable decrease in confidence observed after masking the audio features in Stage 3 as shown in Fig. 1. In the peg insertion task, removing either the RGB or depth modality will degrade the alignment performance. It is worth noting that even without the assistance of the depth or RGB modality, our MS-Bot achieves an equal success rate compared to concat model and Du et al. [63], demonstrating the superiority of stage comprehension and dynamic fusion (Q1).

Method Overall
MS-Bot 3.88 ±\pm 2.55
- Attention Blur 4.01 ±\pm 2.54
- Stage Comprehend 4.62 ±\pm 2.34
- State Tokenizer 5.19 ±\pm 2.21
Table 2: Impact of each component of our method in pouring task (Q2). ‘-’ indicates further removing the module from the model in the previous line. Overall error on all settings is shown.
Method Pouring Insertion
Color Color Mess
Concat 8.75 ±\pm 2.02 2/10 3/10
Du et al. [63] 8.04 ±\pm 1.85 3/10 3/10
MULSA [11] 5.56 ±\pm 1.13 4/10 5/10
MS-Bot 2.41 ±\pm 1.47 6/10 6/10
Table 3: Comparison of models in scenes with visual distractors (Q3). “Color” indicates the color of the cylinder or the base is changed and “Mess” indicates there are some clutter near the base.

4.4 Why is stage-based dynamic fusion superior to fusion based solely on the input itself?

To better understand the working principles of MS-Bot and the reasons for performance improvement, we visualize the attention scores of each modality and the scores of the stage tokens. As shown in Fig. 4, MS-Bot accurately predicts the rapid stage changes. Due to the model’s coarse-to-fine task stage comprehension, the aggregated attention scores of the three modalities remain relatively stable, exhibiting clear inter-stage changes and minor intra-stage adjustment (Q2). Vision initially dominates during the first stage. Upon entering the second stage, the model begins to use the sound of beads to find the appropriate angle and distinguishes this stage from the last stage through touch. During the holding still stage, the model primarily relies on audio and tactile deformation to assess the mass of beads. In the final stage, the model discerns the completion of the pouring based on tactile deformation and the rise of tactile attention scores is observed. The changes in attention scores for the peg insertion task (see Fig. 12(b)) also follow a similar pattern, demonstrating the validity and explainability of the proposed stage-guided dynamic multi-sensory fusion (Q2).

We also conduct ablation experiments on the pouring task to examine the contributions of each module, as shown in Tab. 3. We observe significant contributions to performance from both the state tokenizer and the stage comprehension module (Q2). These two modules provide critical task state understanding and coarse-to-fine stage comprehension. In addition, removing random attention blur results in a minor performance decline. The complete ablation results are located in Tab. 8.

4.5 Can stage-guided dynamic fusion maintain its advantage in out-of-distribution settings?

To verify the generalization of our method to distractors, we conduct experiments with visual distractors in both tasks. In the pouring task, we changed the cylinder’s color from white to red. For the peg insertion task, we altered the base color from black to green (“Color”), and placed clutter around the base (“Mess”). As shown in Tab. 3, our method exhibits minimal impact across various scenarios with distractors and consistently maintains performance superiority, demonstrating the generalizability of stage comprehension (Q1, Q3). Two baseline methods, the concat model and Du et al. [63], suffer from performance degradation when the visual modality is disturbed. Due to their lack of ability to understand the stages and dynamically adjust modality weights, even when information from visual modality is not needed in a particular stage, the model is still affected by distractors. This leads to unsatisfying performance. Our method dynamically allocates modality weights based on the understanding of the current stage, thereby reducing the impact of visual distractors on the fused features (Q2, Q3). Consequently, it outperforms the two baseline methods using static fusion and the MULSA that solely relies on the current observation for dynamic fusion.

5 Conclusion and Limitations

We present MS-Bot, a stage-guided dynamic multi-sensory fusion method with coarse-to-fine stage comprehension, which dynamically focuses on relevant modalities based on the state within the current stage. Evaluated on the challenging pouring and peg insertion with keyway tasks, our method outperforms previous static and dynamic fusion methods. We find that stage comprehension can be generalized to scenes with various distractors, reducing the impact of interference from one modality on multi-sensory fusion. We hope our work can inspire future work on multi-sensory robots.

Limitations: Our stage division involves human factors. An improvement could be leveraging LLMs for automatic stage labeling. Additionally, proposing more challenging multi-sensory manipulation tasks would be interesting for future works. More limitations are discussed in Appendix L.

Acknowledgments

The project was supported by National Natural Science Foundation of China (NO.62106272).

References

  • Yuan et al. [2017] W. Yuan, S. Dong, and E. H. Adelson. Gelsight: High-resolution robot tactile sensors for estimating geometry and force. Sensors, 17(12):2762, 2017.
  • Schmidt et al. [2019] P. Schmidt, J. Scaife, M. Harville, S. Liman, and A. Ahmed. Intel® realsense™ tracking camera t265 and intel® realsense™ depth camera d435-tracking and depth. Real Sense, 2019.
  • Tesla [2024] Tesla. Tesla bot. Website, 2024. https://www.tesla.com/AI.
  • Figure [2024] Figure. Figure 01 + openai: Speech-to-speech reasoning and end-to-end neural network. Website, 2024. https://www.figure.ai/.
  • Luo et al. [2024] J. Luo, C. Xu, X. Geng, G. Feng, K. Fang, L. Tan, S. Schaal, and S. Levine. Multi-stage cable routing through hierarchical imitation learning. IEEE Transactions on Robotics, 2024.
  • Sharma et al. [2019] P. Sharma, D. Pathak, and A. Gupta. Third-person visual imitation learning via decoupled hierarchical controller. Advances in Neural Information Processing Systems, 32, 2019.
  • Nachum et al. [2018] O. Nachum, S. S. Gu, H. Lee, and S. Levine. Data-efficient hierarchical reinforcement learning. Advances in neural information processing systems, 31, 2018.
  • Belkhale et al. [2024] S. Belkhale, T. Ding, T. Xiao, P. Sermanet, Q. Vuong, J. Tompson, Y. Chebotar, D. Dwibedi, and D. Sadigh. Rt-h: Action hierarchies using language. arXiv preprint arXiv:2403.01823, 2024.
  • Huang et al. [2023] W. Huang, F. Xia, T. Xiao, H. Chan, J. Liang, P. Florence, A. Zeng, J. Tompson, I. Mordatch, Y. Chebotar, et al. Inner monologue: Embodied reasoning through planning with language models. In Conference on Robot Learning, pages 1769–1782. PMLR, 2023.
  • Sermanet et al. [2023] P. Sermanet, T. Ding, J. Zhao, F. Xia, D. Dwibedi, K. Gopalakrishnan, C. Chan, G. Dulac-Arnold, N. J. Joshi, P. Florence, et al. Robovqa: Multimodal long-horizon reasoning for robotics. In 2nd Workshop on Language and Robot Learning: Language as Grounding, 2023.
  • Li et al. [2023] H. Li, Y. Zhang, J. Zhu, S. Wang, M. A. Lee, H. Xu, E. Adelson, L. Fei-Fei, R. Gao, and J. Wu. See, hear, and feel: Smart sensory fusion for robotic manipulation. In Conference on Robot Learning, pages 1368–1378. PMLR, 2023.
  • Han et al. [2021] Y. Han, K. Yu, R. Batra, N. Boyd, C. Mehta, T. Zhao, Y. She, S. Hutchinson, and Y. Zhao. Learning generalizable vision-tactile robotic grasping strategy for deformable objects via transformer. arXiv preprint arXiv:2112.06374, 2021.
  • Guzey et al. [2023] I. Guzey, B. Evans, S. Chintala, and L. Pinto. Dexterity from touch: Self-supervised pre-training of tactile representations with robotic play. In Conference on Robot Learning, pages 3142–3166. PMLR, 2023.
  • Liang et al. [2020] H. Liang, C. Zhou, S. Li, X. Ma, N. Hendrich, T. Gerkmann, F. Sun, M. Stoffel, and J. Zhang. Robust robotic pouring using audition and haptics. In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pages 10880–10887. IEEE, 2020.
  • Fu et al. [2023] L. Fu, H. Huang, L. Berscheid, H. Li, K. Goldberg, and S. Chitta. Safe self-supervised learning in real of visuo-tactile feedback policies for industrial insertion. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 10380–10386. IEEE, 2023.
  • Sferrazza et al. [2023] C. Sferrazza, Y. Seo, H. Liu, Y. Lee, and P. Abbeel. The power of the senses: Generalizable manipulation from vision and touch through masked multimodal learning. 2023.
  • Vaswani et al. [2017] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017.
  • Fang et al. [2020] H.-S. Fang, C. Wang, M. Gou, and C. Lu. Graspnet-1billion: A large-scale benchmark for general object grasping. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 11444–11453, 2020.
  • Yu et al. [2022] S. Yu, D.-H. Zhai, Y. Xia, H. Wu, and J. Liao. Se-resunet: A novel robotic grasp detection method. IEEE Robotics and Automation Letters, 7(2):5238–5245, 2022.
  • Song et al. [2022] Y. Song, J. Wen, D. Liu, and C. Yu. Deep robotic grasping prediction with hierarchical rgb-d fusion. International Journal of Control, Automation and Systems, 20(1):243–254, 2022.
  • Qin et al. [2023] R. Qin, H. Ma, B. Gao, and D. Huang. Rgb-d grasp detection via depth guided learning with cross-modal attention. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 8003–8009. IEEE, 2023.
  • Pang et al. [2024] X. Pang, W. Xia, Z. Wang, B. Zhao, D. Hu, D. Wang, and X. Li. Depth helps: Improving pre-trained rgb-based policy with depth information injection. arXiv preprint arXiv:2408.05107, 2024.
  • Wellhausen et al. [2020] L. Wellhausen, R. Ranftl, and M. Hutter. Safe robot navigation via multi-modal anomaly detection. IEEE Robotics and Automation Letters, 5(2):1326–1333, 2020.
  • Huang et al. [2023] C. Huang, O. Mees, A. Zeng, and W. Burgard. Visual language maps for robot navigation. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 10608–10615. IEEE, 2023.
  • Xu et al. [2020] C. Xu, J. Chen, M. Yao, J. Zhou, L. Zhang, and Y. Liu. 6dof pose estimation of transparent object from a single rgb-d image. Sensors, 20(23):6790, 2020.
  • Tian et al. [2020] M. Tian, L. Pan, M. H. Ang, and G. H. Lee. Robust 6d object pose estimation by learning rgb-d features. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 6218–6224. IEEE, 2020.
  • He et al. [2021] Y. He, H. Huang, H. Fan, Q. Chen, and J. Sun. Ffb6d: A full flow bidirectional fusion network for 6d pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 3003–3013, 2021.
  • Cui et al. [2020a] S. Cui, R. Wang, J. Wei, J. Hu, and S. Wang. Self-attention based visual-tactile fusion learning for predicting grasp outcomes. IEEE Robotics and Automation Letters, 5(4):5827–5834, 2020a.
  • Cui et al. [2020b] S. Cui, R. Wang, J. Wei, F. Li, and S. Wang. Grasp state assessment of deformable objects using visual-tactile fusion perception. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 538–544. IEEE, 2020b.
  • Smith et al. [2020] E. Smith, R. Calandra, A. Romero, G. Gkioxari, D. Meger, J. Malik, and M. Drozdzal. 3d shape reconstruction from vision and touch. Advances in Neural Information Processing Systems, 33:14193–14206, 2020.
  • Wang et al. [2021] Y. Wang, W. Huang, B. Fang, F. Sun, and C. Li. Elastic tactile simulation towards tactile-visual perception. In Proceedings of the 29th ACM International Conference on Multimedia, pages 2690–2698, 2021.
  • Lee et al. [2019] M. A. Lee, Y. Zhu, K. Srinivasan, P. Shah, S. Savarese, L. Fei-Fei, A. Garg, and J. Bohg. Making sense of vision and touch: Self-supervised learning of multimodal representations for contact-rich tasks. In 2019 International Conference on Robotics and Automation (ICRA), pages 8943–8950. IEEE, 2019.
  • Wilson and Lin [2020] J. Wilson and M. C. Lin. Avot: Audio-visual object tracking of multiple objects for robotics. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 10045–10051. IEEE, 2020.
  • Qian et al. [2022] X. Qian, Z. Wang, J. Wang, G. Guan, and H. Li. Audio-visual cross-attention network for robotic speaker tracking. IEEE/ACM Transactions on Audio, Speech, and Language Processing, 31:550–562, 2022.
  • Gan et al. [2020] C. Gan, Y. Zhang, J. Wu, B. Gong, and J. B. Tenenbaum. Look, listen, and act: Towards audio-visual embodied navigation. In 2020 IEEE International Conference on Robotics and Automation (ICRA), pages 9701–9707. IEEE, 2020.
  • Chen et al. [2020] C. Chen, U. Jain, C. Schissler, S. V. A. Gari, Z. Al-Halah, V. K. Ithapu, P. Robinson, and K. Grauman. Soundspaces: Audio-visual navigation in 3d environments. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part VI 16, pages 17–36. Springer, 2020.
  • Younes et al. [2023] A. Younes, D. Honerkamp, T. Welschehold, and A. Valada. Catch me if you hear me: Audio-visual navigation in complex unmapped environments with moving sounds. IEEE Robotics and Automation Letters, 8(2):928–935, 2023.
  • Gao et al. [2022a] R. Gao, Y.-Y. Chang, S. Mall, L. Fei-Fei, and J. Wu. Objectfolder: A dataset of objects with implicit visual, auditory, and tactile representations. In Conference on Robot Learning, pages 466–476. PMLR, 2022a.
  • Gao et al. [2022b] R. Gao, Z. Si, Y.-Y. Chang, S. Clarke, J. Bohg, L. Fei-Fei, W. Yuan, and J. Wu. Objectfolder 2.0: A multisensory object dataset for sim2real transfer. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 10598–10608, 2022b.
  • Yang et al. [2021] X. Yang, Z. Ji, J. Wu, Y.-K. Lai, C. Wei, G. Liu, and R. Setchi. Hierarchical reinforcement learning with universal policies for multistep robotic manipulation. IEEE Transactions on Neural Networks and Learning Systems, 33(9):4727–4741, 2021.
  • Beyret et al. [2019] B. Beyret, A. Shafti, and A. A. Faisal. Dot-to-dot: Explainable hierarchical reinforcement learning for robotic manipulation. In 2019 IEEE/RSJ International Conference on intelligent robots and systems (IROS), pages 5014–5019. IEEE, 2019.
  • Botvinick [2012] M. M. Botvinick. Hierarchical reinforcement learning and decision making. Current opinion in neurobiology, 22(6):956–962, 2012.
  • Yang et al. [2018] Z. Yang, K. Merrick, L. Jin, and H. A. Abbass. Hierarchical deep reinforcement learning for continuous action control. IEEE transactions on neural networks and learning systems, 29(11):5174–5184, 2018.
  • Zhang et al. [2021] D. Zhang, Q. Li, Y. Zheng, L. Wei, D. Zhang, and Z. Zhang. Explainable hierarchical imitation learning for robotic drink pouring. IEEE Transactions on Automation Science and Engineering, 19(4):3871–3887, 2021.
  • Xie et al. [2020] F. Xie, A. Chowdhury, M. De Paolis Kaluza, L. Zhao, L. Wong, and R. Yu. Deep imitation learning for bimanual robotic manipulation. Advances in neural information processing systems, 33:2327–2337, 2020.
  • Li et al. [2021] B. Li, J. Li, T. Lu, Y. Cai, and S. Wang. Hierarchical learning from demonstrations for long-horizon tasks. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 4545–4551. IEEE, 2021.
  • Hakhamaneshi et al. [2021] K. Hakhamaneshi, R. Zhao, A. Zhan, P. Abbeel, and M. Laskin. Hierarchical few-shot imitation with skill transition models. In International Conference on Learning Representations, 2021.
  • Jia et al. [2023] Z. Jia, F. Liu, V. Thumuluri, L. Chen, Z. Huang, and H. Su. Chain-of-thought predictive control. arXiv preprint arXiv:2304.00776, 2023.
  • Arevalo et al. [2017] J. Arevalo, T. Solorio, M. Montes-y Gómez, and F. A. González. Gated multimodal units for information fusion. arXiv preprint arXiv:1702.01992, 2017.
  • Kim et al. [2018] J. Kim, J. Koh, Y. Kim, J. Choi, Y. Hwang, and J. W. Choi. Robust deep multi-modal learning based on gated information fusion network. In Asian Conference on Computer Vision, pages 90–106. Springer, 2018.
  • Wang et al. [2018] S. Wang, J. Zhang, and C. Zong. Learning multimodal word representation via dynamic fusion methods. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018.
  • Arevalo et al. [2020] J. Arevalo, T. Solorio, M. Montes-y Gomez, and F. A. González. Gated multimodal networks. Neural Computing and Applications, 32:10209–10228, 2020.
  • Han et al. [2022] Z. Han, F. Yang, J. Huang, C. Zhang, and J. Yao. Multimodal dynamics: Dynamical fusion for trustworthy multimodal classification. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 20707–20717, 2022.
  • Zhang et al. [2023] Q. Zhang, H. Wu, C. Zhang, Q. Hu, H. Fu, J. T. Zhou, and X. Peng. Provable dynamic fusion for low-quality multimodal data. In International conference on machine learning, pages 41753–41769. PMLR, 2023.
  • Tellamekala et al. [2023] M. K. Tellamekala, S. Amiriparian, B. W. Schuller, E. André, T. Giesbrecht, and M. Valstar. Cold fusion: Calibrated and ordinal latent distribution fusion for uncertainty-aware multimodal emotion recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  • Nagrani et al. [2021] A. Nagrani, S. Yang, A. Arnab, A. Jansen, C. Schmid, and C. Sun. Attention bottlenecks for multimodal fusion. Advances in neural information processing systems, 34:14200–14213, 2021.
  • Li et al. [2020] S. Li, C. Zou, Y. Li, X. Zhao, and Y. Gao. Attention-based multi-modal fusion network for semantic scene completion. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pages 11402–11409, 2020.
  • Chudasama et al. [2022] V. Chudasama, P. Kar, A. Gudmalwar, N. Shah, P. Wasnik, and N. Onoe. M2fnet: Multi-modal fusion network for emotion recognition in conversation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4652–4661, 2022.
  • Li et al. [2022] G. Li, Y. Wei, Y. Tian, C. Xu, J.-R. Wen, and D. Hu. Learning to answer questions in dynamic audio-visual scenarios. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 19108–19118, 2022.
  • Guhur et al. [2023] P.-L. Guhur, S. Chen, R. G. Pinel, M. Tapaswi, I. Laptev, and C. Schmid. Instruction-driven history-aware policies for robotic manipulations. In Conference on Robot Learning, pages 175–187. PMLR, 2023.
  • Li et al. [2023] X. Li, M. Liu, H. Zhang, C. Yu, J. Xu, H. Wu, C. Cheang, Y. Jing, W. Zhang, H. Liu, et al. Vision-language foundation models as effective robot imitators. In The Twelfth International Conference on Learning Representations, 2023.
  • Hochreiter and Schmidhuber [1997] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
  • Du et al. [2022] M. Du, O. Y. Lee, S. Nair, and C. Finn. Play it by ear: Learning skills amidst occlusion through audio-visual imitation learning. arXiv preprint arXiv:2205.14850, 2022.
  • He et al. [2016] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778, 2016.
  • Kingma and Ba [2014] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • Xia et al. [2024] W. Xia, D. Wang, X. Pang, Z. Wang, B. Zhao, D. Hu, and X. Li. Kinematic-aware prompting for generalizable articulated object manipulation with llms. In 2024 IEEE International Conference on Robotics and Automation (ICRA), pages 2073–2080. IEEE, 2024.
  • Anthropic [2024] Anthropic. Claude 3.5 sonnet. Website, 2024. https://www.anthropic.com/news/claude-3-5-sonnet.
  • Chi et al. [2023] C. Chi, S. Feng, Y. Du, Z. Xu, E. Cousineau, B. Burchfiel, and S. Song. Diffusion policy: Visuomotor policy learning via action diffusion. In Proceedings of Robotics: Science and Systems (RSS), 2023.
  • Wei et al. [2024] Y. Wei, R. Feng, Z. Wang, and D. Hu. Enhancing multimodal cooperation via sample-level modality valuation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 27338–27347, 2024.

Appendix

Appendix A Details of the Task Setup

In Sec. 4.1 of the main paper, we briefly introduced the basic information of the pouring task and the peg insertion with keyway task, including the task objectives and stage divisions. In this section, we provide a more detailed introduction to the setup of these two tasks. We control the robot arm through a keyboard to complete the tasks and collect human demonstrations.

A.1 Pouring

Refer to caption
Figure 5: Illustration of the pouring task. We randomly shift the fixed target cylinder sideways by 030\sim 3cm (indicated by the blue arrow) in training demonstrations, and shift by 060\sim 6cm (indicated by the orange arrow) during testing.

Setup Details. For the pouring task, the robot needs to pour tiny steel beads of specific quality from the cylinder in the hand into another cylinder. In the demonstrations, we randomly shift the fixed target cylinder sideways by 030\sim 3cm, while during testing, this range expands to 060\sim 6cm, as illustrated in Fig. 5. Following the previous work [11], we use small beads with a diameter of 1mm to simulate liquids. The initial mass of the beads is 90g/120g, while the target mass for pouring out is 40g/60g, both indicated by prompts. Since the camera cannot capture the interior of the target cylinder, other modalities are needed to assess whether the poured-out quantity meets the target. Hence, we use vision (RGB), audio and touch modalities in this task.

Robot Action Space. In this task, the robot can act in a 2-dimensional action space along the axis xx and ϕ\phi, where xx represents the horizontal movement and ϕ\phi represents the rotation of the gripper. The action step size is Δx=0.5mm\Delta x=0.5mm and Δϕ=0.12\Delta\phi=0.12^{\circ}. There are a total of 55 possible actions (±Δx\pm\Delta x, ±Δϕ\pm\Delta\phi and 0), corresponding to the two directions of the two dimensions of (x,ϕ)(x,\phi) and holding still.

Stage Division. This task consists of four stages [44]: 1) Aligning, where the robot needs to move a graduated cylinder containing beads and align it with the target cylinder, 2) Start Pouring, where the robot rotates the end effector to pour out the beads at an appropriate speed, 3) Holding Still, where the robot maintains its posture to continue pouring out beads steadily, and 4) End Pouring, where the robot rotates the end effector at the appropriate time to stop the flow of beads. In trajectories, we consider the timesteps of the first downward rotation of the gripper (Δϕ-\Delta\phi), the first holding still after rotation, and the first upward rotation of the gripper (+Δϕ+\Delta\phi) as stage transition points.

A.2 Peg Insertion with Keyway.

Refer to caption
Figure 6: Illustration of the peg insertion with keyway task. We randomly shift the fixed base along a 1212cm-long parallel line on the desktop (the yellow arrow), and randomly initialize the position of the robot arm inside a 40cm ×\times 20cm rectangular area (the green rectangle) in training demonstrations. During testing, we fix the position of the base (the red points) and the robot arm (the green points) to several pre-defined points.

Setup Details. This task is an upgraded version of peg insertion, where the robot needs to first align a peg with a key to the keyway on the base by rotating, and then insert the peg fully. The alignment between the key and the keyway in this task primarily relies on tactile feedback, as the camera cannot observe inside the hole. Hence, we use RGB, depth and touch modalities in this task. Considering generalization, we randomly fix the base at any position along a 1212cm-long parallel line on the desktop in demonstrations, as illustrated in Fig. 6. The robot arm holding the peg can also be initialized inside a 40cm ×\times 20cm rectangular area around the base. During testing, to ensure fairness, the positions of the base and the robot arm are several pre-defined points.

Robot Action Space. In this task, the robot arm can move on the three axes x,y,zx,y,z of Cartesian coordinate, where x,yx,y represents the horizontal movement and zz represents the vertical movement. The gripper can also rotate along axis ϕ\phi to align the key with the keyway. Since the vertical movement of the robot arm (Δz-\Delta z) and the rotation of the gripper (+Δϕ+\Delta\phi) are both unidirectional, there are a total of 77 possible actions (±Δx\pm\Delta x, ±Δy\pm\Delta y, Δz-\Delta z, +Δϕ+\Delta\phi and 0).

Stage Division. This task consists of three stages: 1) First Insertion, where the robot aligns the peg with the hole and inserts it until the key collides with the base, 2) Rotating, where the robot aligns the key with the keyway on the base by rotating the peg based on tactile feedback, and 3) Second Insertion, where the robot further inserts the peg to the bottom. Similar to the pouring task, we consider the timesteps of the first gripper rotation (+Δϕ+\Delta\phi) and the first downward movement (Δz-\Delta z) after rotation as stage transition points.

A.3 Generalization Experiments with Distractors

Refer to caption
(a) Color (Pouring)
Refer to caption
(b) Color (Insertion)
Refer to caption
(c) Mess (Insertion)
Figure 7: Illustration of tasks with distractors. For the pouring task, we change the color of the cylinder from white to red (denoted as “Color (Pouring)”). For the peg insertion with keyway task, we respectively change the color of the base from black to green (denoted as “Color (Insertion)”), and scatter some clutter around the base (denoted as “Mess (Insertion)”).

To verify the generalization of our framework to distractors in the environment, we conduct experiments with visual distractors on both tasks. For the pouring task, we change the color of the cylinder from white to red. As for the peg insertion task, we respectively change the color of the base from black to green, and scatter some clutter around the base. These task settings are illustrated in Fig. 7. The distractors only exist during testing. Besides introducing distractors, the settings for these tasks remain consistent with the main experiments.

Appendix B Implementation Details

Following [11], we resize visual and tactile frames to 140×105140\times 105 and randomly crop them to 128×96128\times 96 during training. We also use color jitter for image augmentation. For audio modality, we resample the wave signal at 16kHZ and generate a 64×5064\times 50 mel-spectrogram through short-time Fourier transform, with 400 FFT windows, hop length of 160, and mel bin of 64. We employ ResNet-18 [64] network as the uni-modal encoder. Each encoder takes a brief history of observations spanning T=6T=6 timesteps. We train all the models using Adam [65] with a learning rate of 10410^{-4} for 75 epochs. We perform linear learning rate decay every two epochs, with a decay factor of 0.9.

For action history, we use a buffer of length 200 to store actions. Each action is encoded using one-hot encoding. Action sequences shorter than 200 are padded with zeros. For both tasks, we consider the 15 timesteps near the stage transition point as the soft constraint range (γ=15\gamma=15). We set λ=5.0\lambda=5.0 for both tasks to control the intensity of the penalty. For the learnable stage token [stagei][stage_{i}], we initialize it with the mean of all state tokens on samples within the ii-th stage after warmup training for 1 epoch. We use β=0.5\beta=0.5 for the calculation of the stage-injected state token ztz^{*}_{t}. Moreover, to prevent gradients from becoming too large, we also truncated the gradients of the stage score penalty, restricting them to solely influencing the gate network.

Appendix C Random Attention Blur

Refer to caption
Figure 8: Illustration of random attention blur in the stage-guided dynamic fusion module. We randomly replace the attention scores on all feature tokens with the same average value 1M×T\frac{1}{M\times T} with a probability pp, which is a hyper-parameter.

In order to prevent the model from simply memorizing the actions corresponding to attention score patterns, we introduce a random attention blur mechanism to the stage-guided dynamic fusion module. For each input, we replace the attention scores on all feature tokens with the same average value 1M×T\frac{1}{M\times T} with a probability pp, where MM is the number of modalities and TT is the timestep number of the brief observation history, as illustrated in Fig. 8. We set p=0.25p=0.25 in both tasks.

We introduce this mechanism due to the potential overfitting issue in the model, where the policy head (MLP in the dynamic fusion module) learns the correspondence between the distribution of attention scores and actions. For instance, when using the trained MS-Bot model (without random attention blur) to complete pouring tasks, manually increasing the attention scores on tactile feature tokens invariably leads the model to predict the next action as upward rotation (+Δϕ+\Delta\phi), regardless of the current stage of the task. This phenomenon suggests that the stage comprehension module in the model partially assumes the role of action prediction rather than focusing solely on stage understanding. Therefore, randomly blurring attention scores can compel the policy head to focus on the information from the feature tokens and better decouple the stage comprehension module from the dynamic fusion module.

Appendix D LLM-Based Automatic Stage Labeling

Our strategy for stage labeling is to segment the action sequence of the trajectory based on pre-set rules (such as considering the timestep of the first gripper rotation as a stage transition point). This does indeed require some time to pre-define the stages and design specific rules for division. However, extracting stage information without human annotation is also feasible. One possible approach is to use LLMs, which have demonstrated strong comprehension ability in robotic manipulation [66], for automatic stage annotation of action sequences. We use Claude 3.5 Sonnet [67] as our LLM, and construct prompts to describe the task setup and include an example action sequence to better help the LLM understand the process of this task. We then ask the LLM to provide the specific stage division, as shown in Fig. 9.

Refer to caption
Figure 9: An example of automatic stage division by LLM. We provide the task setup, robot action space and an example action sequence to the LLM to obtain the main stage division. For ease of presentation, the action sequence in the figure has been abbreviated.

After obtaining the stage divisions provided by the LLM, we then construct another prompt and ask the LLM to label the action sequence according to the stage divisions, as shown in Fig. 10. This strategy allows for obtaining stage labels for all samples without requiring human annotation. We find that the stage divisions provided by the LLM are usually more detailed. For example, it divides the peg insertion with keyway task into four stages (splitting our First Insertion stage into two stages: Initial Positioning and Vertical Approach), which is also reasonable. Since our model does not explicitly require that the modality importance must be different between stages, a more detailed stage division is acceptable to some extent.

Refer to caption
Figure 10: An example of automatic stage labeling by LLM. We provide the task setup, robot action space and an target action sequence to the LLM to obtain the stage labels. For ease of presentation, the action sequence in the figure has been abbreviated.

Through this method, we perform automatic stage annotation on our imitation learning dataset. To evaluate the accuracy of the LLM’s annotations, we merge the first two stages divided by the LLM into a single First Insertion stage, and use the human-annotated stage labels as ground truth. The evaluation results show that the accuracy of the stage labels annotated by the LLM is 95.68%95.68\% (10,096 / 10,552, about 15 errors per trajectory).

For tasks like peg insertion with keyway, where many actions are continuous, we can merge consecutive actions to obtain a compressed action sequence like “[action*num,…,action*num]”, as shown in Fig. 11. This approach helps reduce annotation errors caused by LLM counting mistakes and increases the accuracy to 99.65%99.65\% (10,521 / 10,552, about 3 errors per trajectory). These results demonstrate that automatic stage annotation using LLMs without pre-defined stages is feasible. The experiment is only a relatively simple attempt. We believe that with more detailed prompt engineering and the use of more powerful LLMs, the accuracy can be further improved.

Refer to caption
Figure 11: An example of automatic stage labeling by LLM. We provide the task setup, robot action space and an compressed target action sequence to the LLM to obtain the stage labels.

Appendix E Comparison of Attention Scores between MULSA and MS-Bot

Refer to caption
(a) Pouring
Refer to caption
(b) Peg insertion with keyway
Figure 12: Visualization of the aggregated attention scores for each modality and stage scores of MS-Bot in both tasks. At each timestep, we average the attention scores on all feature tokens of each modality separately. The stage score is the softmax normalized output of the gate network.
Refer to caption
(a) Pouring
Refer to caption
(b) Peg insertion with keyway
Figure 13: Visualization of the aggregated attention scores of MULSA for each modality in both tasks. At each timestep, we average the attention scores on all feature tokens of each modality separately. The range of the attention score axis in the figure is consistent with Fig. 12.

In Sec. 4.4 of the main paper, we illustrate the aggregated attention scores for each modality and the stage scores of MS-Bot in the pouring task, as shown in Fig. 12(a). We also record the changes in attention scores and stage scores in the peg insertion with keyway task, as shown in Fig. 12(b). The results in the figure demonstrate that our model accurately predicts the rapid changes in stages for both tasks. As a result, the attention scores across modalities guided by stage comprehension are also relatively stable, exhibiting clear inter-stage changes and minor intra-stage adjustment.

As a comparison, we also visualize the attention scores of another baseline MULSA [11] with self-attention fusion, as shown in Fig. 13. The range of the attention score axis in the figure is consistent with Fig. 12 (0.01.00.0\sim 1.0). It is evident that the attention scores of the modalities in the MULSA model are close and exhibit a small variation along the time axis, lacking clear stage characteristics. This indicates that the MULSA model fails to fully leverage the advantages of dynamic fusion compared to the concat model.

Appendix F Combination with Diffusion Policy

Fusion Module Policy Network Success Rate
Concat MLP 7/15
Concat Diffusion Policy [68] 11/15
Stage-guided Dynamic Fusion MLP 13/15
Stage-guided Dynamic Fusion Diffusion Policy [68] 13/15
Table 4: Comparison of performance with Concat and Diffusion Policy baselines on the peg insertion with keyway task.

To verify the applicability of our proposed stage-guided dynamic multi-sensory fusion to different policy networks, we replace the policy network with the Diffusion Policy [68] and conduct experiments on the peg insertion with keyway task. Specifically, we replace the MLP policy network with an UNet-based Diffusion Policy in the Concat baseline and our MS-Bot model, and evaluate the performance of stage-guided dynamic fusion on models using different policy networks.

The results in Tab. 4 show that replacing the policy network in the Concat baseline with the Diffusion Policy leads to a significant performance improvement, but there is still a gap compared to MS-Bot. This is because the Diffusion Policy still lacks an explicit feature fusion module (apart from simple concatenation). The fusion of multi-sensory features in this baseline still occurs within the policy network, similar to the concat baseline, albeit with a more capable policy network. After further replacing its fusion module with our proposed stage-guided dynamic fusion, it achieves performance comparable to MS-Bot. This demonstrates that our proposed explicit stage understanding and stage-guided dynamic fusion are still effective for models with more powerful policy networks. Our proposed fusion module can help the policy network focus more on generating actions using the already fused multi-sensor features. However, under the same conditions of using stage-guided dynamic fusion, the model with diffusion policy as the policy network does not outperform the model with MLP. We think this may be due to the relatively small scale of data in this task, which does not allow the Diffusion Policy to fully leverage its advantages.

Appendix G Comparison with Stage-Aware and Context-Aware Baselines

Method Success Rate
Concat 7/15
MULSA 11/15
Concat + Stage Auxiliary Task 9/15
MULSA + Stage Auxiliary Task 11/15
MULSA + Observation Context 11/15
MS-Bot 13/15
Table 5: Comparison of performance with stage-aware and context-aware baselines on the peg insertion with keyway task. We introduced stage prediction auxiliary tasks and multi-sensory observation context into the Concat and MULSA baselines for a more comprehensive comparison.

Due to not fully considering the importance of explicit stage understanding in guiding multi-sensory fusion, the original Concat and MULSA baselines does not utilize contextual inputs and additional stage information. To make a fairer comparison with our MS-Bot, we respectively construct Concat and MULSA baselines that incorporate a stage prediction auxiliary task and multi-sensory observation context. Specifically, we add an extra head (an MLP identical to the gate network in MS-Bot) to both the Concat and MULSA models to predict the stage, trained with the same form of loss used to penalize the scores of the gate network in MS-Bot. We also attempt to modify the structure of the MULSA model to accept observation context as input. We input the observations from 5s, 7.5s, 10s, 12.5s, 15s, 17.5s, and 20s before the current timestep into the model, and encode these observation contexts by each modality’s encoder into 7 tokens (per modality). These tokens are then concatenated with the original 6 tokens of the current timestep observation of each modality, resulting in a total of 13 tokens for each modality. We evaluate these baselines on the peg insertion with keyway task. The results in Tab. 5 shows that introducing an explicit stage prediction task could improve the performance of the Concat baseline, but it provides minimal benefit for the MULSA method. These baselines still have a performance gap compared to our MS-Bot. This indicates that explicit stage understanding is meaningful, but introducing context to aid stage understanding and specially designed stage-guided dynamic fusion modules are also necessary. In addition, the results show that inputting observation contexts into MULSA results in virtually no improvement in the model’s performance. This indicates that mixing the contexts and current observations in self-attention calculations is highly inefficient for helping the model understand stages and allocate modality weights, as it reduces focus on the more important current observations and significantly increases GPU memory usage. In contrast, our MS-Bot leverages a long action history as context to enhance its understanding of task stages. This approach simplifies processing for the model and offers finer granularity, effectively guiding dynamic multi-sensor fusion.

Appendix H Comparison under the Multi-Camera Setting

Method Success Rate
Concat 9/15
MULSA 12/15
MS-Bot 14/15
Table 6: Comparison under the setting of multiple cameras on the peg insertion with keyway task.

In robotic manipulation, a common setup involves using multiple cameras to capture visual information from different views. By treating visual images from different views as distinct modalities, our MS-Bot is also able to handle multi-camera observations. To verify this capability, we conduct a comparison with baseline methods on the peg insert with keyway task under the setting with multiple cameras. Specifically, we place another camera on the side of the base, so that the lines connecting these two cameras to the base roughly form a right angle. We replace the depth observation from the first camera with the RGB observation from the second camera, as the RGB observations from these two cameras already provide sufficient spatial information for this task. The results in Tab. 6 show that in the multi-camera setup, our MS-Bot method still outperforms the baselines. This demonstrates that our proposed stage-guided dynamic fusion can also be applied to fuse features from multiple cameras.

Appendix I Combination with Audio Modality

Method Success Rate
Concat 8/15
MULSA 11/15
MS-Bot 13/15
Table 7: Comparison on the peg insertion with keyway task using RGB, depth, tactile and audio.

For tasks involving using lock-picking tools to unlock in real life, audio can indeed serve as the basis for completing the task, as these locks typically have specific mechanical structures that produce distinct sounds. Similarly, audio can indicate whether the key and keyway are aligned, thereby assisting with the peg insertion with keyway task. To verify the effectiveness of our model after further incorporating audio, we conduct experiments on the peg insertion with keyway task using RGB, depth, tactile and audio modality. The results in Tab. 7 show that the addition of the audio modality only brings improvement to the weakest Concat baseline. For MULSA and our MS-Bot method, the benefits are minimal. This is because the audio in this task is short and minimal due to the 3D-printed materials and is vulnerable to noise disturbance.

Appendix J Detailed Experimental Results of Pouring

Method Pouring Initial (g) Pouring Target (g)
90 120 40 60
MS-Bot 1.60 ±\pm 1.10 5.58 ±\pm 1.79 6.48 ±\pm 1.55 1.80 ±\pm 0.95
- Attention Blur 1.72 ±\pm 1.09 5.70 ±\pm 1.74 6.55 ±\pm 1.38 1.95 ±\pm 1.32
- Stage Comprehension 2.52 ±\pm 1.12 6.15 ±\pm 1.64 6.80 ±\pm 1.59 2.92 ±\pm 1.40
- State Tokenizer 3.05 ±\pm 1.01 6.42 ±\pm 1.98 7.12 ±\pm 1.66 4.19 ±\pm 1.24
Table 8: Impact of each component of our framework in the pouring task (mean ± standard deviation). ‘-’ indicates further removing the module from the model in the previous line.
Method Pouring Initial (g) Pouring Target (g)
90 120 40 60
Concat 8.75 ±\pm 2.02 10.69 ±\pm 2.45 10.72 ±\pm 2.35 8.46 ±\pm 2.03
Du et al. [63] 8.51 ±\pm 1.79 9.54 ±\pm 2.10 9.98 ±\pm 2.20 8.04 ±\pm 1.85
MULSA [11] 4.72 ±\pm 1.30 7.83 ±\pm 1.71 8.45 ±\pm 1.50 5.56 ±\pm 1.13
MS-Bot 2.04 ±\pm 1.40 6.10 ±\pm 1.49 7.62 ±\pm 1.94 2.41 ±\pm 1.47
Table 9: Comparison of performance in scenes with visual distractors in the pouring task(mean ± standard deviation). We change the color of the cylinder from white to red during testing.

In Sec. 4.4 and 4.5 of the main paper, we show the overall error on all settings of the pouring task. In this section, we comprehensively present the detailed result of the component ablation and the distractor experiment on all the settings of the pouring task in Tab. 8 and 9. As shown in Tab. 8, the contributions from both the state tokenizer and the stage comprehension module on all settings demonstrate the importance of the coarse-to-fine stage comprehension. The consistent lead of MS-BOT across all settings of the pouring task in Tab. 3 also demonstrates the generalizability of stage comprehension.

Appendix K Evaluation of Hyper-parameter Settings

Refer to caption
(a) Stage Score Penalty Intensity λ\lambda
Refer to caption
(b) Soft Constraint Range γ\gamma
Refer to caption
(c) Attention Blur Probability pp
Figure 14: Performance of MS-Bot on one setting of the pouring task when changing the hyper-parameters λ\lambda, γ\gamma and pp. We fix the others when changing one hyper-parameter. We report the mean error of the concat model, MULSA and MS-Bot. The error bars represent the standard deviation of errors for MS-Bot.

In this section, we test the sensitivity of MS-Bot to different hyper-parameter settings. Specifically, we test the performance of MS-Bot under different penalty intensity λ\lambda, soft constraint range γ\gamma and probability of random attention blur pp in one setting of the pouring task. We train all the models to pour out 40g of small beads with different initial masses (90g/120g). We set λ=5.0\lambda=5.0, γ=15\gamma=15 and p=0.25p=0.25 by default, and keep the others fixed when we modify one of the hyper-parameters.

We present the evaluation results in Fig. 14. Our MS-Bot consistently outperforms both the concat model and MULSA across various hyper-parameter settings, indicating that the performance of our MS-Bot is relatively stable to hyper-parameter variations. However, we also observe that when setting λ\lambda and γ\gamma to small values (λ=1.0\lambda=1.0 in Fig. 14(a) and γ=0\gamma=0 in Fig. 14(b)), the performance of MS-Bot drops and becomes closer to MULSA with simple self-attention fusion. This is because when λ\lambda is too small, the prediction of the current stage becomes inaccurate due to the insufficient training. When γ\gamma is too small, the model is forced to make drastic changes in the stage score prediction within very few timesteps, leading to unstable stage predictions. Both of these factors weaken the efficacy of stage comprehension within MS-Bot. We also find that setting too large pp (p>0.3p>0.3 in Fig. 14(c)) can bring negative impacts as the attention blur truncates the gradients backpropagated to the stage comprehension module and the state tokenizer. Excessive attention blur can impair the training of these two modules.

Appendix L Limitations and Future Improvements

In this section, we discuss the potential limitations of our work and the corresponding solutions:

In many robotic tasks this kind of dynamic weighting is not necessary. Indeed, there are many tasks that can be completed with just the visual modality, where readings from other sensors may not provide useful information. We believe that one solution is to let the model first select the necessary subset of sensors based on the task (possibly using an LLM), and then perform dynamic multi-sensor fusion based on stage understanding. For tasks with only one stage, this can be considered as a special case where |S|=1|S|=1.

The stage labeling may be costly. Currently, our strategy for stage labeling is to segment the action sequence of the trajectory based on pre-set rules. This requires time to set up detailed rules and becomes more challenging when there are more stages and more complex stage transitions. A possible solution is the LLM-based automatic labeling method mentioned earlier, although a small amount of human effort is still necessary, such as constructing prompts to accurately describe the task setup and providing example for in-context learning.

The lack of comparison to larger models. Currently, there is no well-trained open-source multi-sensory large foundation model for robot manipulation (including vision, hearing, touch, etc.). Therefore, we are temporarily unable to verify the role of explicit stage understanding in large models, nor can we directly compare our model with multi-sensory large models. However, we believe that even for larger models, explicit stage understanding remains meaningful when introducing new modalities and training is required using a small amount of paired multi-sensory data, or when fine-tuning in new scenarios with limited data scale.

Robustness issue in the case of missing modalities. Since in both of our tasks there are stages that depend on a single modality, masking uni-modal features during testing may directly lead to failure. This is a problem common to all multi-sensory models, and we did not make improvements specifically for this issue. In cases where multiple sensors complement each other, the absence of uni-modal features may have a compounding effect on stage understanding and dynamic fusion. We believe that one possible solution is to randomly dropout uni-modal features during the training process [69]. Some other methods for robust multi-modal learning could also be referenced.

Transferability of the current dynamic multi-sensory fusion architecture. Since the multi-sensory fusion architecture we use in the dynamic fusion module is based on cross-attention rather than self-attention, there are difficulties in directly transferring the original form of this fusion method to transformers with multiple self-attention blocks. One potential strategy is to use the stage token as a special token to replace the class token. It could be used to predict the stage under the supervision of stage labels and thereby achieving the effect of stage-guided multi-modal fusion.

Low integration level of current multi-sensory robot system. Our current multi-sensory robot system has not well integrated the various sensors: the camera is positioned at a third-person perspective, the microphone is placed on the desktop, and the tactile sensor is at the end of the robotic arm. They are located in different positions. This low level of integration means that the system can only complete relatively simple tasks, while more complex multi-sensory tasks that require mobility cannot be accomplished. This multi-sensory system still needs further refinement and upgrading.