RoboBrain A Unified Brain Model for Robotic Manipulation from Abstract to Concrete

Source

@misc{ji_2025_robobrain,
    title={RoboBrain: A Unified Brain Model for Robotic Manipulation from Abstract to Concrete},
    author={Yuheng Ji and Huajie Tan and Jiayu Shi and Xiaoshuai Hao and Yuan Zhang and Hengyuan Zhang and Pengwei Wang and Mengdi Zhao and Yao Mu and Pengju An and Xinda Xue and Qinghang Su and Huaihai Lyu and Xiaolong Zheng and Jiaming Liu and Zhongyuan Wang and Shanghang Zhang},
      year={2025},
      eprint={2502.21257},
      archivePrefix={arXiv},
      primaryClass={cs.RO},
      url={https://arxiv.org/abs/2502.21257}, 
}

(Peking University, Chinese Academy of Sciences) | arXiv

TL;DR

General pipeline

Flash Reading

References

Extension

The starting point is to, instead of updating the full weight matrix $W$, we freeze $W$ and inject trainable rank decomposition matrices $A$ and $B$ such that the weight update $\Delta W = BA$ (the rank of $A$ and $B$ can be chosen). The weight update is then added to the original weight matrix: $W’ = W + \Delta W$. This approach reduces the number of trainable parameters and allows for efficient fine-tuning of large models. The theory is that the update matrix $\Delta W$ lies in a low-rank subspace, which is often sufficient for adapting large models to new tasks.

Another hyperparameter is the scaling factor $\alpha$. The weight update is scaled as $W’ = W + \alpha \Delta W$. This helps to control the magnitude of the updates and can improve training stability. Dropout can also be applied to the intermediate activations to prevent overfitting.