Robotics
Seed-Robotics 团队致力于研究攻克通用智能机器人有挑战的课题,研发业界领先技术,聚焦机器人基础模型、智能感知-操作-交互方面,孵化智能机器人系统

课题方向

机器人多模态大模型
从事大规模机器人基础模型的预训练、微调和优化研究,以模型为中心,提升机器人的智能边界;探索多模态大模型前沿问题,推动在机器人上的大规模扩展,包括但不限于抓取操作、运动控制、世界建模等
Foundation Model
Multimodal
Large-scale Application
Foundation Model
Multimodal

机器人强化学习
围绕机器人在深度强化学习开展前沿技术研究工作,推动最新强化学习算法在机器人上的大规模扩展
Reinforcement Learning
Large-scale Application
Reinforcement Learning
Large-scale Application

机器人数据算法
突破机器人的数据瓶颈,探索数据采集新硬件与新系统,探索真机、仿真等机器人数据生成方式,参与全链路数据闭环建设,构建自动化、大规模的机器人数据引擎
Data
Agent
Training
Data
Agent
精选论文

2025.12.02
GR-RL: Going Dexterous and Precise for Long-Horizon Robotic Manipulation
We present GR-RL, a robotic learning framework that turns a generalist vision-language-action (VLA) policy into a highly capable specialist for long-horizon dexterous manipulation. Assuming the optimality of human demonstrations is core to existing VLA policies. However, we claim that in highly dexterous and precise manipulation tasks, human demonstrations are noisy and suboptimal. GR-RL proposes a multi-stage training pipeline that filters, augments, and reinforces the demonstrations by reinforcement learning. First, GR-RL learns a vision-language-conditioned task progress, filters the demonstration trajectories, and only keeps the transitions that contribute positively to the progress. Specifically, we show that by directly applying offline RL with sparse reward, the resulting Q-values can be treated as a robust progress function. Next, we introduce morphological symmetry augmentation that greatly improves the generalization and performance of GR-RL. Lastly, to better align the VLA policy with its deployment behaviors for high-precision control, we perform online RL by learning a latent space noise predictor. With this pipeline, GR-RL is, to our knowledge, the first learning-based policy that can autonomously lace up a shoe by threading shoelaces through multiple eyelets with an 83.3% success rate, a task requiring long-horizon reasoning, millimeter-level precision, and compliant soft-body interaction. We hope GR-RL provides a step toward enabling generalist robot foundations models to specialize into reliable real-world experts.
Yunfei Li, Xiao Ma, Jiafeng Xu, Yu Cui, Zhongren Cui, Zhigang Han, Liqun Huang, Tao Kong, Yuxiao Liu, Hao Niu, Wanli Peng, Jingchao Qiao, Zeyu Ren, Haixin Shi, Zhi Su, Jiawen Tian, Yuyang Xiao, Shenyu Zhang, Liwei Zheng, Hang Li, Yonghui Wu
Robotics
2025.12.02
GR-RL: Going Dexterous and Precise for Long-Horizon Robotic Manipulation
Yunfei Li, Xiao Ma, Jiafeng Xu, Yu Cui, Zhongren Cui, Zhigang Han, Liqun Huang, Tao Kong, Yuxiao Liu, Hao Niu, Wanli Peng, Jingchao Qiao, Zeyu Ren, Haixin Shi, Zhi Su, Jiawen Tian, Yuyang Xiao, Shenyu Zhang, Liwei Zheng, Hang Li, Yonghui Wu
Robotics

2025.07.21
GR-3 Technical Report
We report our recent progress towards building generalist robot policies, the development of GR-3. GR-3 is a large-scale vision-language-action (VLA) model. It showcases exceptional capabilities in generalizing to novel objects, environments, and instructions involving abstract concepts. Furthermore, it can be efficiently fine-tuned with minimal human trajectory data, enabling rapid and cost-effective adaptation to new settings. GR-3 also excels in handling long-horizon and dexterous tasks, including those requiring bi-manual manipulation and mobile movement, showcasing robust and reliable performance. These capabilities are achieved through a multi-faceted training recipe that includes co-training with web-scale vision-language data, efficient fine-tuning from human trajectory data collected via VR devices, and effective imitation learning with robot trajectory data. In addition, we introduce ByteMini, a versatile bi-manual mobile robot designed with exceptional flexibility and reliability, capable of accomplishing a wide range of tasks when integrated with GR-3. Through extensive real-world experiments, we show GR-3 surpasses the state-of-the-art baseline method, π0, on a wide variety of challenging tasks. We hope GR-3 can serve as a step towards building generalist robots capable of assisting humans in daily life.
Seed Robotics Team
Robotics
2025.07.21
GR-3 Technical Report
Seed Robotics Team
Robotics

2025.07.04
Dexterous Teleoperation of 20-DoF ByteDexter Hand via Human Motion Retargeting
Replicating human--level dexterity remains a fundamental robotics challenge, requiring integrated solutions from mechatronic design to the control of high degree--of--freedom (DoF) robotic hands. While imitation learning shows promise in transferring human dexterity to robots, the efficacy of trained policies relies on the quality of human demonstration data. We bridge this gap with a hand--arm teleoperation system featuring: (1) a 20--DoF linkage--driven anthropomorphic robotic hand for biomimetic dexterity, and (2) an optimization--based motion retargeting for real--time, high--fidelity reproduction of intricate human hand motions and seamless hand--arm coordination. We validate the system via extensive empirical evaluations, including dexterous in-hand manipulation tasks and a long--horizon task requiring the organization of a cluttered makeup table randomly populated with nine objects. Experimental results demonstrate its intuitive teleoperation interface with real--time control and the ability to generate high--quality demonstration data. Please refer to the accompanying video for further details.
Ruoshi Wen, Jiajun Zhang, Guangzeng Chen, Zhongren Cui, Min Du, Yang Gou, Zhigang Han, Junkai Hu, Liqun Huang, Hao Niu, Wei Xu, Haoxiang Zhang, Zhengming Zhu, Hang Li, Zeyu Ren
Robotics
2025.07.04
Dexterous Teleoperation of 20-DoF ByteDexter Hand via Human Motion Retargeting
Ruoshi Wen, Jiajun Zhang, Guangzeng Chen, Zhongren Cui, Min Du, Yang Gou, Zhigang Han, Junkai Hu, Liqun Huang, Hao Niu, Wei Xu, Haoxiang Zhang, Zhengming Zhu, Hang Li, Zeyu Ren
Robotics
查看更多论文
技术能力展示

Seed GR-RL
GR-RL 是一个面向长周期灵巧操作的强化学习框架,让机器人在真实场景中稳定完成多步骤、高精度的操作任务,在业界首次实现“让机器人给整只鞋连续穿鞋带”任务。它还具备自动重试、主动调整场景、识别错位并纠偏的能力;也可应对不同颜色、尺寸和材质的鞋子。

Seed GR-3
GR-3 是一个大规模的视觉 - 语言 - 动作(VLA)模型。它对新物体、新环境以及含抽象概念的新指令展现出较好的泛化能力。GR-3 支持少量人类轨迹数据的高效微调,可快速且经济地适应新任务。同时,在处理长周期和灵巧性任务上,也展现出稳健、可靠的性能。

ByteDexter
ByteDexter 作为自研高主动自由度灵巧手 (21DoF),搭载了高灵敏度触觉传感器,搭配灵巧手遥操作系统,能够实现复杂人类手部动作的实时高保真复现,以及手&臂协同的无缝衔接。
热招岗位
具身智能大模型负责人-Seed
具身智能3D仿真专家-Seed
机器人多模态交互算法研究员-Seed
机器人产品负责人-Seed
机器人工程技术负责人-Seed
机器人运动控制算法工程师-Seed