| 1 |
TextOp: Real-time Interactive Text-Driven Humanoid Robot Motion Generation and Control |
TextOp:实时交互式文本驱动的人形机器人运动生成与控制框架 |
humanoid humanoid robot humanoid control |
|
|
| 2 |
Bridging Speech, Emotion, and Motion: a VLM-based Multimodal Edge-deployable Framework for Humanoid Robots |
提出SeM²框架,利用VLM驱动人形机器人实现情感丰富的多模态交互,并支持边缘部署。 |
humanoid humanoid robot multimodal |
|
|
| 3 |
VividFace: Real-Time and Realistic Facial Expression Shadowing for Humanoid Robots |
VividFace:面向人型机器人的实时逼真面部表情模仿系统 |
humanoid humanoid robot human-to-humanoid |
✅ |
|
| 4 |
Differentiate-and-Inject: Enhancing VLAs via Functional Differentiation Induced by In-Parameter Structural Reasoning |
提出iSTAR框架,通过参数空间结构推理增强视觉-语言-动作模型的任务级理解能力。 |
manipulation vision-language-action VLA |
|
|
| 5 |
UEREBot: Learning Safe Quadrupedal Locomotion under Unstructured Environments and High-Speed Dynamic Obstacles |
UEREBot:学习非结构化环境和高速动态障碍下的四足机器人安全运动 |
quadruped locomotion Unitree |
|
|
| 6 |
Trace-Focused Diffusion Policy for Multi-Modal Action Disambiguation in Long-Horizon Robotic Manipulation |
提出Trace-Focused Diffusion Policy,解决长时程机器人操作中的多模态动作歧义问题。 |
manipulation diffusion policy |
|
|
| 7 |
Going with the Flow: Koopman Behavioral Models as Implicit Planners for Visuo-Motor Dexterity |
提出基于Koopman行为模型的统一行为模型(UBM),用于灵巧操作中的隐式规划。 |
manipulation dexterous manipulation |
|
|
| 8 |
Why Look at It at All?: Vision-Free Multifingered Blind Grasping Using Uniaxial Fingertip Force Sensing |
提出一种仅依赖单轴指尖力感知的无视觉多指盲抓取方法 |
manipulation teacher-student |
|
|