cs.LG(2025-12-31)
📊 共 22 篇论文 | 🔗 2 篇有代码
🎯 兴趣领域导航
支柱九:具身大模型 (Embodied Foundation Models) (13 🔗1)
支柱二:RL算法与架构 (RL & Architecture) (8 🔗1)
支柱八:物理动画 (Physics-based Animation) (1)
🔬 支柱九:具身大模型 (Embodied Foundation Models) (13 篇)
🔬 支柱二:RL算法与架构 (RL & Architecture) (8 篇)
| # | 题目 | 一句话要点 | 标签 | 🔗 | ⭐ |
|---|---|---|---|---|---|
| 14 | Dichotomous Diffusion Policy Optimization | 提出DIPOLE:一种用于稳定和可控扩散策略优化的强化学习算法。 | reinforcement learning diffusion policy vision-language-action | ||
| 15 | Sparse Offline Reinforcement Learning with Corruption Robustness | 提出基于稀疏鲁棒估计的Actor-Critic算法,解决离线稀疏RL中的数据污染问题。 | reinforcement learning offline RL offline reinforcement learning | ||
| 16 | From Perception to Punchline: Empowering VLM with the Art of In-the-wild Meme | 提出HUMOR框架,赋能VLM生成更幽默、符合人类偏好的野生表情包 | reinforcement learning HuMoR multimodal | ||
| 17 | GRL-SNAM: Geometric Reinforcement Learning with Path Differential Hamiltonians for Simultaneous Navigation and Mapping in Unknown Environments | 提出GRL-SNAM,通过几何强化学习与路径微分哈密顿量实现未知环境下的同步定位与建图。 | reinforcement learning policy learning | ✅ | |
| 18 | Many Minds from One Model: Bayesian-Inspired Transformers for Population Diversity | 提出Population Bayesian Transformers,从单一预训练LLM中采样多样化模型实例,提升生成多样性。 | reinforcement learning large language model | ||
| 19 | ResponseRank: Data-Efficient Reward Modeling through Preference Strength Learning | ResponseRank:通过偏好强度学习实现数据高效的奖励建模 | reinforcement learning preference learning RLHF | ||
| 20 | Attribution-Guided Distillation of Matryoshka Sparse Autoencoders | 提出DMSAE,通过归因引导蒸馏Matryoshka稀疏自编码器,提升特征一致性和可迁移性。 | distillation | ||
| 21 | Robust Bayesian Dynamic Programming for On-policy Risk-sensitive Reinforcement Learning | 提出鲁棒贝叶斯动态规划,用于解决策略风险敏感强化学习中的转移不确定性问题 | reinforcement learning |
🔬 支柱八:物理动画 (Physics-based Animation) (1 篇)
| # | 题目 | 一句话要点 | 标签 | 🔗 | ⭐ |
|---|---|---|---|---|---|
| 22 | Learning Temporally Consistent Turbulence Between Sparse Snapshots via Diffusion Models | 提出基于扩散模型的稀疏快照间湍流重建方法 | spatiotemporal |