cs.CL(2025-12-23)
📊 共 20 篇论文 | 🔗 5 篇有代码
🎯 兴趣领域导航
支柱九:具身大模型 (Embodied Foundation Models) (11 🔗1)
支柱二:RL算法与架构 (RL & Architecture) (7 🔗4)
支柱一:机器人控制 (Robot Control) (2)
🔬 支柱九:具身大模型 (Embodied Foundation Models) (11 篇)
🔬 支柱二:RL算法与架构 (RL & Architecture) (7 篇)
| # | 题目 | 一句话要点 | 标签 | 🔗 | ⭐ |
|---|---|---|---|---|---|
| 12 | Nemotron 3 Nano: Open, Efficient Mixture-of-Experts Hybrid Mamba-Transformer Model for Agentic Reasoning | 提出Nemotron 3 Nano,一种高效的混合专家Mamba-Transformer模型,用于Agent推理。 | Mamba | ||
| 13 | FaithLens: Detecting and Explaining Faithfulness Hallucination | 提出FaithLens,用于检测并解释大语言模型中的忠实性幻觉问题。 | reinforcement learning large language model | ||
| 14 | Fun-Audio-Chat Technical Report | Fun-Audio-Chat:通过双分辨率语音表示和核心鸡尾酒训练,实现高效且强大的大型音频语言模型 | DPO instruction following | ✅ | |
| 15 | Multi-hop Reasoning via Early Knowledge Alignment | 提出早期知识对齐(EKA)模块,提升迭代RAG多跳推理性能与效率 | reinforcement learning large language model | ✅ | |
| 16 | Memory-T1: Reinforcement Learning for Temporal Reasoning in Multi-session Agents | Memory-T1:利用强化学习进行多轮对话Agent中的时序推理 | reinforcement learning | ✅ | |
| 17 | Distilling to Hybrid Attention Models via KL-Guided Layer Selection | 提出基于KL散度的层选择方法,用于将Softmax注意力Transformer蒸馏为混合注意力模型。 | linear attention distillation | ||
| 18 | SpidR: Learning Fast and Stable Linguistic Units for Spoken Language Models Without Supervision | SpidR:一种无需监督即可学习快速稳定语音单元的语音语言模型 | representation learning distillation | ✅ |
🔬 支柱一:机器人控制 (Robot Control) (2 篇)
| # | 题目 | 一句话要点 | 标签 | 🔗 | ⭐ |
|---|---|---|---|---|---|
| 19 | Semantic Deception: When Reasoning Models Can't Compute an Addition | 提出语义欺骗框架,揭示LLM在符号推理中易受语义误导的缺陷 | manipulation large language model chain-of-thought | ||
| 20 | AprielGuard | 提出AprielGuard,统一安全风险与对抗威胁,提升LLM安全防护能力 | manipulation large language model |