| 1 |
SSAM: Singular Subspace Alignment for Merging Multimodal Large Language Models |
提出SSAM,通过奇异子空间对齐实现多模态大语言模型的无训练融合 |
large language model multimodal |
|
|
| 2 |
Riemannian Geometry Speaks Louder Than Words: From Graph Foundation Model to Next-Generation Graph Intelligence |
提出Riemannian Foundation Model (RFM),利用黎曼几何构建下一代图智能。 |
large language model foundation model |
|
|
| 3 |
AdditiveLLM2: A Multi-modal Large Language Model for Additive Manufacturing |
提出AdditiveLLM2,一种面向增材制造的多模态大语言模型,通过领域自适应预训练实现专业化。 |
large language model |
|
|
| 4 |
Extending Precipitation Nowcasting Horizons via Spectral Fusion of Radar Observations and Foundation Model Priors |
PW-FouCast:通过雷达观测与气象大模型先验的频谱融合,扩展降水临近预报时效 |
foundation model |
✅ |
|
| 5 |
Towards Multimodal Time Series Anomaly Detection with Semantic Alignment and Condensed Interaction |
提出MindTS模型,通过语义对齐和精简交互实现多模态时间序列异常检测。 |
multimodal |
✅ |
|
| 6 |
ROM: Real-time Overthinking Mitigation via Streaming Detection and Intervention |
提出ROM,通过流式检测和干预缓解大型推理模型中的过度思考问题。 |
large language model chain-of-thought |
|
|
| 7 |
Data-Free Layer-Adaptive Merging via Fisher Information for Long-to-Short Reasoning LLMs |
提出基于Fisher信息的层自适应模型融合方法,提升长链推理LLM性能。 |
large language model chain-of-thought |
|
|
| 8 |
Noise Titration: Exact Distributional Benchmarking for Probabilistic Time Series Forecasting |
提出噪声滴定方法以解决时间序列预测的评估问题 |
foundation model |
|
|
| 9 |
SPA: A Simple but Tough-to-Beat Baseline for Knowledge Injection |
SPA:一种简单但效果极佳的知识注入基线方法 |
large language model |
✅ |
|
| 10 |
Revisiting Quantum Code Generation: Where Should Domain Knowledge Live? |
通过推理时增强,提升LLM在量子代码生成中的性能,无需领域微调。 |
large language model |
|
|
| 11 |
Causal Evidence that Language Models use Confidence to Drive Behavior |
揭示大语言模型利用置信度驱动行为决策,为自主智能体发展奠定基础 |
large language model |
|
|
| 12 |
Holistic Scaling Laws for Optimal Mixture-of-Experts Architecture Optimization |
提出MoE架构优化框架,通过联合约束和降维搜索,实现任意计算预算下的最优架构配置。 |
large language model |
|
|
| 13 |
Thinking Deeper, Not Longer: Depth-Recurrent Transformers for Compositional Generalization |
提出深度递归Transformer,解决Transformer在组合泛化中计算深度受限问题 |
chain-of-thought |
|
|
| 14 |
Kolmogorov Complexity Bounds for LLM Steganography and a Perplexity-Based Detection Proxy |
提出基于柯尔莫哥洛夫复杂度的LLM隐写术理论界限及基于困惑度的检测代理。 |
large language model |
|
|
| 15 |
Generalization Limits of In-Context Operator Networks for Higher-Order Partial Differential Equations |
ICONs模型扩展至高阶偏微分方程,保持解的动态特性 |
foundation model |
|
|