| 1 |
ACDC: Autoregressive Coherent Multimodal Generation using Diffusion Correction |
提出ACDC,结合自回归模型与扩散模型,实现高质量连贯的多模态生成。 |
large language model multimodal |
|
|
| 2 |
fLSA: Learning Semantic Structures in Document Collections Using Foundation Models |
fLSA:利用基础模型学习文档集合中的语义结构,提升文本重建与生成质量。 |
large language model foundation model |
✅ |
|
| 3 |
From Sparse Dependence to Sparse Attention: Unveiling How Chain-of-Thought Enhances Transformer Sample Efficiency |
揭示CoT提升Transformer样本效率的机制:从稀疏依赖到稀疏注意力 |
large language model chain-of-thought |
|
|
| 4 |
RespLLM: Unifying Audio and Text with Multimodal LLMs for Generalized Respiratory Health Prediction |
RespLLM:利用多模态LLM统一音频和文本,实现广义呼吸系统健康预测 |
large language model multimodal |
|
|
| 5 |
Compression via Pre-trained Transformers: A Study on Byte-Level Multimodal Data |
利用预训练Transformer实现高效数据压缩,超越传统算法。 |
foundation model multimodal |
|
|
| 6 |
Wireless-Friendly Window Position Optimization for RIS-Aided Outdoor-to-Indoor Networks based on Multi-Modal Large Language Model |
提出基于多模态大语言模型的无线友好型窗户位置优化方法,用于RIS辅助的室外到室内网络。 |
large language model |
|
|
| 7 |
PrefixQuant: Eliminating Outliers by Prefixed Tokens for Large Language Models Quantization |
PrefixQuant通过前缀化异常token解决LLM量化中的token级离群点问题 |
large language model |
✅ |
|
| 8 |
GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models |
GSM-Symbolic:揭示大语言模型在数学推理上的局限性 |
large language model |
|
|
| 9 |
Recent Advances of Multimodal Continual Learning: A Comprehensive Survey |
首个多模态持续学习综述,系统梳理方法并展望未来方向。 |
multimodal |
✅ |
|
| 10 |
TLDR: Token-Level Detective Reward Model for Large Vision Language Models |
提出TLDR:一种Token级别判别奖励模型,提升大型视觉语言模型性能。 |
large language model multimodal |
|
|
| 11 |
AnyAttack: Towards Large-scale Self-supervised Adversarial Attacks on Vision-language Models |
AnyAttack:面向视觉-语言模型的大规模自监督对抗攻击框架 |
foundation model multimodal |
|
|
| 12 |
Chain and Causal Attention for Efficient Entity Tracking |
提出链式与因果注意力机制,高效解决Transformer在实体追踪任务中的局限性 |
large language model |
|
|
| 13 |
Transformers learn variable-order Markov chains in-context |
研究Transformer上下文学习变阶马尔可夫链能力,并提出CTW算法的Transformer构造。 |
large language model |
|
|
| 14 |
Transformers are Efficient Compilers, Provably |
证明Transformer能以对数复杂度高效编译类C语言,优于RNN |
large language model |
|
|
| 15 |
SecAlign: Defending Against Prompt Injection with Preference Optimization |
SecAlign:利用偏好优化防御大语言模型的提示注入攻击 |
large language model |
✅ |
|
| 16 |
Can LLMs Understand Time Series Anomalies? |
探索LLM在时间序列异常检测中的能力,揭示其理解机制与局限性 |
large language model |
✅ |
|
| 17 |
Density estimation with LLMs: a geometric investigation of in-context learning trajectories |
利用LLM进行密度估计:上下文学习轨迹的几何分析 |
large language model |
✅ |
|
| 18 |
Interactive Event Sifting using Bayesian Graph Neural Networks |
提出基于贝叶斯图神经网络的交互式事件筛选方法,用于法庭分析中社交媒体数据的快速过滤。 |
multimodal |
|
|
| 19 |
Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild |
Model-GLUE:面向大规模模型库的普适性LLM扩展方案 |
large language model |
✅ |
|
| 20 |
TidalDecode: Fast and Accurate LLM Decoding with Position Persistent Sparse Attention |
TidalDecode:利用位置持久稀疏注意力加速LLM解码并保持精度 |
large language model |
|
|
| 21 |
TOAST: Transformer Optimization using Adaptive and Simple Transformations |
TOAST:利用自适应简单变换优化Transformer,无需额外训练。 |
foundation model |
|
|
| 22 |
Strong Model Collapse |
揭示大规模模型训练中由合成数据引起的强模型崩溃现象 |
large language model |
|
|
| 23 |
Generating CAD Code with Vision-Language Models for 3D Designs |
提出CADCodeVerify,利用视觉-语言模型迭代验证并改进CAD代码生成的3D对象。 |
large language model |
|
|
| 24 |
Deeper Insights Without Updates: The Power of In-Context Learning Over Fine-Tuning |
ICL在隐式模式学习上优于微调,无需模型更新即可获得更深层次的理解 |
large language model |
|
|