| 1 |
TS-RAG: Retrieval-Augmented Generation based Time Series Foundation Models are Stronger Zero-Shot Forecaster |
提出TS-RAG,一种基于检索增强生成的时间序列基础模型,显著提升零样本预测能力。 |
large language model foundation model |
✅ |
|
| 2 |
Transferable Foundation Models for Geometric Tasks on Point Cloud Representations: Geometric Neural Operators |
提出几何神经算子(GNP)作为点云几何任务的可迁移基础模型。 |
foundation model |
|
|
| 3 |
Wanda++: Pruning Large Language Models via Regional Gradients |
Wanda++:利用区域梯度剪枝大语言模型,显著提升性能。 |
large language model |
|
|
| 4 |
Incentivizing Multi-Tenant Split Federated Learning for Foundation Models at the Network Edge |
提出PRINCE机制,激励多租户分割联邦学习在边缘侧高效微调基础模型。 |
foundation model |
|
|
| 5 |
Predictable Scale: Part I, Step Law -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining |
提出Step Law:大规模语言模型预训练超参数优化通用Scaling Law |
large language model |
✅ |
|
| 6 |
Leveraging Large Language Models to Address Data Scarcity in Machine Learning: Applications in Graphene Synthesis |
利用大语言模型解决石墨烯合成中机器学习的数据稀缺问题 |
large language model |
|
|
| 7 |
Large Language Models for Zero-shot Inference of Causal Structures in Biology |
利用大型语言模型零样本推断生物学因果结构 |
large language model |
|
|
| 8 |
The Challenge of Identifying the Origin of Black-Box Large Language Models |
提出PlugAE:一种主动追踪黑盒大语言模型来源的技术 |
large language model |
|
|
| 9 |
RCRank: Multimodal Ranking of Root Causes of Slow Queries in Cloud Database Systems |
RCRank:提出云数据库系统中慢查询根因多模态排序方法,提升问题诊断与修复效率。 |
multimodal |
|
|
| 10 |
TimeFound: A Foundation Model for Time Series Forecasting |
TimeFound:用于时间序列预测的Transformer基础模型,实现零样本预测。 |
foundation model |
|
|
| 11 |
Continual Pre-training of MoEs: How robust is your router? |
研究MoE模型持续预训练的鲁棒性,揭示路由算法对性能的影响 |
foundation model |
|
|
| 12 |
Universality of Layer-Level Entropy-Weighted Quantization Beyond Model Architecture and Size |
提出层级熵权重量化(EWQ),实现模型架构和尺寸无关的LLM选择性量化。 |
large language model |
|
|
| 13 |
CLDyB: Towards Dynamic Benchmarking for Continual Learning with Pre-trained Models |
提出CLDyB动态基准测试框架,解决持续学习中数据污染和基准饱和问题。 |
foundation model |
✅ |
|
| 14 |
Know Thy Judge: On the Robustness Meta-Evaluation of LLM Safety Judges |
评估LLM安全评判器的鲁棒性:揭示提示敏感性和对抗攻击下的脆弱性 |
large language model |
|
|
| 15 |
Speculative MoE: Communication Efficient Parallel MoE Inference with Speculative Token and Expert Pre-scheduling |
Speculative MoE:通过推测Token和专家预调度,提升MoE模型通信效率 |
large language model |
|
|
| 16 |
How to Mitigate Overfitting in Weak-to-strong Generalization? |
提出双阶段框架,提升弱监督到强泛化中的过拟合问题 |
large language model |
|
|
| 17 |
ThrowBench: Benchmarking LLMs by Predicting Runtime Exceptions |
提出ThrowBench基准测试,用于评估LLM预测运行时异常的能力 |
large language model |
|
|
| 18 |
PokéChamp: an Expert-level Minimax Language Agent |
PokéChamp:基于LLM的专家级Minimax宝可梦对战智能体 |
large language model |
|
|