| 1 |
HackerRank-ASTRA: Evaluating Correctness & Consistency of Large Language Models on cross-domain multi-file project problems |
HackerRank-ASTRA:评估大语言模型在跨领域多文件项目中的正确性和一致性 |
large language model |
|
|
| 2 |
Should You Use Your Large Language Model to Explore or Exploit? |
评估LLM在探索-利用权衡中的能力,发现其在探索语义化动作空间中具有潜力 |
large language model |
|
|
| 3 |
Cache Me If You Must: Adaptive Key-Value Quantization for Large Language Models |
AQUA-KV:自适应键值量化,提升大语言模型KV缓存压缩率并保持精度 |
large language model |
|
|
| 4 |
Pivoting Factorization: A Compact Meta Low-Rank Representation of Sparsity for Efficient Inference in Large Language Models |
提出PIFA:一种紧凑的低秩表示方法,用于加速大语言模型推理。 |
large language model |
✅ |
|
| 5 |
Towards the Worst-case Robustness of Large Language Models |
提出针对大语言模型的最坏情况鲁棒性评估方法 |
large language model |
|
|
| 6 |
Symmetric Pruning of Large Language Models |
提出对称剪枝理论,并结合激活与权重重要性,显著提升大语言模型剪枝效果。 |
large language model |
|
|
| 7 |
Improving Rule-based Reasoning in LLMs using Neurosymbolic Representations |
提出神经符号表示方法,提升LLM在数学推理任务中的规则遵循能力 |
large language model chain-of-thought |
✅ |
|
| 8 |
Spend Wisely: Maximizing Post-Training Gains in Iterative Synthetic Data Bootstrapping |
优化迭代合成数据自举训练:最大化后训练阶段性能提升 |
large language model foundation model |
|
|
| 9 |
Byzantine-Resilient Zero-Order Optimization for Communication-Efficient Heterogeneous Federated Learning |
提出CyBeR-0以解决异构联邦学习中的拜占庭攻击问题 |
large language model |
|
|
| 10 |
Federated Sketching LoRA: A Flexible Framework for Heterogeneous Collaborative Fine-Tuning of LLMs |
提出联邦草图LoRA,解决异构联邦环境下LLM高效协同微调问题 |
large language model |
|
|
| 11 |
Judge Decoding: Faster Speculative Sampling Requires Going Beyond Model Alignment |
提出Judge Decoding,通过训练判别模块显著加速LLM推断,突破模型对齐限制。 |
large language model |
|
|
| 12 |
Offline Learning for Combinatorial Multi-armed Bandits |
提出Off-CMAB框架,解决组合多臂老虎机离线学习问题 |
large language model |
|
|
| 13 |
Elucidating Subspace Perturbation in Zeroth-Order Optimization: Theory and Practice at Scale |
提出MeZO-BCD,加速零阶优化在大语言模型微调中的应用,提升高达2.77倍。 |
large language model |
|
|
| 14 |
TeZO: Empowering the Low-Rankness on the Temporal Dimension in the Zeroth-Order Optimization for Fine-tuning LLMs |
TeZO:利用时序维度低秩性,提升大语言模型零阶优化微调效率 |
large language model |
|
|
| 15 |
Importing Phantoms: Measuring LLM Package Hallucination Vulnerabilities |
评估LLM代码生成中软件包幻觉漏洞,揭示潜在软件供应链攻击风险 |
large language model |
|
|
| 16 |
TabFSBench: Tabular Benchmark for Feature Shifts in Open Environments |
提出TabFSBench,首个表格数据特征偏移基准,评估模型在开放环境下的泛化能力。 |
large language model |
✅ |
|
| 17 |
LLM Program Optimization via Retrieval Augmented Search |
提出检索增强搜索(RAS)优化LLM程序,并用AEGIS提升可解释性。 |
large language model |
|
|
| 18 |
Scaling Laws for Differentially Private Language Models |
提出差分隐私语言模型的缩放规律以优化训练配置 |
large language model |
|
|
| 19 |
Predictive Prompt Analysis |
提出基于稀疏自编码器的预测性提示分析方法SPA,加速LLM提示工程。 |
large language model |
|
|
| 20 |
Partially Rewriting a Transformer in Natural Language |
提出一种部分重写Transformer的方法,旨在用自然语言解释替换网络组件,提升模型可解释性。 |
large language model |
|
|