| 1 |
ASER: Activation Smoothing and Error Reconstruction for Large Language Model Quantization |
ASER:通过激活平滑与误差重构实现大语言模型低比特量化 |
large language model |
|
|
| 2 |
Zer0-Jack: A Memory-efficient Gradient-based Jailbreaking Method for Black-box Multi-modal Large Language Models |
Zer0-Jack:一种面向黑盒多模态大语言模型的内存高效梯度越狱方法 |
large language model |
|
|
| 3 |
Retrieval Augmented Time Series Forecasting |
提出检索增强时间序列预测框架RAF,提升时间序列基础模型在多样化场景下的零样本预测精度。 |
foundation model |
|
|
| 4 |
NVCiM-PT: An NVCiM-assisted Prompt Tuning Framework for Edge LLMs |
提出NVCiM辅助的Prompt Tuning框架,解决边缘LLM领域迁移问题。 |
large language model |
|
|
| 5 |
FRUGAL: Memory-Efficient Optimization by Reducing State Overhead for Scalable Training |
FRUGAL:通过减少状态开销实现内存高效优化,用于可扩展训练 |
large language model |
|
|
| 6 |
Efficient Federated Finetuning of Tiny Transformers with Resource-Constrained Devices |
提出一种高效联邦微调Tiny Transformer的层微调方案,解决资源受限设备上的内存和计算瓶颈。 |
large language model |
|
|
| 7 |
Federated Low-Rank Adaptation with Differential Privacy over Wireless Networks |
提出基于无线网络差分隐私的联邦低秩自适应框架,解决边缘设备微调大模型时的计算和隐私挑战。 |
foundation model |
|
|
| 8 |
What Do Learning Dynamics Reveal About Generalization in LLM Reasoning? |
通过学习动态揭示LLM推理泛化能力:提出预记忆训练准确率指标 |
large language model |
|
|
| 9 |
Circuit Complexity Bounds for RoPE-based Transformer Architecture |
证明RoPE Transformer在特定复杂度类下的表达能力存在根本限制 |
large language model |
|
|
| 10 |
Model Stealing for Any Low-Rank Language Model |
针对低秩语言模型的模型窃取算法,提升了窃取效率和适用性 |
large language model |
|
|