| 1 |
SEFE: Superficial and Essential Forgetting Eliminator for Multimodal Continual Instruction Tuning |
提出SEFE,通过消除表面和本质遗忘提升多模态持续指令调优性能 |
large language model multimodal |
|
|
| 2 |
Radio: Rate-Distortion Optimization for Large Language Model Compression |
提出基于率失真优化的Radio量化方法,用于大规模语言模型压缩。 |
large language model |
|
|
| 3 |
HSplitLoRA: A Heterogeneous Split Parameter-Efficient Fine-Tuning Framework for Large Language Models |
HSplitLoRA:异构拆分参数高效微调框架,用于大规模语言模型 |
large language model |
|
|
| 4 |
Knowledge Graphs for Enhancing Large Language Models in Entity Disambiguation |
利用知识图谱增强大语言模型在实体消歧中的表现 |
large language model |
|
|
| 5 |
A Note on Statistically Accurate Tabular Data Generation Using Large Language Models |
提出概率驱动的提示方法,利用大语言模型更准确地生成表格数据 |
large language model |
|
|
| 6 |
Enhancing Chemical Reaction and Retrosynthesis Prediction with Large Language Model and Dual-task Learning |
ChemDual:利用大语言模型和双任务学习提升化学反应和逆合成预测 |
large language model |
|
|
| 7 |
EntroLLM: Entropy Encoded Weight Compression for Efficient Large Language Model Inference on Edge Devices |
EntroLLM:面向边缘设备,通过熵编码压缩LLM权重以实现高效推理。 |
large language model |
|
|
| 8 |
Towards Cross-Modality Modeling for Time Series Analytics: A Survey in the LLM Era |
综述:面向时间序列分析的跨模态建模,聚焦LLM时代的方法与应用 |
large language model multimodal |
|
|
| 9 |
RetroInfer: A Vector-Storage Approach for Scalable Long-Context LLM Inference |
RetroInfer:一种向量存储方法,用于可扩展的长上下文LLM推理。 |
large language model |
|
|
| 10 |
Towards Quantifying the Hessian Structure of Neural Networks |
揭示神经网络Hessian矩阵近块对角结构的成因:架构与训练的双重作用 |
large language model |
|
|
| 11 |
When Your Own Output Becomes Your Training Data: Noise-to-Meaning Loops and a Formal RSI Trigger |
提出Noise-to-Meaning递归自提升模型(N2M-RSI),揭示AI自反馈学习中复杂性增长的机制。 |
large language model |
|
|
| 12 |
Less is More: Efficient Weight Farcasting with 1-Layer Neural Network |
提出基于单层神经网络的权重远距离预测方法,提升大模型训练效率 |
large language model |
|
|
| 13 |
Unlearning vs. Obfuscation: Are We Truly Removing Knowledge? |
区分遗忘与混淆:提出DF-MCQ方法,实现LLM的真实知识移除与拒绝行为。 |
large language model |
|
|
| 14 |
Rewriting Pre-Training Data Boosts LLM Performance in Math and Code |
重写预训练数据提升LLM在数学和代码领域的性能 |
large language model |
|
|