| 1 |
HiDe-LLaVA: Hierarchical Decoupling for Continual Instruction Tuning of Multimodal Large Language Model |
HiDe-LLaVA:提出分层解耦方法,用于多模态大语言模型持续指令调优 |
large language model multimodal |
✅ |
|
| 2 |
Levels of Analysis for Large Language Models |
借鉴认知科学分析框架,提升大型语言模型的可理解性 |
large language model |
|
|
| 3 |
LLM-Match: An Open-Sourced Patient Matching Model Based on Large Language Models and Retrieval-Augmented Generation |
LLM-Match:基于大语言模型和RAG的开源患者匹配模型,优于现有方案。 |
large language model |
|
|
| 4 |
TablePilot: Recommending Human-Preferred Tabular Data Analysis with Large Language Models |
TablePilot:利用大语言模型推荐更符合人类偏好的表格数据分析方案 |
large language model |
|
|
| 5 |
Code-Driven Inductive Synthesis: Enhancing Reasoning Abilities of Large Language Models with Sequences |
提出CodeSeq数据集,提升大语言模型在归纳推理中的代码生成能力 |
large language model |
|
|
| 6 |
A Multi-Stage Framework with Taxonomy-Guided Reasoning for Occupation Classification Using Large Language Models |
提出基于分类体系引导推理的多阶段框架,利用大语言模型进行职业分类 |
large language model |
|
|
| 7 |
HICD: Hallucination-Inducing via Attention Dispersion for Contrastive Decoding to Mitigate Hallucinations in Large Language Models |
提出HICD,通过注意力分散诱导幻觉,用于对比解码以缓解大语言模型中的幻觉问题 |
large language model |
|
|
| 8 |
Can Language Models Follow Multiple Turns of Entangled Instructions? |
MultiTurnInstruct:系统评估LLM在多轮纠缠指令下的执行能力,揭示模型在记忆、推理和冲突解决间的权衡。 |
large language model instruction following |
✅ |
|
| 9 |
ThinkPatterns-21k: A Systematic Study on the Impact of Thinking Patterns in LLMs |
系统性研究思维模式对大语言模型影响,提出ThinkPatterns-21k数据集。 |
large language model instruction following |
|
|
| 10 |
Mitigating KV Cache Competition to Enhance User Experience in LLM Inference |
CacheOPT通过缓解KV缓存竞争,提升LLM推理用户体验 |
large language model |
|
|
| 11 |
AccelGen: Heterogeneous SLO-Guaranteed High-Throughput LLM Inference Serving for Diverse Applications |
AccelGen:面向多样化应用,提供异构SLO保障的高吞吐量LLM推理服务系统 |
large language model |
|
|
| 12 |
CoDet-M4: Detecting Machine-Generated Code in Multi-Lingual, Multi-Generator and Multi-Domain Settings |
CoDet-M4:提出一种多语言、多生成器、多领域环境下的机器生成代码检测框架 |
large language model |
|
|
| 13 |
Pensez: Less Data, Better Reasoning -- Rethinking French LLM |
Pensez:通过少量高质量数据,提升法语LLM的推理能力 |
large language model |
|
|
| 14 |
MetaScale: Test-Time Scaling with Evolving Meta-Thoughts |
MetaScale:通过演进的元思考实现大语言模型测试时自适应缩放 |
large language model |
|
|
| 15 |
DLPO: Towards a Robust, Efficient, and Generalizable Prompt Optimization Framework from a Deep-Learning Perspective |
DLPO:从深度学习视角构建鲁棒、高效、泛化的提示优化框架 |
large language model |
✅ |
|
| 16 |
KVShare: An LLM Service System with Efficient and Effective Multi-Tenant KV Cache Reuse |
KVShare:通过高效多租户KV缓存复用提升LLM服务系统性能 |
large language model |
|
|
| 17 |
Valid Text-to-SQL Generation with Unification-based DeepStochLog |
提出基于Unification的DeepStochLog框架,保证Text-to-SQL生成结果的有效性。 |
large language model |
✅ |
|
| 18 |
Computation Mechanism Behind LLM Position Generalization |
揭示LLM位置泛化能力背后的计算机制,发现注意力logits的解耦现象 |
large language model |
|
|
| 19 |
A Survey on Transformer Context Extension: Approaches and Evaluation |
Transformer长文本扩展综述:方法与评估 |
large language model |
|
|
| 20 |
Improving Complex Reasoning with Dynamic Prompt Corruption: A soft prompt Optimization Approach |
提出动态Prompt腐蚀(DPC)方法,提升大语言模型在复杂推理任务中的Prompt Tuning效果。 |
large language model |
|
|
| 21 |
REPA: Russian Error Types Annotation for Evaluating Text Generation and Judgment Capabilities |
提出REPA数据集,评估LLM在俄语环境下的文本生成和判断能力,揭示俄语LLM Judge的性能差距。 |
large language model |
|
|
| 22 |
ClusComp: A Simple Paradigm for Model Compression and Efficient Finetuning |
ClusComp:一种用于模型压缩和高效微调的简单范式 |
large language model |
|
|
| 23 |
A Framework to Assess Multilingual Vulnerabilities of LLMs |
提出多语言LLM脆弱性评估框架,揭示低资源语言中的安全风险。 |
large language model |
|
|
| 24 |
nvBench 2.0: Resolving Ambiguity in Text-to-Visualization through Stepwise Reasoning |
提出nvBench 2.0基准与Step-Text2Vis模型,解决文本到可视化任务中歧义查询的难题。 |
large language model |
|
|