| 1 |
ChemEval: A Comprehensive Multi-Level Chemical Evaluation for Large Language Models |
提出ChemEval以解决化学领域LLM评估不足问题 |
large language model instruction following |
✅ |
|
| 2 |
Instruction Following without Instruction Tuning |
揭示隐式指令调优:仅凭响应或领域数据微调即可实现指令遵循 |
instruction following |
|
|
| 3 |
Interpreting Arithmetic Mechanism in Large Language Models through Comparative Neuron Analysis |
通过比较神经元分析解读大语言模型中的算术机制 |
large language model |
✅ |
|
| 4 |
Role-Play Paradox in Large Language Models: Reasoning Performance Gains and Ethical Dilemmas |
大型语言模型角色扮演悖论:推理性能提升与伦理困境 |
large language model |
|
|
| 5 |
Exploring Automated Keyword Mnemonics Generation with Large Language Models via Overgenerate-and-Rank |
提出基于大语言模型的Overgenerate-and-Rank方法,自动生成关键词助记法,辅助词汇学习。 |
large language model |
|
|
| 6 |
GroupDebate: Enhancing the Efficiency of Multi-Agent Debate Using Group Discussion |
GroupDebate:利用分组讨论提升多智能体辩论效率,降低token成本 |
large language model chain-of-thought |
|
|
| 7 |
Obliviate: Neutralizing Task-agnostic Backdoors within the Parameter-efficient Fine-tuning Paradigm |
Obliviate:中和参数高效微调中与任务无关的后门攻击 |
large language model |
✅ |
|
| 8 |
PTD-SQL: Partitioning and Targeted Drilling with LLMs in Text-to-SQL |
PTD-SQL:利用LLM进行文本到SQL的查询分组划分与针对性训练,提升模型推理能力。 |
large language model |
✅ |
|
| 9 |
Rephrase and Contrast: Fine-Tuning Language Models for Enhanced Understanding of Communication and Computer Networks |
提出RaC框架,通过重述和对比微调语言模型,提升通信网络理解能力 |
large language model |
|
|
| 10 |
Probing Context Localization of Polysemous Words in Pre-trained Language Model Sub-Layers |
探究预训练语言模型子层中多义词上下文定位能力 |
large language model |
|
|