| 1 |
Internalized Self-Correction for Large Language Models |
提出InSeC:一种基于内部自纠正的大语言模型训练方法 |
large language model instruction following |
|
|
| 2 |
Assessing Social Alignment: Do Personality-Prompted Large Language Models Behave Like Humans? |
评估社会一致性:人格提示的大语言模型行为是否与人类相似? |
large language model |
|
|
| 3 |
Beyond Partisan Leaning: A Comparative Analysis of Political Bias in Large Language Models |
提出基于主题的无角色扮演方法,评估大型语言模型在政治倾向和参与度上的偏见 |
large language model |
|
|
| 4 |
Do Multimodal Language Models Really Understand Direction? A Benchmark for Compass Direction Reasoning |
提出CDR基准测试,评估多模态语言模型在指南针方向推理上的能力。 |
multimodal |
|
|
| 5 |
Privacy in Fine-tuning Large Language Models: Attacks, Defenses, and Future Directions |
综述:针对微调大语言模型的隐私攻击、防御及未来方向 |
large language model |
|
|
| 6 |
Enhancing Conflict Resolution in Language Models via Abstract Argumentation |
利用抽象论证增强语言模型在冲突解决中的能力 |
large language model chain-of-thought |
|
|
| 7 |
Text2midi: Generating Symbolic Music from Captions |
提出Text2midi,一种端到端模型,用于从文本描述生成MIDI音乐文件。 |
large language model multimodal |
✅ |
|
| 8 |
The Task Shield: Enforcing Task Alignment to Defend Against Indirect Prompt Injection in LLM Agents |
Task Shield:通过任务对齐防御LLM Agent中的间接Prompt注入攻击 |
large language model |
|
|
| 9 |
LearnLM: Improving Gemini for Learning |
LearnLM:通过教学指令跟随改进Gemini模型,提升教育场景下的AI表现 |
instruction following |
|
|
| 10 |
TimeRAG: BOOSTING LLM Time Series Forecasting via Retrieval-Augmented Generation |
TimeRAG:通过检索增强生成提升LLM时间序列预测能力 |
large language model |
|
|
| 11 |
AIGCodeSet: A New Annotated Dataset for AI Generated Code Detection |
提出AIGCodeSet数据集,用于AI生成代码检测,并验证贝叶斯分类器性能。 |
large language model |
|
|