| 1 |
Morality is Contextual: Learning Interpretable Moral Contexts from Human Data with Probabilistic Clustering and Large Language Models |
COMETH框架通过概率聚类和LLM学习可解释的道德上下文,提升道德判断准确性。 |
large language model |
|
|
| 2 |
Semi-Supervised Learning for Large Language Models Safety and Content Moderation |
提出半监督学习方法,提升大语言模型安全与内容审核能力 |
large language model |
|
|
| 3 |
Neural Probe-Based Hallucination Detection for Large Language Models |
提出基于神经探针的大语言模型幻觉检测框架,提升低误报下的检测精度。 |
large language model |
|
|
| 4 |
ClarifyMT-Bench: Benchmarking and Improving Multi-Turn Clarification for Conversational Large Language Models |
提出ClarifyMT-Bench,用于评测和提升会话大语言模型的多轮澄清能力。 |
large language model |
|
|
| 5 |
Foundation Model-based Evaluation of Neuropsychiatric Disorders: A Lifespan-Inclusive, Multi-Modal, and Multi-Lingual Study |
提出基于大模型的神经精神疾病评估框架FEND,实现多模态、多语言和全生命周期的诊断。 |
foundation model |
|
|
| 6 |
Evaluating Novelty in AI-Generated Research Plans Using Multi-Workflow LLM Pipelines |
利用多工作流LLM评估AI生成研究计划的新颖性 |
large language model multimodal |
|
|
| 7 |
Rethinking Supervised Fine-Tuning: Emphasizing Key Answer Tokens for Improved LLM Accuracy |
SFTKey:通过强化关键答案token优化LLM监督微调的准确性 |
large language model chain-of-thought |
|
|
| 8 |
Reflection Pretraining Enables Token-Level Self-Correction in Biological Sequence Models |
提出反射预训练,使生物序列模型具备token级自纠错能力 |
large language model chain-of-thought |
|
|
| 9 |
ReaSeq: Unleashing World Knowledge via Reasoning for Sequential Modeling |
ReaSeq:通过推理释放世界知识,用于序列建模,提升推荐系统性能。 |
large language model chain-of-thought |
|
|
| 10 |
Teaching People LLM's Errors and Getting it Right |
研究LLM错误模式教学,提升用户识别LLM失效场景的能力 |
large language model |
|
|
| 11 |
C2LLM Technical Report: A New Frontier in Code Retrieval via Adaptive Cross-Attention Pooling |
C2LLM:通过自适应跨注意力池化实现代码检索的新突破 |
large language model |
|
|
| 12 |
LLM_annotate: A Python package for annotating and analyzing fiction characters |
LLM_annotate:用于小说人物分析的Python工具包,提升标注和分析效率。 |
large language model |
|
|
| 13 |
Architectural Trade-offs in Small Language Models Under Compute Constraints |
研究计算约束下小型语言模型架构权衡,揭示不同架构和训练预算对性能的影响 |
large language model |
|
|