| 1 |
Caterpillar of Thoughts: The Optimal Test-Time Algorithm for Large Language Models |
提出Caterpillar of Thoughts (CaT),优化大语言模型测试时计算,提升效率。 |
large language model chain-of-thought |
|
|
| 2 |
Assessing the Robustness of Climate Foundation Models under No-Analog Distribution Shifts |
评估气候基础模型在非相似分布偏移下的鲁棒性 |
foundation model |
|
|
| 3 |
Can Graph Foundation Models Generalize Over Architecture? |
提出自适应图算子混合框架,提升图基础模型在异构架构任务上的泛化能力 |
foundation model |
|
|
| 4 |
TreeTeaming: Autonomous Red-Teaming of Vision-Language Models via Hierarchical Strategy Exploration |
TreeTeaming:通过分层策略探索实现视觉-语言模型的自主红队测试 |
large language model multimodal |
|
|
| 5 |
Sparser, Faster, Lighter Transformer Language Models |
提出稀疏Transformer语言模型,提升推理和训练效率并降低资源消耗 |
large language model foundation model |
|
|
| 6 |
Post-Selection Distributional Model Evaluation |
提出PS-DME框架,用于模型预选后对KPI分布进行可靠评估,解决后选择偏差问题。 |
large language model |
|
|
| 7 |
Multitask-Informed Prior for In-Context Learning on Tabular Data: Application to Steel Property Prediction |
提出多任务学习先验的TabPFN,用于钢材性能预测的上下文学习。 |
foundation model |
|
|