| 1 |
Diffusion-based supervised learning of generative models for efficient sampling of multimodal distributions |
提出基于扩散模型的监督学习方法,高效采样多峰分布,解决贝叶斯推断难题。 |
multimodal |
|
|
| 2 |
Efficient Split Federated Learning for Large Language Models over Communication Networks |
提出SflLLM框架,通过拆分联邦学习和LoRA高效微调边缘端大语言模型 |
large language model |
|
|
| 3 |
NoWag: A Unified Framework for Shape Preserving Compression of Large Language Models |
NoWag:一种统一的LLM压缩框架,保持模型结构并实现高效压缩 |
large language model |
✅ |
|
| 4 |
Deep Learning with Pretrained 'Internal World' Layers: A Gemma 3-Based Modular Architecture for Wildfire Prediction |
提出基于Gemma 3预训练“内部世界”层的模块化架构,用于野火预测。 |
multimodal |
|
|
| 5 |
Evaluating Temporal Plasticity in Foundation Time Series Models for Incremental Fine-tuning |
评估时间序列基础模型的时间可塑性,用于增量微调 |
foundation model |
|
|
| 6 |
SlimPipe: Memory-Thrifty and Efficient Pipeline Parallelism for Long-Context LLM Training |
SlimPipe:面向长文本LLM训练的内存高效流水线并行方法 |
large language model |
|
|
| 7 |
Less is More: Adaptive Coverage for Synthetic Training Data |
提出基于最大覆盖的自适应采样算法,提升合成数据训练分类器性能 |
large language model |
|
|