| 1 |
TEA: Trajectory Encoding Augmentation for Robust and Transferable Policies in Offline Reinforcement Learning |
提出轨迹编码增强TEA,提升离线强化学习策略在未知环境中的泛化性 |
reinforcement learning offline reinforcement learning |
|
|
| 2 |
Puzzle: Distillation-Based NAS for Inference-Optimized LLMs |
Puzzle:基于蒸馏的NAS优化LLM推理,实现单H100 GPU部署 |
distillation large language model |
|
|
| 3 |
Pre-Training Graph Contrastive Masked Autoencoders are Strong Distillers for EEG |
提出EEG-DisGCMAE,解决低密度脑电图数据下的知识迁移与蒸馏问题 |
masked autoencoder distillation |
✅ |
|
| 4 |
MSEMG: Surface Electromyography Denoising with a Mamba-based Efficient Network |
提出MSEMG:一种基于Mamba的高效网络,用于表面肌电信号降噪。 |
Mamba state space model |
|
|
| 5 |
LD-EnSF: Synergizing Latent Dynamics with Ensemble Score Filters for Fast Data Assimilation with Sparse Observations |
LD-EnSF:融合潜在动力学与集成评分滤波,加速稀疏观测下的数据同化 |
latent dynamics |
|
|
| 6 |
ICLERB: In-Context Learning Embedding and Reranker Benchmark |
提出ICLERB基准测试与RLRAIF算法,优化上下文学习的检索增强生成。 |
reinforcement learning large language model |
|
|