| 1 |
Dynamic Multimodal Fusion via Meta-Learning Towards Micro-Video Recommendation |
提出基于元学习的动态多模态融合框架MetaMMF,用于提升微视频推荐效果。 |
representation learning multimodal |
✅ |
|
| 2 |
MSV-Mamba: A Multiscale Vision Mamba Network for Echocardiography Segmentation |
提出MSV-Mamba,用于提升超声心动图分割精度,尤其针对复杂结构。 |
Mamba spatiotemporal |
|
|
| 3 |
Dataset Distillation via Committee Voting |
提出基于委员会投票的数据集蒸馏方法CV-DD,提升小数据集泛化能力。 |
distillation |
✅ |
|
| 4 |
Localization-Aware Multi-Scale Representation Learning for Repetitive Action Counting |
提出LMRL框架,通过定位感知多尺度表示学习提升重复动作计数精度。 |
representation learning |
|
|
| 5 |
Skip Mamba Diffusion for Monocular 3D Semantic Scene Completion |
提出Skip Mamba扩散模型,用于单目3D语义场景补全,显著提升性能。 |
Mamba |
✅ |
|
| 6 |
EdgeTAM: On-Device Track Anything Model |
提出EdgeTAM,通过2D空间感知器加速SAM 2,实现移动端视频分割。 |
distillation foundation model |
|
|
| 7 |
SAMKD: Spatial-aware Adaptive Masking Knowledge Distillation for Object Detection |
提出空间感知自适应掩码知识蒸馏(SAMKD)框架,提升目标检测性能。 |
distillation |
|
|
| 8 |
Representation Learning of Point Cloud Upsampling in Global and Local Inputs |
ReLPU:通过全局-局部特征学习提升点云上采样性能 |
representation learning |
|
|
| 9 |
Rethinking Knowledge in Distillation: An In-context Sample Retrieval Perspective |
提出IC-KD框架,通过上下文样本检索视角重新定义知识蒸馏,提升模型性能。 |
distillation |
|
|
| 10 |
CSTA: Spatial-Temporal Causal Adaptive Learning for Exemplar-Free Video Class-Incremental Learning |
提出CSTA框架,通过时空因果自适应学习解决免样本视频类增量学习问题 |
distillation spatiotemporal |
|
|