cs.LG(2024-12-20)

📊 共 20 篇论文 | 🔗 2 篇有代码

🎯 兴趣领域导航

支柱二:RL算法与架构 (RL & Architecture) (10) 支柱九:具身大模型 (Embodied Foundation Models) (10 🔗2)

🔬 支柱二:RL算法与架构 (RL & Architecture) (10 篇)

#题目一句话要点标签🔗
1 Beyond Human Data: Aligning Multimodal Large Language Models by Iterative Self-Evolution 提出迭代自进化框架,无需人工标注对齐多模态大语言模型 DPO direct preference optimization large language model
2 Offline Reinforcement Learning for LLM Multi-Step Reasoning 提出OREO:一种用于LLM多步推理的离线强化学习方法 reinforcement learning offline RL offline reinforcement learning
3 Mamba-based Deep Learning Approach for Sleep Staging on a Wireless Multimodal Wearable System without Electroencephalography 提出基于Mamba的深度学习方法,利用可穿戴设备多模态数据实现无脑电睡眠分期 Mamba multimodal
4 SGAC: A Graph Neural Network Framework for Imbalanced and Structure-Aware AMP Classification SGAC:用于不平衡和结构感知AMP分类的图神经网络框架 representation learning contrastive learning distillation
5 FedRLHF: A Convergence-Guaranteed Federated Framework for Privacy-Preserving and Personalized RLHF 提出FedRLHF:一种保证收敛的联邦RLHF框架,用于保护隐私和实现个性化。 reinforcement learning policy learning RLHF
6 SORREL: Suboptimal-Demonstration-Guided Reinforcement Learning for Learning to Branch 提出SORREL以解决MILP求解中的分支学习问题 reinforcement learning offline reinforcement learning imitation learning
7 Novelty-Guided Data Reuse for Efficient and Diversified Multi-Agent Reinforcement Learning 提出基于新颖性引导的数据重用方法,提升多智能体强化学习效率与多样性 reinforcement learning distillation
8 Decoding fairness: a reinforcement learning perspective 基于强化学习在最后通牒博弈中解码公平性行为 reinforcement learning imitation learning
9 Multi Agent Reinforcement Learning for Sequential Satellite Assignment Problems 提出基于多智能体强化学习的卫星序列分配方法,显著提升大规模任务分配效率。 reinforcement learning
10 Graph Structure Refinement with Energy-based Contrastive Learning 提出基于能量的对比学习图结构优化框架ECL-GSR,提升图神经网络节点分类性能。 contrastive learning

🔬 支柱九:具身大模型 (Embodied Foundation Models) (10 篇)

#题目一句话要点标签🔗
11 Towards scientific discovery with dictionary learning: Extracting biological concepts from microscopy foundation models 结合稀疏字典学习与PCA白化的ICFL方法,从细胞显微图像中提取生物学概念。 large language model foundation model
12 Continual Learning Using a Kernel-Based Method Over Foundation Models 提出基于核方法的KLDA持续学习算法,有效应对灾难性遗忘和类间分离问题。 foundation model
13 A Machine Learning Approach for Emergency Detection in Medical Scenarios Using Large Language Models 利用大语言模型和提示工程实现医疗场景下的紧急情况自动检测 large language model
14 Statistical Modeling of Univariate Multimodal Data 提出一种基于密度谷点递归分割的单变量多模态数据统计建模方法 multimodal
15 Measuring Cross-Modal Interactions in Multimodal Models 提出InterSHAP,用于量化多模态模型中的跨模态交互,并应用于医疗领域。 multimodal
16 Extracting Interpretable Task-Specific Circuits from Large Language Models for Faster Inference 提出一种从大语言模型中提取任务特定电路的方法,加速推理。 large language model
17 Deciphering the Underserved: Benchmarking LLM OCR for Low-Resource Scripts 评估LLM在低资源文字OCR中的潜力,揭示零样本学习的局限性 large language model
18 WebLLM: A High-Performance In-Browser LLM Inference Engine WebLLM:一个高性能的浏览器内LLM推理引擎 large language model
19 Inference Scaling vs Reasoning: An Empirical Analysis of Compute-Optimal LLM Problem-Solving 探究推理能力与计算效率的权衡:LLM问题求解的实证分析 large language model
20 PreNeT: Leveraging Computational Features to Predict Deep Neural Network Training Time PreNeT:利用计算特征预测深度神经网络训练时间,优化硬件配置。 large language model

⬅️ 返回 cs.LG 首页 · 🏠 返回主页