cs.LG(2025-12-23)

📊 共 14 篇论文 | 🔗 2 篇有代码

🎯 兴趣领域导航

支柱二:RL算法与架构 (RL & Architecture) (8 🔗1) 支柱九:具身大模型 (Embodied Foundation Models) (5 🔗1) 支柱一:机器人控制 (Robot Control) (1)

🔬 支柱二:RL算法与架构 (RL & Architecture) (8 篇)

#题目一句话要点标签🔗
1 QE-Catalytic: A Graph-Language Multimodal Base Model for Relaxed-Energy Prediction in Catalytic Adsorption 提出QE-Catalytic,融合图和语言模型,提升催化吸附中弛豫能量预测精度。 MAE large language model multimodal
2 Sample-Efficient Policy Constraint Offline Deep Reinforcement Learning based on Sample Filtering 提出基于样本过滤的策略约束离线深度强化学习方法,提升样本效率。 reinforcement learning deep reinforcement learning offline RL
3 Recurrent Off-Policy Deep Reinforcement Learning Doesn't Have to be Slow 提出RISE,通过简化编码提升图像Off-Policy强化学习中循环网络的效率 reinforcement learning deep reinforcement learning
4 TableGPT-R1: Advancing Tabular Reasoning Through Reinforcement Learning TableGPT-R1:通过强化学习提升表格推理能力,实现SOTA性能。 reinforcement learning reward shaping large language model
5 Emergent temporal abstractions in autoregressive models enable hierarchical reinforcement learning 提出内部强化学习,利用自回归模型中的时间抽象实现分层强化学习 reinforcement learning foundation model
6 Performative Policy Gradient: Optimality in Performative Reinforcement Learning 提出PePG算法,解决强化学习中策略执行带来的环境动态变化问题,实现策略的执行最优性。 reinforcement learning
7 Jensen-Shannon Divergence Message-Passing for Rich-Text Graph Representation Learning 提出JSDMP框架,利用Jensen-Shannon散度提升富文本图表示学习 representation learning
8 Generalisation in Multitask Fitted Q-Iteration and Offline Q-learning 提出多任务离线Q学习方法以提升统计效率与泛化能力 reinforcement learning offline reinforcement learning

🔬 支柱九:具身大模型 (Embodied Foundation Models) (5 篇)

#题目一句话要点标签🔗
9 Unified Multimodal Brain Decoding via Cross-Subject Soft-ROI Fusion 提出BrainROI模型,通过跨被试软ROI融合实现统一的多模态脑解码,提升脑活动到自然语言描述的泛化性和可解释性。 multimodal
10 LoFT-LLM: Low-Frequency Time-Series Forecasting with Large Language Models LoFT-LLM:结合低频学习与大语言模型的时间序列预测框架 large language model
11 Fail Fast, Win Big: Rethinking the Drafting Strategy in Speculative Decoding via Diffusion LLMs FailFast:利用扩散LLM加速推测解码,显著提升自回归LLM推理速度 large language model
12 Learning to Reason in LLMs by Expectation Maximization 提出基于期望最大化的LLM推理学习框架,提升复杂推理任务性能 large language model
13 Reliable LLM-Based Edge-Cloud-Expert Cascades for Telecom Knowledge Systems 提出基于级联LLM的电信知识系统,优化成本与可靠性,实现自动化决策。 large language model

🔬 支柱一:机器人控制 (Robot Control) (1 篇)

#题目一句话要点标签🔗
14 Explainable time-series forecasting with sampling-free SHAP for Transformers 提出SHAPformer,一种基于Transformer的快速、无采样的可解释时间序列预测模型。 manipulation

⬅️ 返回 cs.LG 首页 · 🏠 返回主页