| 1 |
Can Active Sampling Reduce Causal Confusion in Offline Reinforcement Learning? |
主动采样减少离线强化学习中的因果混淆 |
reinforcement learning offline reinforcement learning |
|
|
| 2 |
Generalizable Visual Reinforcement Learning with Segment Anything Model |
提出SAM-G框架,利用SAM提升视觉强化学习在未知环境中的泛化能力 |
reinforcement learning foundation model |
|
|
| 3 |
Beyond PID Controllers: PPO with Neuralized PID Policy for Proton Beam Intensity Control in Mu2e |
提出基于神经PID策略的PPO算法,用于Mu2e实验中的质子束强度控制 |
reinforcement learning PPO |
|
|
| 4 |
Rethinking Model-based, Policy-based, and Value-based Reinforcement Learning via the Lens of Representation Complexity |
提出表示复杂性层次以重构强化学习范式 |
reinforcement learning model-based RL |
|
|
| 5 |
RLPlanner: Reinforcement Learning based Floorplanning for Chiplets with Fast Thermal Analysis |
RLPlanner:基于强化学习的Chiplet Floorplanning,加速热分析 |
reinforcement learning MAE |
|
|
| 6 |
Think Before You Duel: Understanding Complexities of Preference Learning under Constrained Resources |
提出一种基于EXP3的对抗算法以解决资源约束下的偏好学习问题 |
preference learning |
|
|
| 7 |
Resilient Constrained Reinforcement Learning |
提出弹性约束强化学习,解决约束条件未知下的强化学习问题 |
reinforcement learning |
|
|
| 8 |
Improving Intrusion Detection with Domain-Invariant Representation Learning in Latent Space |
提出基于领域不变表征学习的入侵检测方法,提升零日攻击检测能力 |
representation learning |
|
|
| 9 |
FedSDD: Scalable and Diversity-enhanced Distillation for Model Aggregation in Federated Learning |
FedSDD:面向联邦学习的可扩展、多样性增强的蒸馏模型聚合方法 |
distillation |
|
|
| 10 |
Agnostic Interactive Imitation Learning: New Theory and Practical Algorithms |
提出Agnostic交互式模仿学习算法MFTPL-P与Bootstrap-Dagger,解决专家策略非策略类问题。 |
imitation learning |
|
|
| 11 |
Layer Attack Unlearning: Fast and Accurate Machine Unlearning via Layer Level Attack and Knowledge Distillation |
提出层攻击卸载学习,通过层级攻击和知识蒸馏实现快速精确的机器卸载学习。 |
distillation |
|
|