cs.CL(2026-04-01)
📊 共 21 篇论文 | 🔗 2 篇有代码
🎯 兴趣领域导航
支柱九:具身大模型 (Embodied Foundation Models) (13 🔗1)
支柱二:RL算法与架构 (RL & Architecture) (7 🔗1)
支柱七:动作重定向 (Motion Retargeting) (1)
🔬 支柱九:具身大模型 (Embodied Foundation Models) (13 篇)
🔬 支柱二:RL算法与架构 (RL & Architecture) (7 篇)
| # | 题目 | 一句话要点 | 标签 | 🔗 | ⭐ |
|---|---|---|---|---|---|
| 14 | Brainstacks: Cross-Domain Cognitive Capabilities via Frozen MoE-LoRA Stacks for Continual LLM Learning | Brainstacks:基于冻结MoE-LoRA堆栈的跨领域认知能力持续LLM学习 | DPO large language model instruction following | ||
| 15 | Agentic Tool Use in Large Language Models | 综述性研究:大型语言模型中的Agentic工具使用方法与演进 | policy learning large language model | ||
| 16 | Embarrassingly Simple Self-Distillation Improves Code Generation | 提出简单自蒸馏方法SSD,无需外部资源提升代码生成能力 | reinforcement learning distillation large language model | ||
| 17 | LangMARL: Natural Language Multi-Agent Reinforcement Learning | LangMARL:提出基于自然语言的多智能体强化学习框架,解决LLM智能体在动态环境中协同策略演化难题。 | reinforcement learning large language model | ||
| 18 | TR-ICRL: Test-Time Rethinking for In-Context Reinforcement Learning | TR-ICRL:面向上下文强化学习的测试时重思考框架,提升推理和知识密集型任务性能 | reinforcement learning large language model | ✅ | |
| 19 | Agent Q-Mix: Selecting the Right Action for LLM Multi-Agent Systems through Reinforcement Learning | Agent Q-Mix:通过强化学习为LLM多智能体系统选择最优动作 | reinforcement learning large language model | ||
| 20 | Optimsyn: Influence-Guided Rubrics Optimization for Synthetic Data Generation | Optimsyn:利用影响引导的规则优化合成数据生成,提升下游任务性能 | reinforcement learning large language model |
🔬 支柱七:动作重定向 (Motion Retargeting) (1 篇)
| # | 题目 | 一句话要点 | 标签 | 🔗 | ⭐ |
|---|---|---|---|---|---|
| 21 | Emotion Entanglement and Bayesian Inference for Multi-Dimensional Emotion Understanding | 提出EmoScene基准,并结合情感纠缠与贝叶斯推理提升多维度情感理解。 | motion prediction large language model |