| 1 |
Multimodal RewardBench 2: Evaluating Omni Reward Models for Interleaved Text and Image |
提出Multimodal RewardBench 2,用于评估处理交错文本和图像的通用奖励模型。 |
large language model multimodal |
|
|
| 2 |
An Information-Theoretic Framework for Robust Large Language Model Editing |
提出IBKE,基于信息瓶颈理论实现鲁棒的大语言模型知识编辑 |
large language model |
|
|
| 3 |
DualGuard: Dual-stream Large Language Model Watermarking Defense against Paraphrase and Spoofing Attack |
提出DualGuard,一种可防御复述攻击和欺骗攻击的双流大语言模型水印算法 |
large language model |
|
|
| 4 |
Hearing to Translate: The Effectiveness of Speech Modality Integration into LLMs |
首个SpeechLLM综合评测:对比端到端与级联架构语音翻译性能 |
large language model foundation model |
|
|
| 5 |
From Facts to Conclusions : Integrating Deductive Reasoning in Retrieval-Augmented LLMs |
提出推理追踪增强的RAG框架,解决检索信息冲突和主观性问题。 |
large language model |
|
|
| 6 |
Refusal Steering: Fine-grained Control over LLM Refusal Behaviour for Sensitive Topics |
Refusal Steering:通过激活向量干预实现对LLM在敏感话题上拒绝行为的细粒度控制 |
large language model |
|
|
| 7 |
Sigma-Moe-Tiny Technical Report |
Sigma-MoE-Tiny:提出一种高稀疏MoE语言模型,解决专家负载均衡问题,实现高效扩展。 |
foundation model |
✅ |
|
| 8 |
From Essence to Defense: Adaptive Semantic-aware Watermarking for Embedding-as-a-Service Copyright Protection |
提出SemMark:一种自适应语义感知水印方法,用于保护Embedding-as-a-Service的版权 |
large language model |
|
|
| 9 |
Evaluating OpenAI GPT Models for Translation of Endangered Uralic Languages: A Comparison of Reasoning and Non-Reasoning Architectures |
评估OpenAI GPT模型在濒危乌拉尔语翻译中的表现,对比推理与非推理架构。 |
large language model |
|
|
| 10 |
LoPA: Scaling dLLM Inference via Lookahead Parallel Decoding |
LoPA:通过前瞻并行解码加速扩散大语言模型推理 |
large language model |
✅ |
|