| 1 |
Semi-Supervised Learning for Large Language Models Safety and Content Moderation |
提出半监督学习方法,提升大语言模型安全性和内容审核能力 |
large language model |
|
|
| 2 |
Neural Probe-Based Hallucination Detection for Large Language Models |
提出基于神经探针的大语言模型幻觉检测框架,提升低误报率下的检测性能。 |
large language model |
|
|
| 3 |
ClarifyMT-Bench: Benchmarking and Improving Multi-Turn Clarification for Conversational Large Language Models |
提出ClarifyMT-Bench,用于评估和提升会话大语言模型的多轮澄清能力。 |
large language model |
|
|
| 4 |
Foundation Model-based Evaluation of Neuropsychiatric Disorders: A Lifespan-Inclusive, Multi-Modal, and Multi-Lingual Study |
提出FEND框架,用于基于多模态融合和预训练模型评估神经精神疾病。 |
foundation model |
|
|
| 5 |
Rethinking Supervised Fine-Tuning: Emphasizing Key Answer Tokens for Improved LLM Accuracy |
SFTKey:通过强化关键答案token,提升LLM监督微调的准确率 |
large language model chain-of-thought |
|
|
| 6 |
Reflection Pretraining Enables Token-Level Self-Correction in Biological Sequence Models |
提出反射预训练,使生物序列模型具备token级自纠错能力 |
large language model chain-of-thought |
|
|
| 7 |
ReaSeq: Unleashing World Knowledge via Reasoning for Sequential Modeling |
ReaSeq:通过推理释放世界知识,用于序列建模,提升淘宝推荐系统性能 |
large language model chain-of-thought |
|
|
| 8 |
Architectural Trade-offs in Small Language Models Under Compute Constraints |
研究计算约束下小型语言模型的架构权衡,揭示不同架构选择对性能的影响。 |
large language model |
|
|
| 9 |
C2LLM Technical Report: A New Frontier in Code Retrieval via Adaptive Cross-Attention Pooling |
C2LLM:通过自适应跨注意力池化实现代码检索的新突破 |
large language model |
|
|