| 1 |
IDProxy: Cold-Start CTR Prediction for Ads and Recommendation at Xiaohongshu with Multimodal LLMs |
IDProxy:利用多模态LLM解决小红书广告和推荐中冷启动CTR预测问题 |
large language model multimodal |
|
|
| 2 |
Orchestrating Multimodal DNN Workloads in Wireless Neural Processing |
提出O-WiN框架,通过通信-计算流水线加速无线神经处理中的多模态DNN推理。 |
multimodal |
|
|
| 3 |
CoVAE: correlated multimodal generative modeling |
提出CoVAE模型,通过捕捉模态间相关性,提升多模态生成建模的性能和不确定性量化。 |
multimodal |
|
|
| 4 |
Causal Circuit Tracing Reveals Distinct Computational Architectures in Single-Cell Foundation Models: Inhibitory Dominance, Biological Coherence, and Cross-Model Convergence |
提出因果回路追踪方法,揭示单细胞Foundation模型中独特的计算架构。 |
foundation model |
|
|
| 5 |
SafeSci: Safety Evaluation of Large Language Models in Science Domains and Beyond |
SafeSci:构建科学领域大语言模型安全评估与提升的综合框架 |
large language model |
|
|
| 6 |
Frontier Models Can Take Actions at Low Probabilities |
前沿模型能以极低概率执行特定动作,需警惕恶意利用 |
chain-of-thought |
|
|
| 7 |
Symbol-Equivariant Recurrent Reasoning Models |
提出符号等变循环推理模型,提升神经推理的泛化性和鲁棒性 |
large language model |
✅ |
|
| 8 |
Multi-Head Low-Rank Attention |
提出多头低秩注意力(MLRA),解决大模型长文本推理中KV缓存的张量并行瓶颈。 |
large language model |
✅ |
|
| 9 |
Adam Converges Without Any Modification On Update Rules |
证明Adam在适当超参数下收敛,揭示其收敛-发散相变现象 |
large language model |
|
|
| 10 |
Probabilistic Retrofitting of Learned Simulators |
通过概率追溯拟合,将预训练的确定性模拟器转化为概率模型,提升偏微分方程建模性能。 |
foundation model |
|
|
| 11 |
Probing Materials Knowledge in LLMs: From Latent Embeddings to Reliable Predictions |
评估LLM在材料科学中的知识:从潜在嵌入到可靠预测 |
large language model |
|
|
| 12 |
Modular Memory is the Key to Continual Learning Agents |
提出模块化记忆架构,融合In-Weight Learning和In-Context Learning,解决持续学习中的灾难性遗忘问题。 |
foundation model |
|
|
| 13 |
DeLo: Dual Decomposed Low-Rank Experts Collaboration for Continual Missing Modality Learning |
提出DeLo,通过双重分解低秩专家协作解决持续缺失模态学习中的模态干扰问题。 |
multimodal |
|
|
| 14 |
One Operator to Rule Them All? On Boundary-Indexed Operator Families in Neural PDE Solvers |
揭示神经PDE求解器局限性:学习边界条件索引算子族而非通用算子 |
foundation model |
|
|
| 15 |
Quasar: Quantized Self-Speculative Acceleration for Rapid Inference via Memory-Efficient Verification |
Quasar:通过量化自推测加速和内存高效验证,实现快速LLM推理。 |
large language model |
✅ |
|
| 16 |
3BASiL: An Algorithmic Framework for Sparse plus Low-Rank Compression of LLMs |
提出3BASiL-TM算法框架,用于大语言模型的稀疏加低秩分解压缩,提升性能。 |
large language model |
✅ |
|