Credibility Governance: A Social Mechanism for Collective Self-Correction under Weak Truth Signals
作者: Wanying He, Yanxi Lin, Ziheng Zhou, Xue Feng, Min Peng, Qianqian Xie, Zilong Zheng, Yipeng Kang
分类: cs.CY, cs.AI, cs.CL, cs.MA, cs.SI
发布日期: 2026-03-03
🔗 代码/项目: GITHUB
💡 一句话要点
提出可信度治理机制以解决在线平台集体判断脆弱性问题
🎯 匹配领域: 支柱一:机器人控制 (Robot Control)
关键词: 可信度治理 集体判断 在线平台 动态评分 信息过滤
📋 核心要点
- 现有方法在面对弱真相信号和噪声反馈时,集体判断的脆弱性导致决策失误和资源浪费。
- 本文提出的可信度治理机制通过动态更新代理和观点的可信度评分,提升集体判断的可靠性。
- 实验结果显示,CG机制在多种复杂环境下表现优异,显著提高了恢复真实状态的速度和系统的鲁棒性。
📝 摘要(中文)
在线平台越来越依赖意见聚合来分配现实世界的关注和资源,但常见的信号如参与投票或资本加权承诺容易被放大,往往追踪可见性而非可靠性。这使得在弱真相信号、噪声反馈、早期流行激增和战略操控下,集体判断变得脆弱。本文提出可信度治理(CG)机制,通过学习哪些代理和观点能够持续跟踪不断演变的公共证据,重新分配影响力。CG为代理和观点维护动态可信度评分,通过可信度加权的背书更新观点影响力,并根据支持的观点的长期表现更新代理可信度,奖励与新兴证据的早期和持续一致性,同时过滤短期噪声。实验结果表明,CG在多种设置下优于基于投票、资本加权和无治理的基线,能够更快恢复真实状态,减少锁定和路径依赖,并在对抗压力下提高鲁棒性。
🔬 方法详解
问题定义:本文旨在解决在线平台在弱真相信号下集体判断脆弱的问题。现有方法容易受到噪声和操控的影响,导致决策失误和资源浪费。
核心思路:可信度治理机制通过动态维护代理和观点的可信度评分,学习哪些代理和观点能够持续跟踪公共证据,从而重新分配影响力,提升集体判断的可靠性。
技术框架:CG机制包括三个主要模块:动态可信度评分、可信度加权背书和代理可信度更新。动态评分根据代理和观点的表现进行调整,背书机制则通过可信度加权来更新观点影响力。
关键创新:CG机制的创新在于其动态调整的可信度评分和基于长期表现的代理可信度更新,与传统的静态投票或资本加权方法本质上不同,能够更有效地过滤短期噪声。
关键设计:在设计中,CG机制采用了动态评分算法,设置了特定的损失函数以优化代理的表现,同时考虑了噪声和信息污染对决策的影响。
🖼️ 关键图片
📊 实验亮点
实验结果表明,可信度治理机制在面对初始多数不一致、观察噪声和虚假信息冲击的情况下,显著优于基于投票和资本加权的基线,能够更快恢复到真实状态,减少锁定和路径依赖,提高系统在对抗压力下的鲁棒性。
🎯 应用场景
该研究的潜在应用领域包括社交媒体平台、在线评论系统和众包决策机制等。通过提升集体判断的可靠性,可信度治理机制能够帮助平台更有效地分配资源,减少错误决策的发生,具有重要的实际价值和未来影响。
📄 摘要(原文)
Online platforms increasingly rely on opinion aggregation to allocate real-world attention and resources, yet common signals such as engagement votes or capital-weighted commitments are easy to amplify and often track visibility rather than reliability. This makes collective judgments brittle under weak truth signals, noisy or delayed feedback, early popularity surges, and strategic manipulation. We propose Credibility Governance (CG), a mechanism that reallocates influence by learning which agents and viewpoints consistently track evolving public evidence. CG maintains dynamic credibility scores for both agents and opinions, updates opinion influence via credibility-weighted endorsements, and updates agent credibility based on the long-run performance of the opinions they support, rewarding early and persistent alignment with emerging evidence while filtering short-lived noise. We evaluate CG in POLIS, a socio-physical simulation environment that models coupled belief dynamics and downstream feedback under uncertainty. Across settings with initial majority misalignment, observation noise and contamination, and misinformation shocks, CG outperforms vote-based, stake-weighted, and no-governance baselines, yielding faster recovery to the true state, reduced lock-in and path dependence, and improved robustness under adversarial pressure. Our implementation and experimental scripts are publicly available at https://github.com/Wanying-He/Credibility_Governance.