We use cookies to improve your experience with our site.

基于负类原型对比的偏标签学习算法

NPC: Negative Prototypical Contrasting for Label Disambiguation of Partial Label Learning

  • 摘要:
    研究背景 在现代机器学习领域,偏标签学习(Partial Label Learning, PLL)是弱监督学习里一个重要的研究方向。PLL处理的是每个训练样本被标注为一组候选标签的情况,而其中只有一个标签是真实的。这种标签模糊性在现实世界的数据标注中很常见,例如网络挖掘、疾病诊断和自动图像标注等应用场景。然而,现有的基于深度学习的PLL方法往往过于强调由不准确的伪标签引出的正样本,而忽视了负样本在学习足够可区分的表示空间中的重要性。
    目的 本研究的目的是解决现有PLL方法在表示空间学习中类内紧凑性和类间可分性之间平衡不足的问题。具体而言,研究旨在通过利用非候选标签集中提取的纯负监督信息来增强表示学习的质量,从而实现更优的标签消歧效果。这种方法的提出是为了克服现有方法在处理标签模糊性时的局限性,提高PLL的性能。
    方法 为了实现上述目的,我们提出了一种新的框架,Negative Prototypical Contrasting (NPC)。NPC对比每个样本与其候选原型和负原型,其优化目标是减小样本与候选原型的距离,增大样本与负原型的距离,实现一个足够可区分的表示空间。基于学习到的表示,标签消歧过程以移动平均的方式进行。理论上,NPC的目标等价于求解一个带约束的最大似然优化问题。
    结果 实验结果表明,NPC方法在多个PLL基准数据集上实现了最先进的分类性能,甚至可与完全监督的方法相比拟。具体来说,在受控的合成数据集及真实世界数据集上,NPC方法均优于所有基线方法,性能提升显著。此外,NPC方法在处理不同版本的数据增强时具有独立性,这使得它在实际应用中更加实用和高效。
    结论 NPC框架通过利用负监督信息显著提高了从部分标记数据集中学习的效果。该方法不仅有理论上的合理解释,在实验中也证明了其有效性,实际应用潜力巨大。NPC方法在收敛速度上仍有提升空间,未来将探索使用方差减少技术来进一步加速所提出的在线EM (Expectation-Maximization)类型算法的收敛速度。

     

    Abstract: Partial label learning (PLL) learns under label ambiguity where each training instance is annotated with a set of candidate labels, among which only one is the ground-truth label. Recent advances showed that PLL can be promoted by combining label disambiguation with representation learning coherently, which achieved state-of-the-art performance. However, most of the existing deep PLL methods over-emphasize pulling the inaccurate pseudo-label-induced positive samples and fail to achieve a balance between the intra-class compactness and the inter-class separability, thus leading to a sub-optimal representation space. In this paper, we solve this issue by taking into account the pure negative supervision information which can be extracted perfectly from the non-candidate label set. Methodologically, we propose a novel framework Negative Prototypical Contrasting (NPC). The optimization objective of NPC contrasts each instance with its candidate prototypes against its negative prototypes, aiming at a sufficiently distinguishable representation space. Based on the learned representations, the label disambiguation process is performed in a moving-average style. Theoretically, we show that the objective of NPC is equivalent to solving a constrained maximum likelihood optimization. We also justify applying the moving average from the stochastic expectation-maximization perspective. Empirically, extensive experiments demonstrate that the proposed NPC method achieves state-of-the-art classification performance on various datasets, and even competes with its supervised counterparts.

     

/

返回文章
返回