当前共找到 2 篇文献分享。
1.
王昊 (2022-09-01 14:34):
#paper doi:10.1109/ICCV48922.2021.00014 ZHOU X, LIU X, WANG C, 等. Learning with Noisy Labels via Sparse Regularization[C/OL]//2021 IEEE/CVF International Conference on Computer Vision (ICCV). 2021: 72-81. https://doi.org/10.1109/ICCV48922.2021.00014. 本文使用稀疏正则化的方法,将输出尽可能地往one-hot上引导,使得输出锐化(一个是1,其它都是0,相当于有很大的确信度就是那一个答案,其它的概率都很低), 具体使用使用Lp Norm方法来达成. 该方法属于噪声标签去噪的损失函数方法的paper。噪声标签去噪综述可参见: SONG H, KIM M, PARK D, 等. Learning From Noisy Labels With Deep Neural Networks: A Survey[J/OL]. IEEE Transactions on Neural Networks and Learning Systems, 2022: 1-19. https://doi.org/10.1109/TNNLS.2022.3152527
Abstract:
Learning with noisy labels is an important and challenging task for training accurate deep neural networks. Some commonly-used loss functions, such as Cross Entropy (CE), suffer from severe overfitting to … >>>
Learning with noisy labels is an important and challenging task for training accurate deep neural networks. Some commonly-used loss functions, such as Cross Entropy (CE), suffer from severe overfitting to noisy labels. Robust loss functions that satisfy the symmetric condition were tailored to remedy this problem, which however encounter the underfitting effect. In this paper, we theoretically prove that any loss can be made robust to noisy labels by restricting the network output to the set of permutations over a fixed vector. When the fixed vector is one-hot, we only need to constrain the output to be one-hot, which however produces zero gradients almost everywhere and thus makes gradient-based optimization difficult. In this work, we introduce the sparse regularization strategy to approximate the one-hot constraint, which is composed of network output sharpening operation that enforces the output distribution of a net-work to be sharp and the ℓ p -norm (p ≤ 1) regularization that promotes the network output to be sparse. This simple approach guarantees the robustness of arbitrary loss functions while not hindering the fitting ability. Experimental results demonstrate that our method can significantly improve the performance of commonly-used loss functions in the presence of noisy labels and class imbalance, and out-perform the state-of-the-art methods. The code is available at https://github.com/hitcszx/lnl_sr. <<<
翻译
2.
王昊 (2022-08-05 23:08):
#paper doi:10.1109/ICCV48922.2021.01307 [ZHANG F Z, CAMPBELL D, GOULD S. Spatially Conditioned Graphs for Detecting Human–Object Interactions[C/OL]//2021 IEEE/CVF International Conference on Computer Vision (ICCV). 2021: 13299-13307. https://doi.org/10.1109/ICCV48922.2021.01307. 本文使用GNN处理图像中人-物交互(HOI)的任务。在传统方法中,节点向它们的每个邻居发送与其它节点同质消息,本文根据它们的空间关系来调节节点对之间的消息传递内容,从而使得不同的消息发送到同一节点的邻居。其中用到了配对的二向图的概念和各向异性消息传递算法.多维度的数据的融合使用了MBF网络.本文是2021ICCV文章,在当年性能还行.可作为场景图生成(SGG)任务的子任务.
Abstract:
We address the problem of detecting human–object interactions in images using graphical neural networks. Unlike conventional methods, where nodes send scaled but otherwise identical messages to each of their neighbours, … >>>
We address the problem of detecting human–object interactions in images using graphical neural networks. Unlike conventional methods, where nodes send scaled but otherwise identical messages to each of their neighbours, we propose to condition messages between pairs of nodes on their spatial relationships, resulting in different messages going to neighbours of the same node. To this end, we explore various ways of applying spatial conditioning under a multi-branch structure. Through extensive experimentation we demonstrate the advantages of spatial conditioning for the computation of the adjacency structure, messages and the refined graph features. In particular, we empirically show that as the quality of the bounding boxes increases, their coarse appearance features contribute relatively less to the disambiguation of interactions compared to the spatial information. Our method achieves an mAP of 31.33% on HICO-DET and 54.2% on V-COCO, significantly outperforming state-of-the-art on fine-tuned detections. <<<
翻译
回到顶部