当前位置:网站首页>Cvpr2022 𞓜 loss problem in weakly supervised multi label classification

Cvpr2022 𞓜 loss problem in weakly supervised multi label classification

2022-06-21 23:37:00 CV technical guide (official account)

Preface   This paper proposes a new weakly supervised multi label classification (WSML) Method , This method rejects or corrects large loss samples , To prevent the model from remembering noisy labels . Because there are no heavy and complex components , The proposed method sets labels in several parts ( Include Pascal VOC 2012、MS COCO、NUSWIDE、CUB and OpenImages V3 Data sets ) Superior to the most advanced before WSML Method . Various analyses also show that , The practical effect of this method is very good , It is proved that it is very important to deal with the loss correctly in the weakly supervised multi label classification .

Welcome to the official account CV Technical guide , Focus on computer vision technology summary 、 The latest technology tracking 、 Interpretation of classic papers 、CV Recruitment information .

The paper :Large Loss Matters in Weakly Supervised Multi-Label Classification

The paper :http://arxiv.org/pdf/2206.03740

Code :https://github.com/snucml/LargeLossMatters

background

Weakly supervised multi label classification (WSML) The task is to use part of each image to observe labels to learn multi label classification , Because of its huge labeling cost , Becoming more and more important .

at present , There are two simple ways to train a model using partial tags . One is to use only observed tags to train the model , And ignore the unobserved labels . The other is to assume that all unobserved labels are negative , And incorporate it into your training , Because in multi label settings , Most labels are negative .

But the second method has one limitation , That is, this assumption will generate some noise in the tag , Thus hindering model learning , Therefore, most of the previous work followed the first method , And try to use various technologies ( Such as bootstrap or regularization ) Explore clues to unobserved tags . However , These methods include extensive calculations or complex optimization of pipelines .

Based on the above ideas , The author assumes that , If the label noise can be properly handled , The second approach may be a good starting point , Because it has the advantage of incorporating many real negative labels into model training . therefore , The author looks at it from the perspective of noise label learning WSML problem .

as everyone knows , When training models with noise labels , The model first adapts to clean labels , Then start remembering noise labels . Although previous studies have shown that memory effect only exists in noisy multi category classification scenes , But the author found that , The same effect also exists in noisy multi label classification scenarios . Pictured 1 Shown , During training , From the clean label ( True negative sample ) The loss value of is reduced from the beginning , And from the noise tag ( False negative sample ) The loss of is reduced from the middle .

chart 1 WSML Memory effect in

Based on this discovery , The author has developed three different schemes , By rejecting or correcting large loss samples during training , To prevent false positive labels from being memorized into the multi label classification model .

 

contribution

1) It is proved by experiments for the first time , Memory effect occurs in the process of multi label classification with noise .

2) A new weakly supervised multi label classification scheme is proposed , This scheme explicitly utilizes the learning technology with noise labels .

3) The proposed method is light and simple , The most advanced classification performance is achieved on various partial label datasets .

 

Method

In this paper , The author puts forward a new WSML Method , The motivation is based on the idea of noise multiclass learning , It ignores the huge loss in the process of model training . The weight term is further introduced into the loss function λi:

The authors propose three ways to provide weights λi Different schemes , The schematic diagram is described in Figure 2 Shown .

chart 2 The overall pipeline of the proposed method

1. Loss rejection

One way to handle large loss samples is by setting λi=0 To reject it . In multi class tasks with noise ,B.Han Et al. Proposed a method to gradually increase the rejection rate in the training process . The author also sets the function λi,

Because the model learns clean patterns at the initial stage , So in t=1 Do not reject any loss value . Use small batches instead of full batches in each iteration D′ To form a loss set . The author calls this method LL-R.

2. Loss correction ( temporary )

Another way to deal with a large loss sample is to correct it rather than reject it . In multi label settings , This can be easily achieved by switching the corresponding annotation from a negative value to a positive value .“ temporary ” The word means , It does not change the actual label , Only the loss calculated according to the modified label is used , Will function λi Defined as

The author named this method LL-Ct. The advantage of this method is , It increases the number of true positive tags in the tags that have never been observed .

3. Loss correction ( permanent )

Deal more aggressively with larger loss values by permanently correcting labels . Directly change the label from negative to positive , And use the modified tag during the next training . So , Define... For each case λi=1, And modify the label as follows :

The author named this method LL-Cp.

 

experiment

surface 2 Quantitative results of artificially created partial label data sets

surface 3 OpenImages V3 Quantitative results in the dataset

chart 3 Artificially generated COCO Qualitative results of some label datasets

chart 4 COCO Accuracy analysis of the proposed method on the dataset

chart 5 LL-Ct Yes COCO The hyperparametric effect of data sets

chart 6 Training with fewer images

surface 4 Pointing Game

Conclusion

In this paper , The author puts forward a loss modification scheme , This scheme rejects or corrects the large loss samples when training multi label classification models with partial label annotations . This comes from empirical observation , That is, the memory effect also occurs in noisy multi label classification scenarios .

Although it does not include heavy and complex components , But the author's scheme successfully prevents the multi label classification model from remembering false negative labels with noise , State of the art performance on a variety of partially labeled multi label datasets .

---------------------------------------------------------------------------------

 

 

CV The technical guide creates a computer vision technology exchange group and a free version of the knowledge planet , At present, the number of people on the planet has 700+, The number of topics reached 200+.

The knowledge planet will release some homework every day , It is used to guide people to learn something , You can continue to punch in and learn according to your homework .CV Every day in the technology group, the top conference papers published in recent days will be sent , You can choose the papers you are interested in to read , continued follow Latest technology , If you write an interpretation after reading it and submit it to us , You can also receive royalties . in addition , The technical group and my circle of friends will also publish various periodicals 、 Notice of solicitation of contributions for the meeting , If you need it, please scan your friends , And pay attention to .

Add groups and planets : Official account CV Technical guide , Get and edit wechat , Invite to join .

Welcome to the official account CV Technical guide , Focus on computer vision technology summary 、 The latest technology tracking 、 Interpretation of classic papers 、CV Recruitment information .

Welcome to contact me if you can write the following ( Get the contact information after following the official account ).

  1. Interpretation of the latest summit . For example, the most recent CVPR2022 The paper .
  2. Systematic overview of all directions 、 Development and evolution of main models 、 Innovative ideas, advantages and disadvantages of each model 、 Code analysis, etc . Such as the summary of target detection : A summary of target detection from traditional methods to deep learning , It mainly includes traditional method detection 、RCNN series 、YOLO series 、anchor-free series 、 Summary of small target detection methods 、 Summary of small sample target detection methods 、 Summary of object detection methods in video 、 Summary of loss function used in target detection . Support writing while learning .
  3. TVM Introduction to practice tutorial
  4. MNN Introduction to practice tutorial
  5. OpenVINO Introduction to practice tutorial
  6. libtorch Introduction to practice tutorial
  7. Oneflow Introduction to practice tutorial
  8. Detectron Introduction to practice tutorial
  9. caffe Source code reading
  10. Deep learning from introduction to mastery ( Start with convolutional neural networks )
  11. If you have something you want to write and it's not mentioned above , You can contact me . Statement : Paid , Please contact us for details .

 

Other articles of official account

Introduction to computer vision

Summary of common words in computer vision papers

YOLO Series carding ( Four ) About YOLO Deployment of

YOLO Series carding ( 3、 ... and )YOLOv5

YOLO Series carding ( Two )YOLOv4

YOLO Series carding ( One )YOLOv1-YOLOv3

CVPR2022 | Based on egocentric data OCR assessment

CVPR 2022 | Using contrast regularization method to deal with noise labels

CVPR2022 | Loss problem in weakly supervised multi label classification

CVPR2022 | iFS-RCNN: An incremental small sample instance divider

CVPR2022 | Time 3D: End to end joint monocular 3D object detection and tracking for autonomous driving

CVPR2022 | A ConvNet for the 2020s & How to design neural network Summary

CVPR2022 | PanopticDepth: A unified framework for depth aware panoramic segmentation

CVPR2022 | Reexamine pooling : Your feeling field is not ideal

CVPR2022 | Unknown target detection module STUD: Learn about unknown targets in the video

CVPR2022 | Ranking based siamese Visual tracking

Build from scratch Pytorch Model tutorial ( 6、 ... and ) Write the training process and reasoning process

Build from scratch Pytorch Model tutorial ( 5、 ... and ) Write the training process -- Some basic configuration

Build from scratch Pytorch Model tutorial ( Four ) Write the training process -- Argument parsing

Build from scratch Pytorch Model tutorial ( 3、 ... and ) build Transformer The Internet

Build from scratch Pytorch Model tutorial ( Two ) Build network

Build from scratch Pytorch Model tutorial ( One ) data fetch

Some personal thinking habits and thought summary about learning a new technology or field quickly

原网站

版权声明
本文为[CV technical guide (official account)]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/172/202206212143591850.html