当前位置:网站首页>The latest research progress of domain generalization from CVPR 2022
The latest research progress of domain generalization from CVPR 2022
2022-06-23 22:03:00 【Zhiyuan community】
Domain Adaptation(DA: Domain adaption ),Domain Generalization(DG: Domain generalization ) It has always been a hot research direction of major summit meetings .DA Suppose we have a labeled training set ( Source domain ), At this time, we want the model to perform well on another data set ( Target domain ), Use the unlabeled data of the target domain , Improving the adaptability of the model between domains is DA What is emphasized . On this basis ,DG Further weakening the hypothesis , We only have data from multiple source domains , I have no idea what the target domain is , How to improve model generalization at this time ? The core is how to use the rich information brought by multiple source domains . Four articles are selected in this paper CVPR 2022 Domain generalization related articles to study the latest progress .

BIRM

Paper title :Bayesian Invariant Risk Minimization
Thesis link :
https://openaccess.thecvf.com/content/CVPR2022/papers/Lin_Bayesian_Invariant_Risk_Minimization_CVPR_2022_paper.pdf
1.1 Motivation
Generalization under distribution shift is an open challenge of machine learning . Constant risk minimization (IRM) This problem is solved by extracting invariant features . although IRM It has a complete theoretical system , But the performance in depth model is not very good , This paper holds that this failure is mainly The depth model is easy to over fit Caused by the , It is theoretically verified that when the model is over fitted ,IRM Degenerate into traditional ERM. This article will Bayesian inference introduce IRM Put forward Bayesian Invariant Risk Min-imization(BIRM) To alleviate this problem to a certain extent and achieved good results .
1.2 Background
Here is a brief introduction OOD Problems and IRM, So-called OOD The question can be written in the following form

That is to find the best classifier ,encoder Parameters Make the model have better performance in the worst domain ,i.e., . there Is domain Negative log likelihood of data in :
Invariant Risk Minimization(IRM). IRM To solve the following problems :

That is, he wants to learn a encoder Parameters , This encoder For all classifier parameters Both are optimal at the same time . To achieve this goal ,encoder Need to get rid of spurious feature. But this optimization form bi-level Of , Very difficult to solve , So he put forward an approximate target.

DARLING

Paper title :Towards Unsupervised Domain Generalization
Thesis link :
https://arxiv.org/abs/2107.06219
2.1 Motivation
current DG Most algorithms are pre trained on large data sets backbone, Then the algorithm is developed on the downstream data set finetune, Last in unseen Test on the data set of . But the pre - training data set introduces the idea that bias, such as imagenet Most of the images come from the real world , Therefore, its pre trained model is not real data on downstream tasks ( Hand drawn image , Cartoon image ) Performance will be poor . This paper considers a new setting, The model First, perform unsupervised pre training on multiple source domains , And then traditional DG The process of , Erase the tradition pretrain Introduced bias.
2.2 Method
The method name in this article is Domain-Aware Representation Learn-ING(DARLING), This method tries to use self supervised learning technology to improve the multi domain pre training method, so as to improve the generalization performance . The traditional self supervised learning loss is :

there It is in self-monitoring anchor and positive features , and It's two separate encoder, It's a temperature parameter . But this method can not build the confidence of the model domain , because It is actually different in different domains . Every domain The conditional distribution of can be written :

The model architecture is as follows , there It's below similarity predictor Output . That is to say, it has become each domain Compare the weighted form of learning loss :

PCL

Paper title :PCL: Proxy-based Contrastive Learning for Domain Generalization
Thesis link :
https://openaccess.thecvf.com/content/CVPR2022/papers/Yao_PCL_Proxy-Based_Contrastive_Learning_for_Domain_Generalization_CVPR_2022_paper.pdf
3.1 Motivation
This article is also related to contrastive learning A related article , A simple method is to pull the positive sample pairs of different fields closer , At the same time, push other negative sample pairs further . This paper finds that It is not good to directly adopt this kind of supervised comparative learning , This paper considers that there is a significant distribution gap between domains , It makes the distance between positive sample pairs close directly, but hinders the generalization of the model . Therefore, this paper proposes a new prototype based approach (proxy) Comparative learning method .
3.2 Method: Proxy-based Contrastive Learning
Let's look at it first , What is the difference between prototype based self supervised learning and traditional self supervised learning . As shown in the figure below :

PACS Dataset is a typical domain generalization benchmark , It includes four areas : art 、 comic 、 Photos and sketches , There are seven categories in each field .DG From multiple source domains ( Like art 、 Photo 、 Sketch ) Training models , And in the target area ( Like comics ) Test on . In the training phase , The target data set cannot be accessed .
Traditional comparison based losses ( for example , Monitor and compare losses ) Using its sample to sample relationship , Among them, samples from different domains of the same class can be regarded as positive . We think , Optimizing some difficult sample pairs may deteriorate the generalization ability of the model . This article calls it Positive alignment problem ( positive alignment problem). Because between the domains gap It could be very big , Therefore, directly aligning positive samples from different domains may be harmful to the model . This paper proposes a prototype based contrast loss to solve this problem . The specific differences between the two are as follows :

Style Neophile

Paper title :Style Neophile: Constantly Seeking Novel Styles for Domain Generalization
Thesis link :
https://openaccess.thecvf.com/content/CVPR2022/papers/Kang_Style_Neophile_Constantly_Seeking_Novel_Styles_for_Domain_Generalization_CVPR_2022_paper.pdf
4.1 Motivation
At present, most DG Methods all mention learning one domain-invariant Characteristics of , Such methods usually assume that each domain has a different image style , then enforce The characteristics of images of the same class under different styles are as similar as possible . and , These methods are limited to a limited set of styles ( The number of fields we get is finite ), Because they get enhanced patterns from a set of fixed training images or by interpolating training data . This paper presents a new method , Can generate more styles of data , These styles haven't even appeared in the training set .
4.2 Method
This paper describes the so-called image style by using the mean and variance of the image , For images feature map , We have .

The overall framework of this paper is as follows , Let's introduce each part in turn .

summary
Four articles are selected in this paper CVPR And DG Related articles , They are from cause and effect invariance , Pre training mode , Self supervised learning +DG And data enhancement , Overcome the shortcomings of the past . But for now , There is still no milestone approach that can be applied in the vast majority of OOD Of benchmark Show overwhelming effect on , Better, faster and stronger methods are still lacking at present .
边栏推荐
- 高阶柱状图之极环图与极扇图
- Redis encapsulation instance
- 从CVPR 2022看域泛化(Domain Generalization)最新研究进展
- Take you through the distributed file system
- Lighthouse open source application practice: snipe it
- 嵌入式开发:嵌入式基础——重启和重置的区别
- How does the hybrid cloud realize the IP sec VPN cloud networking dedicated line to realize the interworking between the active and standby intranet?
- Raid card with hardware knowledge (5)
- Leetcode algorithm interview sprint sorting algorithm theory (32)
- Cloud database smooth disassembly scheme
猜你喜欢

Selenium批量查询运动员技术等级

Outlook開機自啟+關閉時最小化

Experiment 5 module, package and Library

Code implementation of CAD drawing online web measurement tool (measuring distance, area, angle, etc.)

北大、加州伯克利大学等联合| Domain-Adaptive Text Classification with Structured Knowledge from Unlabeled Data(基于未标记数据的结构化知识的领域自适应文本分类)

Leetcode algorithm interview sprint sorting algorithm theory (32)

Embedded development: embedded foundation -- the difference between restart and reset

CAD图在线Web测量工具代码实现(测量距离、面积、角度等)

Acl2022 | MVR: multi view document representation for open domain retrieval

Minimisé lorsque Outlook est allumé + éteint
随机推荐
Tencent cloud database tdsql elite challenge Q & A (real-time update)
《scikit-learn机器学习实战》简介
Meaning of the last seven digits of wider face
Code implementation of CAD drawing online web measurement tool (measuring distance, area, angle, etc.)
Go bubbling, cocktail, quick, insert sort
Detailed explanation of logical structure, physical structure and data operation
How to dynamically insert a picture into a QR code
MySQL de duplication query only keeps one latest record
What is the gold content of PMP certificate
Experiment 5 module, package and Library
v-chart
Using barcode software to make certificates
Bcdedit, used to adjust the machine startup parameters (safe mode, BootMenu display name, CPU, memory, etc.)
What is the use of PMP certification?
Open source C # WPF control library -newbeecoder UI usage guide (I)
[js] generate random array
蓝牙芯片|瑞萨和TI推出新蓝牙芯片,试试伦茨科技ST17H65蓝牙BLE5.2芯片
Installation and use of Minio
DM sub database and sub table DDL "optimistic coordination" mode introduction - tidb tool sharing
How to deploy the API gateway? Is it OK not to use the API gateway?