当前位置:网站首页>biggan:large scale gan training for high fidelity natural image synthesis
biggan:large scale gan training for high fidelity natural image synthesis
2022-06-26 01:45:00 【Kun Li】
In depth reading DeepMind new : The strongest in history GAN Image Generator —BigGAN - You know The paper notes recommended in this issue are from PaperWeekly Community users @TwistedW. from DeepMind It brings BigGAN It is the best I have ever seen GAN Model , there Big Not just model parameters and Batch The big , It still seems to imply impressive , The article also does …
https://zhuanlan.zhihu.com/p/46581611 This article is easy to understand , Compared with the previous large-scale pendulum formula ,biggan It's much easier to understand , But the model structure diagram has been put in the appendix ,biggan There are three core points , The first is bigger , Big bs And big parameters , second , From the adjustment of the model , such as latent Hierarchical input of , contribution c, Third latent z Truncated input of , In response to this truncation input , Orthogonal regularization is proposed , Finally, based on the instability of training, some improved skills are put forward .
1.introduction
gan Benefit from scaling , Compared with the prior art , The training parameters are 2-4 times ,batch Big 8 times , Reached 2048 This level , Corresponding to the topic large scale, Actually biggan The discovery is just a simple adjustment ,gan Can have a better effect . As a side effect of the modification , The model becomes suitable for the truncation technique ,truncation trick, This is a simple sampling technique , The trade-off between sample type and fidelity can be made clear , Fine-grained control .
2.background

3.scaling up gans
Exploration and expansion gan How to train , In order to get in larger models and larger batch Advantages under . Use SAGAN framework ,hinge loss, Conditions bn Of G, With projection D, Spectral normalization ,moving averages of G weights, Orthogonal initialization .
First add baseline model Of bs,IS Improve quickly , But the side effect is that our model is better in fewer iterations , But it becomes unstable and prone to mode crash .
Width increase 50%, The effect continues to improve .

Look at the picture above ,c Is Shared , Yes latent z Conduct split, Noise vector z Fed into multiple layers instead of just the initial layer , This is very meaningful , stay stylegan That's what we do in China ,latent z In fact, it is useful for every volume layer in the generator , Convolution layer is actually a process from coarseness to refinement , from latent z Spatial sampling , The intuition behind this design is to allow the generator to use potential space to directly affect features at different levels of resolution and hierarchy . stay bigan in , take z Divided into pieces of each resolution , And link each block to a condition vector c, Condition vector c Mapping to BN Gain and deviation of .
3.1 trading off variety and fidelity with the truncation trick
gan Any prior can be used , But most use Gauss and uniform distribution , There can be better alternatives . The so-called truncation technique , It's through a priori distribution z sampling , Truncate by setting a threshold z Sampling of , Values out of range are resampled and fall into the range , This threshold can be generated according to the quality index IS and FID decision . This experiment can be known by setting the threshold , As the threshold decreases, the quality of the generation will get better , But as the threshold drops , The sampling range is narrowed , It will lead to the simplification of the generative orientation , The problem of insufficient diversity of generation .

Above picture a As the threshold gets lower and lower , The effect is getting better and better , But more and more single ,b The graph is a large model, which is not suitable for truncation , Prone to saturation artifacts . To offset this , We're trying to put G Adjust to smooth to force the adaptability of truncation , In order to z The entire space to a good output sample . Sampling orthogonal regularization .
4.analysis
It mainly analyzes gan Reasons for instability during training , And make some restrictions on the generator and discriminator .
边栏推荐
- Viwi interface
- Interpretation of script corresponding to postman assertion
- Fasttext knowledge points summary
- JSON instance (I)
- 从在浏览器的输入框输入一个网址,到看到网页的内容,这个过程中发生了什么?
- **MySQL例题一(根据不同问题,多条件查询)**
- GUN make (3) Makefile的规则
- PTA class a simulated fifth bomb: 1148-1151
- Operation of simulated examination platform for electrical test questions in 2022
- JQ获取对象的几种方式
猜你喜欢

Talking about interface test (2)

23. histogram equalization

readv & writev

27. template match

What happens from entering a web address in the browser's input box to seeing the contents of the web page?

物联网亿万级通信一站式解决方案EMQ

GUN make (2) 总述

Textcnn paper Interpretation -- revolutionary neural networks for sense classification

清甜女孩李斯霞 受邀担任第六季完美童模全球总决赛小主持人

recv & send
随机推荐
俏皮少女王艺璇 受邀担任第六季完美童模全球总决赛推广大使
easyexcel读取文件
ActivityManager kill reason
Postman断言对应脚本的解释
**MySQL例题一(根据不同问题,多条件查询)**
Simple making of master seal
Rollback Protection
Installing MySQL databases in FreeBSD
tos cos dscp 区别和作用
PTA class a simulated second bullet: 1136-1139
PTA class a simulated third bomb: 1140-1143
2021-1-15 fishing notes ctrl+c /v
JQ user defined attribute value
21. Hoff circle transformation
图文大师印章简易制作
Log4j2 vulnerability
[visual studio code] vscode shortcut keys
Model integration and cascading
recvmsg & sendmsg
程序 编译的步骤 预编译 编译 汇编 连接