当前位置:网站首页>Development Series III of GaN (lapgan, srgan)
Development Series III of GaN (lapgan, srgan)
2022-07-24 17:33:00 【51CTO】
GAN Development Series III of (LapGAN、SRGAN)
We have already introduced it in the previous article GAN The introduction to generating countermeasure networks and some GAN series , In the following album will continue to introduce some of the more classic GAN.
GAN Introduction to generating countermeasure network
GAN The development of the series one (CGAN、DCGAN、WGAN、WGAN-GP、LSGAN、BEGAN)
GAN The development of series 2 (PGGAN、SinGAN)
One 、 LapGAN
The paper :《Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks》
Address of thesis :https://arxiv.org/abs/1506.05751
1、 The basic idea
LapGAN It's based on GAN and CGAN On the basis of , use Laplacian Pyramid The pyramid of Laplace To generate images from thick to thin , So as to generate high-resolution images . At each level of the pyramid, there are learning residuals with adjacent levels , By constantly stacking CGAN Get the final resolution .CGAN As we mentioned in the previous article, it is in GAN Add conditional constraints on the basis of , To alleviate the original GAN The generator generates samples too freely .
original GAN The formula of is :

CGAN The formula of is :

2、 The pyramid of Laplace
Laplacian pyramid is the result of continuous up sampling of images in scale space , Gaussian pyramid is the result of continuous down sampling of images in scale space . First build Gaussian pyramid , To image I0 For continuous K Next sampling , obtain

Is the first K The Laplace pyramid on level is

The Laplace pyramids on other levels are :

Laplacian pyramid No k The layer is equal to the Gaussian pyramid k Layer minus Gaussian pyramid k+1 Upper sampling of layer .
Use the Laplace pyramid to restore the image :

3、LapGAN principle
With K=3 For example , At this time, the pyramid of Laplace is 4 The layer structure , contain 4 A generator G0、G1、G2、G3, Generate... Separately 4 A resolution image 64x64、32x32、16x16、8x8, The lowest resolution image to train the original GAN, Input only noise , Later, higher resolution image training CGAN, Input the image sampled on the Gaussian pyramid image with noise and the same level .
LapGAN Through a series of CGAN In series , Constantly generate higher resolution .


LAPGAN stay CIFAR10、STL10 and LSUN Experiments were conducted on three data sets , The generated image is as follows :

Two 、 SRGAN
The paper 《Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network》
Address of thesis :https://arxiv.org/pdf/1609.04802.pdf
Code address :https://github.com/OUCMachineLearning/OUCML/blob/master/GAN/srgan_celebA/srgan.py
1、 The basic idea
SRGAN Yes, it will GAN Applied to the field of image super-resolution ,CNN Convolutional neural network has achieved very good results in the traditional super-resolution reconstruction , High peak signal-to-noise ratio can be achieved PSNR, With MSE Is the objective function of minimization .SRGAN Is the first to recover 4 The algorithm framework of down sampling image , The author proposes a perceptual loss function , Including confrontation loss and content loss , The counter loss comes from the discriminator , It is used to distinguish the real image from the generated super-resolution image , Content loss focuses on visual similarity .

Usually, image super-resolution algorithm uses the mean square error between the reconstructed super-resolution image and the real image MSE As an objective function , Optimize MSE So as to improve PSNR, however MSE and PSNR The value of is not a good indicator of the visual effect , The following figure PSNR The vision with the highest value is not good .

2、 Network structure
Usually per pixel MSE Due to excessive smoothing, it is difficult to deal with the super-resolution details of the image , This paper designs a new loss function , Will be per pixel MSE Replace loss with content loss . Perceived loss is expressed as the weighted sum of content loss and adversarial loss ,

Content Loss It is the loss per pixel of the feature map of a certain layer as the content loss ,

Adversarial Loss Against the loss

The network structure proposed by the author is as follows , The generator consists of a residual structure Residual blocks form ,

The author uses sub-pixel Network as a generative network , use VGG As a discriminant network GAN Got very good results , But this uses the difference per pixel as the loss function . after , The author tries to use the perceptual loss function proposed by himself as the optimization goal , although PSNR and SSIM Not high , But the visual effect is better than other networks , Avoid the over smooth characteristics of other methods .



边栏推荐
- Explain Apache Hudi schema evolution in detail
- The most powerful programmer on the earth is equipped with a "three piece set". Do you know what it is?
- Preliminary understanding of redis
- Atcoder beginer 202 e - count descendants (heuristic merge on heavy chain split tree for offline query)
- Cann training camp learns the animation stylization and AOE ATC tuning of the second season of 2022 model series
- Getaverse, a distant bridge to Web3
- Apachecon Asia 2022 opens registration: pulsar technology issues make a big debut
- Socat port forwarding
- hcip第三天
- Natbypass port forwarding
猜你喜欢

Iqiyi Tiktok reconciled, Weibo lying gun?

还在用Xshell?你out了,推荐一个更现代的终端连接工具!

Getaverse,走向Web3的远方桥梁

【GNN报告】腾讯AI lab 徐挺洋:图生成模型及其在分子生成中的应用

Kernel development

I'll teach you how to use NPs to build intranet penetration services. When you go out, you can easily connect your lightweight notebook to your home game console to play remotely

Exception handling - a small case that takes you to solve NullPointerException

图像像素的逻辑操作

Array learning navigation
Shardingsphere database read / write separation
随机推荐
Apachecon Asia 2022 opens registration: pulsar technology issues make a big debut
Atcoder beginer 202 e - count descendants (heuristic merge on heavy chain split tree for offline query)
Practical application cases of digital Twins - Smart Park
Heuristic merging (including examples of general formula and tree heuristic merging)
Shardingsphere database read / write separation
2022-07-21 Daily: Wu Enda wrote: how to establish projects suitable for AI career
Natbypass port forwarding
NC port forwarding
Openlayers: point aggregation effect
Iftnews | Christie's launched its venture capital department, aiming at Web3 and metauniverse industries
Yolopose practice: one-stage human posture estimation with hands + code interpretation
【GNN报告】腾讯AI lab 徐挺洋:图生成模型及其在分子生成中的应用
近30所高校,获教育部点名表扬!
Analyze the capabilities and scenarios of Apache pulsar, a cloud native message flow system
通道的分离与合并
[spoken English] 01 - Introduction to atom
It's time to consider slimming down your app
Three.js (7): local texture refresh
DF2NET三维模型部署
Logical operation of image pixels