Changes

Jump to: navigation, search
m
no edit summary
It uses two NNs, specifically two CNNs in the most recent approaches, which are locked in a competition game:
* A generator, which is fed a vector of random numbers i.e. the latent vector and outputs synthesized data i.e. the generated images.
* A discriminator, which is fed a batch of data, in this case a batch of images, and outputs a prediction of it being from the training set or from the generated set, basically learning a binary classification problem. In other words, the generator creates fake data and the discriminator attempts to distinguish these fakes samples from the real ones.
It must be specified that GANs in practice are quite complex and training can be a very challenging task making the generation from scratch of high resolution quality images a non trivial problem. This indeed severely limits the usefulness and the applicability of classic GANs architectures for many kind of practical applications. Fortunately, this issue can be addressed by employing a particular typology of networks developed by NVIDIA and named as proGANs which are characterized by a progressive growing architecture.
This approach leads to a series of advantages:
*The incremental learning process greatly stabilizes training and reduces the chance of mode collapse, since the networks gradually learn a much simple piece of the overall problem.
*The low-to-high resolution trend forces the progressively grown networks to focus on high-level structure first and fill in the details later, resulting in an improvement of the quality of the final images.
*Increasing the network size gradually is more computationally efficient w.r.t. the classic approach of using all the layers from the start (fewer layers are faster to train because there are fewer parameters).
==Results validation==
In the two groups of images below, some examples taken from the generated sets of ''missing'' and ''tombstoning'' classes are shown. For each group, from left to right, you can see:
* Samples of 512×512 and 256×256 resolution ''full'' typology.
* Further on the right, from top to bottom, 256×256 ''upper'' and ''lower'' soldering region typology.
The values obtained by applying the t-SNE algorithm on the compressed data are displayed in several 3D plots reported below respectively for ''missing'' and ''tombstoning'' synthesized images. Interestingly, the 4 plots related to full typology do not show a separation in the data between ''fakes'' (red dots) and ''reals'' (blue dots), sign that the generated images do not differ too much from the original ones, which were used to train the proGANs.
The two figures to the utmost right report the plots showing the results of t-SNE on the compressed data, obtained from the dual stream encoder for ''missing'' and ''tombstoning'' synthesized images. In the first case, for ''missing'' class, there is actually no separation between fakes and real of ''upper'' and ''lower'' region sides, a good sign that the generated images are no not very different from the original ones. This is quite understandable since the two regions look alike. The situation is visibly different in the plot for the ''tombstoning class'', where two distinct clouds of dots can be seen, confirming a well defined separation between the data related to the ''upper'' and ''lower'' typologies. It is also possible to see that fakes and reals dots, belonging to the same soldering region, are very close to each other, while dots belonging to a different soldering region are distant in the plot, confirming once again that the image generation with the proGAN was successful.
{| style="background:transparent; color:black" border="0" align="center" cellpadding="10px" cellspacing="0px" height="550" valign="bottom"
* Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, Aaron Courville, [https://arxiv.org/abs/1704.00028 ''Improved Training of Wasserstein GANs''], December 2017.
*[https://shop.ipc.org/IPC-A-610E-English-D ''IPC-A-610E: Acceptability of Electronic Assemblies''], a standard developed by IPC, April 2010.
*Alejandro Betancourt, [https://www.industryweek.com/technology-and-iiot/digital-tools/article/21122846/making-ai-work-with-small-data Making AI Work with Small Data], February 12th12, 2020
dave_user
207
edits

Navigation menu