{"id":26410,"date":"2021-06-07T06:00:00","date_gmt":"2021-06-07T13:00:00","guid":{"rendered":"https:\/\/insidebigdata.com\/?p=26410"},"modified":"2021-06-08T12:07:57","modified_gmt":"2021-06-08T19:07:57","slug":"applying-gans-to-image-generation-tasks","status":"publish","type":"post","link":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/","title":{"rendered":"Applying GANs to Image Generation Tasks"},"content":{"rendered":"\n<p>This article overviews StyleGAN2 application to image generation task and is based on MobiDev\u2019s logotype synthesis <a href=\"https:\/\/mobidev.biz\/blog\/gan-image-generation-with-stylegan2\" target=\"_blank\" rel=\"noreferrer noopener\">research<\/a>.&nbsp;<\/p>\n\n\n\n<p>When it comes to powerful generative models for image synthesis, the most commonly mentioned are <a href=\"https:\/\/github.com\/NVlabs\/stylegan\" target=\"_blank\" rel=\"noreferrer noopener\">StyleGAN<\/a> and its updated version <a href=\"https:\/\/github.com\/NVlabs\/stylegan2\" target=\"_blank\" rel=\"noreferrer noopener\">StyleGAN2<\/a>. <strong>These models created by Nvidia Labs are able to solve image generation tasks and produce remarkably high fidelity images of non-existent people, animals, landscapes, and other objects given an appropriate training dataset<\/strong>.<\/p>\n\n\n\n<p>StyleGAN, just like the other GAN architectures, features two sub-networks: Discriminator and Generator. During the training, the Generator is tasked with producing synthetic images while the Discriminator is trained to differentiate between the fakes from Generator and the real images.<\/p>\n\n\n\n<p>The first iteration of StyleGAN appeared in 2019. It was applied to produce fake faces with high detailization and natural appearance with resolutions up 1024&#215;1024, not previously achieved by other similar models. However, some AI-generated faces had artifacts, so Nvidia Labs decided to improve the model and presented StyleGAN2. One of the main issues of&nbsp; the original StyleGAN were blob-like artifacts that looked like drops of water.&nbsp;<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"375\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_1.png\" alt=\"\" class=\"wp-image-26411\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_1.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_1-150x80.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_1-300x161.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><figcaption>Figure 1. A blob-like artifact example in the image generated by StyleGan.<strong> <\/strong><a href=\"https:\/\/arxiv.org\/pdf\/1912.04958.pdf\">Source<\/a><\/figcaption><\/figure><\/div>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"525\" height=\"221\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_2.png\" alt=\"\" class=\"wp-image-26412\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_2.png 525w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_2-150x63.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_2-300x126.png 300w\" sizes=\"(max-width: 525px) 100vw, 525px\" \/><figcaption><strong>Figure 2.<\/strong> AdaIN based Style Transfer Network <a href=\"https:\/\/towardsdatascience.com\/fast-and-arbitrary-style-transfer-40e29d308dd3\">Source<\/a>&nbsp;<\/figcaption><\/figure><\/div>\n\n\n\n<p>According to the StyleGAN2 paper, this problem is related to the instance normalization operation applied in AdaIN layers. The basic purpose of AdaIN is to fuse together two images: one containing style (style is a property that is present throughout all the image) and one containing the content of the image. Part of the crucial information of the inputs is being lost during this process. Why is this happening?&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;<\/p>\n\n\n\n<p>AdaIN works separately with the variance and mean of individual feature maps. The feature map is an intermediate representation of an image within the neural network. Feature maps carry the information regarding the image that is being generated and normalize their values.&nbsp;<\/p>\n\n\n\n<p>As each feature map is normalized individually, information about the relative feature map values in regard to each other is lost in this process. As a result, the generator network produces the blob artifact with a strong signal within the image, effectively bypassing the AdaIN in an attempt to preserve the relative information that the normalization destroys.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"383\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_3.png\" alt=\"\" class=\"wp-image-26413\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_3.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_3-150x82.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_3-300x164.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><\/figure><\/div>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"466\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_4.png\" alt=\"\" class=\"wp-image-26414\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_4.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_4-150x100.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_4-300x200.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><figcaption>Figure 3. StyleGAN synthesis network architecture redesign. <a href=\"https:\/\/arxiv.org\/pdf\/1912.04958.pdf\">Source<\/a><\/figcaption><\/figure><\/div>\n\n\n\n<p>The development of StyleGAN2 was related not only to addressing the issue with AdaIN, there were other improvements in comparison with StyleGAN:<\/p>\n\n\n\n<ul><li><strong>Network redesign <\/strong>&#8211; small changes were made in the order and type of operations within the style blocks of the network leading to small positive improvements in the network\u2019s performance.<\/li><li><strong>Lazy regularization <\/strong>&#8211; researchers found out that it is not necessary to compute regularization terms in addition to the main loss function at every training iteration. Instead, it can be computed periodically thus reducing the computational costs while having little to no effect on the quality of the results.<\/li><li><strong>Path length regularization<\/strong> &#8211; it was observed that uniformity of latent space W (this space is sampled from to obtain styles for image generation) had a positive effect on image quality. Uniformity was achieved by making sure that a fixed size step when sampling from W space led to a fixed size change in the generated image regardless of the direction of the step.<\/li><li><strong>Progressive growth replaced<\/strong> &#8211; the original architecture used progressive growing to train the model for higher resolutions (at first the model is trained to generate 8&#215;8 images, then 16&#215;16, and so on), however this method introduced certain artifacts in the produced images. Therefore, the authors decided to use skip connections in generator and discriminator instead, managing to avoid the aforementioned problem.<\/li><\/ul>\n\n\n\n<p>You have read about StyleGAN and StyleGAN2, but you can dive deeper into the peculiarities of architecture by getting acquainted with the next section of the article.&nbsp;<\/p>\n\n\n\n<p><strong>What\u2019s Notable in StyleGAN Architecture?<\/strong><\/p>\n\n\n\n<p>An important differentiating feature about StyleGAN is the fact that the latent vector z is not injected into the model directly like in the traditional architectures but is instead first mapped onto a latent space which has separate vectors controlling the style of the generated image at different resolutions, from 4&#215;4 all the way up to 1024&#215;1024.<\/p>\n\n\n\n<p>The influence of style vectors injected into the generator is localized using AdaIN layers (replaced with demodulation in StyleGan2), making sure one style vector affects only one convolution operation before AdaIN is applied once more. Finally, while style vectors control the overall content of the image (gender, age, hair type, skin color, etc.), randomized noise vectors add some variation into the generated image, e.g skin details, hair placement etc.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"460\" height=\"383\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_5.png\" alt=\"\" class=\"wp-image-26415\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_5.png 460w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_5-150x125.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_5-300x250.png 300w\" sizes=\"(max-width: 460px) 100vw, 460px\" \/><figcaption><strong>Figure 4.<\/strong> Image synthesis process modification. <a href=\"https:\/\/neurohive.io\/en\/state-of-the-art\/stylegan-code-released\/\">Source<\/a><\/figcaption><\/figure><\/div>\n\n\n\n<p>The architecture has attracted much attention. Researchers found ways to generate images better and project the real images into the latent space of the model (this task can also be called <strong>inversion<\/strong>). For a model that was trained on facial images, the projection process essentially results in finding synthetic doubles of the real people. When such a double is found, all manner of manipulations can be done to modify the double\u2019s appearance.<\/p>\n\n\n\n<p>Presumably, with the <a href=\"https:\/\/github.com\/eladrich\/pixel2style2pixel\" target=\"_blank\" rel=\"noreferrer noopener\">Pixel2Style2Pixel<\/a> framework, StyleGAN architecture can essentially be turned into a jack-of-all-trades image editing tool. Some of the tasks it can be used for are improved inversion, face pose editing, super-resolution, facial generation from sketches, or segmentation maps, as illustrated in the Fig. 5.<\/p>\n\n\n\n<p>Many of the features demonstrated in the developed products (e.g. face in editing in <a href=\"https:\/\/www.faceapp.com\/\" target=\"_blank\" rel=\"noreferrer noopener\">FaceApp<\/a>) closely resemble the outputs of Pixel2Style2Pixel, giving us some solid grounds for assumptions about which techniques the businesses use under the hood.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"421\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_6.png\" alt=\"\" class=\"wp-image-26416\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_6.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_6-150x90.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_6-300x180.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><figcaption>Figure 5. Image editing examples from Pixel2Style2Pixel framework for StyleGAN. <a href=\"https:\/\/github.com\/eladrich\/pixel2style2pixel\">Source<\/a><\/figcaption><\/figure><\/div>\n\n\n\n<p><strong>Applying GANs to Logotype Synthesis<\/strong><\/p>\n\n\n\n<p>As a part of internal research in our company, we applied the StyleGAN2 architecture to a problem of logotype synthesis. We used all the recent developments and improvements introduced by researchers &#8211; <a href=\"https:\/\/arxiv.org\/abs\/2006.06676\" target=\"_blank\" rel=\"noreferrer noopener\">adaptive discriminator augmentation<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/1710.03740\" target=\"_blank\" rel=\"noreferrer noopener\">mixed-precision training<\/a>, and <a href=\"https:\/\/arxiv.org\/abs\/1805.08318\" target=\"_blank\" rel=\"noreferrer noopener\">self-attention layers<\/a>. To receive GAN generated logotypes almost 49,000 images were used as a training dataset.&nbsp;<\/p>\n\n\n\n<p>Text-based images were removed because generating textual logotypes requires at least three models. For instance, the language model like <a href=\"https:\/\/github.com\/google-research\/bert\" target=\"_blank\" rel=\"noreferrer noopener\">BERT<\/a> or <a href=\"https:\/\/openai.com\/blog\/gpt-2-1-5b-release\/\" target=\"_blank\" rel=\"noreferrer noopener\">GPT-2<\/a> to produce synthetic logotype text, <a href=\"https:\/\/arxiv.org\/abs\/1905.12502v1\" target=\"_blank\" rel=\"noreferrer noopener\">GlyphGAN<\/a> model that creates characters with unique fonts to visualize the synthetic text, and a third model for generating the logotype itself.&nbsp;<\/p>\n\n\n\n<p>The training dataset was aggregated into 10 clusters; this information helped the model in generating images from various logotype groups. As a result, the model was able to generate logotypes whose quality was ranging from good to medium and poor. The results were carefully analyzed to understand the reason for the model\u2019s outputs and find the ways to improve it in the future.<\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"aligncenter size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"700\" height=\"342\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_7.png\" alt=\"\" class=\"wp-image-26417\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_7.png 700w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_7-150x73.png 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/MobiDev_7-300x147.png 300w\" sizes=\"(max-width: 700px) 100vw, 700px\" \/><figcaption><strong>Figure 6<\/strong>. Examples of logotype images generated by StyleGAN2. <a href=\"https:\/\/mobidev.biz\/blog\/gans-technology-use-cases-for-business-application\">Source<\/a><\/figcaption><\/figure><\/div>\n\n\n\n<p><strong>The Future of GANs<\/strong><\/p>\n\n\n\n<p>The research showed that AI has a huge potential in the area of image creation. But the question remains: what else could be achieved with the help of GANs? The <a href=\"https:\/\/www.nature.com\/articles\/s41524-020-00352-0\" target=\"_blank\" rel=\"noreferrer noopener\">paper published in Nature<\/a> suggests that machine learning could significantly contribute to material science as GAN architecture was found to be able to generate chemically valid hypothetical inorganic compounds. Even though the system is still missing some key components (prediction of hypothetical materials\u2019 crystal structure, addition of strict chemical rule filters) the published results mean one day machine learning may be responsible for inventing new materials such as ultra-light durable alloys, solid state electrolytes for Li-Ion batteries, etc.<\/p>\n\n\n\n<p><strong>About the Author<\/strong><\/p>\n\n\n\n<div class=\"wp-block-image is-style-default\"><figure class=\"alignleft size-large\"><img decoding=\"async\" loading=\"lazy\" width=\"150\" height=\"150\" src=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/maksym-tatariants.jpg\" alt=\"\" class=\"wp-image-26418\" srcset=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/maksym-tatariants.jpg 150w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/maksym-tatariants-110x110.jpg 110w, https:\/\/insidebigdata.com\/wp-content\/uploads\/2021\/06\/maksym-tatariants-50x50.jpg 50w\" sizes=\"(max-width: 150px) 100vw, 150px\" \/><\/figure><\/div>\n\n\n\n<p><em>Maksym Tatariants, PhD is an AI Engineer at <a href=\"https:\/\/mobidev.biz\/\" target=\"_blank\" rel=\"noreferrer noopener\">MobiDev<\/a>. Maksym obtained Master\u2019s Degree in Mechanical Engineering, dealing with design of equipment for Renewable Energy applications. His engineering background helps him gain new insights and skills in applying Machine Learning, Deep Learning, and Computer Vision to technology and science projects.<\/em><\/p>\n\n\n\n<p><em>Sign up for the free insideBIGDATA&nbsp;<a rel=\"noreferrer noopener\" href=\"http:\/\/insidebigdata.com\/newsletter\/\" target=\"_blank\">newsletter<\/a>.<\/em><\/p>\n\n\n\n<p><em>Join us on Twitter:&nbsp;@InsideBigData1 \u2013 <a href=\"https:\/\/twitter.com\/InsideBigData1\" target=\"_blank\" rel=\"noreferrer noopener\">https:\/\/twitter.com\/InsideBigData1<\/a><\/em><\/p>\n","protected":false},"excerpt":{"rendered":"<p>This contributed article by Maksym Tatariants, PhD, AI Engineer at MobiDev, overviews StyleGAN2 application to image generation task and is based on MobiDev\u2019s logotype synthesis research. When it comes to powerful generative models for image synthesis, the most commonly mentioned are StyleGAN and its updated version StyleGAN2.<\/p>\n","protected":false},"author":10513,"featured_media":21334,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"jetpack_post_was_ever_published":false,"footnotes":""},"categories":[526,115,87,180,67,56,97,78,1],"tags":[264,797,1017,263,95],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v20.6 - https:\/\/yoast.com\/wordpress\/plugins\/seo\/ -->\n<title>Applying GANs to Image Generation Tasks - insideBIGDATA<\/title>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Applying GANs to Image Generation Tasks - insideBIGDATA\" \/>\n<meta property=\"og:description\" content=\"This contributed article by Maksym Tatariants, PhD, AI Engineer at MobiDev, overviews StyleGAN2 application to image generation task and is based on MobiDev\u2019s logotype synthesis research. When it comes to powerful generative models for image synthesis, the most commonly mentioned are StyleGAN and its updated version StyleGAN2.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/\" \/>\n<meta property=\"og:site_name\" content=\"insideBIGDATA\" \/>\n<meta property=\"article:publisher\" content=\"http:\/\/www.facebook.com\/insidebigdata\" \/>\n<meta property=\"article:published_time\" content=\"2021-06-07T13:00:00+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2021-06-08T19:07:57+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/insidebigdata.com\/wp-content\/uploads\/2018\/10\/shutterstock_1096541144.jpg\" \/>\n\t<meta property=\"og:image:width\" content=\"1000\" \/>\n\t<meta property=\"og:image:height\" content=\"568\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Editorial Team\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:creator\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:site\" content=\"@insideBigData\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Editorial Team\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"8 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\/\/schema.org\",\"@graph\":[{\"@type\":\"WebPage\",\"@id\":\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/\",\"url\":\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/\",\"name\":\"Applying GANs to Image Generation Tasks - insideBIGDATA\",\"isPartOf\":{\"@id\":\"https:\/\/insidebigdata.com\/#website\"},\"datePublished\":\"2021-06-07T13:00:00+00:00\",\"dateModified\":\"2021-06-08T19:07:57+00:00\",\"author\":{\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9\"},\"breadcrumb\":{\"@id\":\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\/\/insidebigdata.com\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Applying GANs to Image Generation Tasks\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\/\/insidebigdata.com\/#website\",\"url\":\"https:\/\/insidebigdata.com\/\",\"name\":\"insideBIGDATA\",\"description\":\"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies\",\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\/\/insidebigdata.com\/?s={search_term_string}\"},\"query-input\":\"required name=search_term_string\"}],\"inLanguage\":\"en-US\"},{\"@type\":\"Person\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9\",\"name\":\"Editorial Team\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/\",\"url\":\"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g\",\"contentUrl\":\"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g\",\"caption\":\"Editorial Team\"},\"sameAs\":[\"http:\/\/www.insidebigdata.com\"],\"url\":\"https:\/\/insidebigdata.com\/author\/editorial\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Applying GANs to Image Generation Tasks - insideBIGDATA","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/","og_locale":"en_US","og_type":"article","og_title":"Applying GANs to Image Generation Tasks - insideBIGDATA","og_description":"This contributed article by Maksym Tatariants, PhD, AI Engineer at MobiDev, overviews StyleGAN2 application to image generation task and is based on MobiDev\u2019s logotype synthesis research. When it comes to powerful generative models for image synthesis, the most commonly mentioned are StyleGAN and its updated version StyleGAN2.","og_url":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/","og_site_name":"insideBIGDATA","article_publisher":"http:\/\/www.facebook.com\/insidebigdata","article_published_time":"2021-06-07T13:00:00+00:00","article_modified_time":"2021-06-08T19:07:57+00:00","og_image":[{"width":1000,"height":568,"url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2018\/10\/shutterstock_1096541144.jpg","type":"image\/jpeg"}],"author":"Editorial Team","twitter_card":"summary_large_image","twitter_creator":"@insideBigData","twitter_site":"@insideBigData","twitter_misc":{"Written by":"Editorial Team","Est. reading time":"8 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"WebPage","@id":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/","url":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/","name":"Applying GANs to Image Generation Tasks - insideBIGDATA","isPartOf":{"@id":"https:\/\/insidebigdata.com\/#website"},"datePublished":"2021-06-07T13:00:00+00:00","dateModified":"2021-06-08T19:07:57+00:00","author":{"@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9"},"breadcrumb":{"@id":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/insidebigdata.com\/2021\/06\/07\/applying-gans-to-image-generation-tasks\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/insidebigdata.com\/"},{"@type":"ListItem","position":2,"name":"Applying GANs to Image Generation Tasks"}]},{"@type":"WebSite","@id":"https:\/\/insidebigdata.com\/#website","url":"https:\/\/insidebigdata.com\/","name":"insideBIGDATA","description":"Your Source for AI, Data Science, Deep Learning &amp; Machine Learning Strategies","potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/insidebigdata.com\/?s={search_term_string}"},"query-input":"required name=search_term_string"}],"inLanguage":"en-US"},{"@type":"Person","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/2949e412c144601cdbcc803bd234e1b9","name":"Editorial Team","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/insidebigdata.com\/#\/schema\/person\/image\/","url":"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/e137ce7ea40e38bd4d25bb7860cfe3e4?s=96&d=mm&r=g","caption":"Editorial Team"},"sameAs":["http:\/\/www.insidebigdata.com"],"url":"https:\/\/insidebigdata.com\/author\/editorial\/"}]}},"jetpack_featured_media_url":"https:\/\/insidebigdata.com\/wp-content\/uploads\/2018\/10\/shutterstock_1096541144.jpg","jetpack_shortlink":"https:\/\/wp.me\/p9eA3j-6RY","jetpack-related-posts":[{"id":28141,"url":"https:\/\/insidebigdata.com\/2022\/01\/06\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-december-2021\/","url_meta":{"origin":26410,"position":0},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 December 2021","date":"January 6, 2022","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":30176,"url":"https:\/\/insidebigdata.com\/2022\/08\/24\/research-highlights-an-image-is-worth-one-word-personalizing-text-to-image-generation-using-textual-inversion\/","url_meta":{"origin":26410,"position":1},"title":"Research Highlights: An Image is Worth One Word: Personalizing Text-to-Image Generation using Textual Inversion","date":"August 24, 2022","format":false,"excerpt":"In this regular column we take a look at highlights for breaking research topics of the day in the areas of big data, data science, machine learning, AI and deep learning. For data scientists, it\u2019s important to keep connected with the research arm of the field in order to understand\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2022\/08\/Research_highlights_5.png?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":23280,"url":"https:\/\/insidebigdata.com\/2019\/09\/18\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-august-2019\/","url_meta":{"origin":26410,"position":2},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 August 2019","date":"September 18, 2019","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":26425,"url":"https:\/\/insidebigdata.com\/2021\/06\/09\/best-of-arxiv-org-for-ai-machine-learning-and-deep-learning-may-2021\/","url_meta":{"origin":26410,"position":3},"title":"Best of arXiv.org for AI, Machine Learning, and Deep Learning \u2013 May 2021","date":"June 9, 2021","format":false,"excerpt":"In this recurring monthly feature, we will filter all the recent research papers appearing in the arXiv.org preprint server for subjects relating to AI, machine learning and deep learning \u2013 from disciplines including statistics, mathematics and computer science \u2013 and provide you with a useful \u201cbest of\u201d list for the\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2013\/12\/arxiv.jpg?resize=350%2C200&ssl=1","width":350,"height":200},"classes":[]},{"id":25323,"url":"https:\/\/insidebigdata.com\/2020\/12\/06\/comparative-testing-of-gpu-servers-with-new-nvidia-rtx30-video-cards-in-ai-ml-tasks\/","url_meta":{"origin":26410,"position":4},"title":"Comparative Testing of GPU Servers with New NVIDIA RTX30 Video Cards in AI\/ML Tasks","date":"December 6, 2020","format":false,"excerpt":"In early September 2020, NVIDIA debuted its second generation GeForce RTX 30 family of graphics cards, the Ampere RTX architecture. NVIDIA broke with tradition when its new generations of cards were sold more expensive than their predecessors, which means that the cost of training models has remained more or less\u2026","rel":"","context":"In &quot;AI Deep Learning&quot;","img":{"alt_text":"","src":"","width":0,"height":0},"classes":[]},{"id":17874,"url":"https:\/\/insidebigdata.com\/2017\/05\/29\/enterprise-software-tools-ai\/","url_meta":{"origin":26410,"position":5},"title":"Enterprise Software Tools for AI","date":"May 29, 2017","format":false,"excerpt":"This post delves into a variety of enterprise software options for AI. This article is part of a special insideHPC report that explores trends in machine learning and deep learning. Find out how businesses are using machine learning and deep learning, differentiating between AI, machine learning and deep learning, what\u2026","rel":"","context":"In &quot;Featured&quot;","img":{"alt_text":"","src":"https:\/\/i0.wp.com\/insidebigdata.com\/wp-content\/uploads\/2017\/04\/Dell_Nivida_ML_DLguide.jpg?resize=350%2C200","width":350,"height":200},"classes":[]}],"_links":{"self":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/26410"}],"collection":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/users\/10513"}],"replies":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/comments?post=26410"}],"version-history":[{"count":0,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/posts\/26410\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media\/21334"}],"wp:attachment":[{"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/media?parent=26410"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/categories?post=26410"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/insidebigdata.com\/wp-json\/wp\/v2\/tags?post=26410"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}