SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 06 | June 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1595
Creating Objects for Metaverse using GANs and Autoencoders
Aditya Singh1, Tejas Patil1
1Student, Dept. of Computer Engineering, Datta Meghe College of Engineering, Navi Mumbai, Maharashtra, India
---------------------------------------------------------------------***---------------------------------------------------------------------
Abstract – Metaverse refers to online spaces which allow
people to interact with one other in a more captivating way
than a website. This can be achieved through virtual reality
(VR) and augmented reality (AR). To make the Metaverse
more connected to reality, we make use of real-life objects
that translate into the Metaverse, that's where the need to
have state of an art model to achieve this goal. In this paper,
we have proposed a way to create such objects using
Generative Adversarial Networks (GANs) coupled with an
Autoencoder. GANs will create a new object, the object could
be anything from animals to human faces, cartoon figures
etc. With the creation of such objects, the goal is basically
achieved here. However, the output given by the GAN model
is not of very high-quality that's where Autoencoder comes
in with the use of Autoencoder the object is upscaled (super-
resolution).
Key Words: Metaverse, Generative Adversarial
Networks (GANs), Autoencoder
1. INTRODUCTION
Metaverse [1] is like ad immersive world where the user
can interact with others, attend a concert etc. which is
achieved with AR and VR technologies. In this VR world,
we may make our model and react with such different
models this paper aims to make this goal more realistic by
creating objects based on real-life entities with which
these models (people) can react.
1.1 Generative Adversarial Networks (GANs)
Generative Adversarial Networks [2] also known as GANs
consist of a pair of neural networks namely generator and
discriminator that compete with each other. As the name
suggests, the generator has to generate images while the
discriminator has to detect whether the image generated is
real or fake. The discriminator is fed at random real images
from the data set it's trained on and images generated by
the generator, it has to successfully identify the real image.
On the other hand, the generator has to generate a better-
quality image such that the discriminator fails to
discriminate between the real image and generated image.
The basic idea of a GAN model is to create a fake never seen
image with help of existing data. The model stops training
when the Nash equilibrium is achieved.
The following is the general loss function of GAN called
Minimax loss there are many other variations of this loss
function.
Ex[log(D(x))] + Ez[log(1-D(G(z)))]
In the above minimax loss equation, D(x) is the
discriminator’s estimate of the probability that the present
data is real. Ex is the expected value over all real data
instances. G(z) is the output given by the generator from
noise G. D(G(z)) is the estimate of the probability that the
fake instance is real
1.2 Autoencoders
Autoencoder [3] is an interesting variant of a Fully
connected neural network, It consists of three parts the
encoder, the bottleneck and the decoder. The encoder
takes the input and its compression is performed in order
to store the spatial data in the bottleneck which only
consists of three neurons, Therefore the bottleneck stores
the spatial data in a lower dimension. this data is further
fed into the decoder which is again a fully connected layer
and the original image is generated. The bottleneck is
forced to learn important information in order to
compress the data into lower dimensions such that using
the same data the original image can be generated.
Autoencoder just like GAN has many variations depending
upon the fully connected layer sometimes it's replaced by
pooling layers. Depending upon the variation of
Autoencoder the loss function may vary, mentioned below
is general a loss function:
min E(A, B) = min E(xt) = min ∆(A ◦ B(xt), xt)
2. PROPOSED WORK
In this paper, we propose a model which will generate real-
world objects including human faces with help of a
Generative Adversarial Network model and upscale the
pixel quality of the generated object with help of
Autoencoders, this generated object will be further used in
AR/VR application to serve the need of brining real-life feel
in Metaverse. For the GAN model, we made use of the Deep
Convolutional Generative Adversarial Network (DCGAN)
[4] which consists of the deep convolutional layer in the
generator and discriminator. In the case of the
autoencoder, we made use of both the convolutional layer
and pooling layer for both encoder and decoder. the GAN
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 06 | June 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1596
model and Autoencoder model were integrated and the
output generated was ready to be fed into the AR/VR
application. In the next section, we will look into this
process in more detail as we explore the methodology
Fig -1: Structure of the Application
3. METHODOLOGY
3.1 Generating images using GANs
A custom dataset was created by us which consisted of
human faces with more than 5000 images. the images
were preprocessed reshaped in 28x28x1 and fed into the
generator which consisted of a dense layer and three
convolutional layers, Batch normalization [5] was used
and leaky ReLU [6] as the activation function. In the case
of the discriminator three convolutional were used and in
the final layer, the sigmoid activation function was used.
For the optimizer, we chose the Adam optimizer to train
the discriminator.
Fig -1: Architecture of DCGAN Model
3.2 Super-image resolution using Autoencoders
The output generated from the DCGAN model is fed into
the autoencoder model. The autoencoder model consists
of three parts the encoder, the decoder and the bottleneck
which learns the latent space representation of the data.
For the encoder, we made use of 7 which consisted of
convolutional layers, Dropout layers and Pooling layers.
Meanwhile, for the decoder, we used 15 layers consisting
of convolutional layers, pooling layers and upscaling
layers. We made use of ReLU activation and L1
normalization. The training was done in batches.
Fig -2: Output of the Autoencoder Model
4. RESULTS
4.1 Result from the DCGAN Model
The following are outputs produced by the Deep
Convolutional Generative Adversarial Model (DCGAN).
Based on the 5000 image data the model was trained on.
Fig -1: First result generated by the DCGAN model
Fig -2: Second Result generated by the DCGAN model
4.2 Super Image Resolution using Autoencoders
After getting outputs from the DCGAN model the results
were fed into the Autoencoder model below are super
image resolutions of the outputs ready to be fed into
AR/VR applications
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 09 Issue: 06 | June 2022 www.irjet.net p-ISSN: 2395-0072
© 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1597
Fig -3: First Result generated by the Autoencoder model
Fig -4: Second Result generated by the Autoencoder model
5. CONCLUSION
In this paper, we have proposed a way to make the
metaverse more connected to the real world by including
human faces the approach can however be even used to
include other objects like cars, animals etc. The Deep
Convolutional Generative Adversarial network was
responsible for generating the images and Autoencoders
was used for image supper resolution to upscale the
images. The images generated can be used in AR/VR
applications.
REFERENCES
1) Stylianos Mystakidis, “Metaverse”, Encyclopedia,
https://doi.org/10.3390/encyclopedia2010031
2) Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi
Mirza, Bing Xu, David Warde-Farley,
Sherjil Ozair, Aaron Courville, Yoshua Bengio,
“Generative Adversarial
Networks”, arXiv:1406.2661v1
3) Dor Bank, Noam Koenigestein, Raja Giryes,
“Autoencoders”, arXiv:2003.05991v2
4) Alec Radford, Luke Metz, Soumith Chintala,
“Unsupervised Representation Learning with
Deep Convolitional Generative Adversarial
Networks”, arXiv:1511.06434v2
5) Segey Ioffe, Christian Szegedy, “Accelerating Deep
Network Training by Reducing Internal Covariate
Shift”,
https://doi.org/10.48550/arXiv.1502.03167
6) Bing Xu, Naiyan Wang, Tianqi Chen, Mu Li,
“Empirical Evaluation of Rectified Activation in
Convolutional Network”, arXiv:1505.00853v2

More Related Content

PDF
IRJET- Generating 3D Models Using 3D Generative Adversarial Network
PDF
Photo Editing And Sharing Web Application With AI- Assisted Features
PDF
IRJET - Hand Gesture Recognition to Perform System Operations
PDF
An Intelligent approach to Pic to Cartoon Conversion using White-box-cartooni...
PDF
IRJET - Single Image Super Resolution using Machine Learning
PDF
IRJET- Mango Classification using Convolutional Neural Networks
PDF
IRJET- Implementation of Gender Detection with Notice Board using Raspberry Pi
PDF
IRJET - Steering Wheel Angle Prediction for Self-Driving Cars
IRJET- Generating 3D Models Using 3D Generative Adversarial Network
Photo Editing And Sharing Web Application With AI- Assisted Features
IRJET - Hand Gesture Recognition to Perform System Operations
An Intelligent approach to Pic to Cartoon Conversion using White-box-cartooni...
IRJET - Single Image Super Resolution using Machine Learning
IRJET- Mango Classification using Convolutional Neural Networks
IRJET- Implementation of Gender Detection with Notice Board using Raspberry Pi
IRJET - Steering Wheel Angle Prediction for Self-Driving Cars

Similar to Creating Objects for Metaverse using GANs and Autoencoders (20)

PDF
MODELLING AND SYNTHESIZING OF 3D SHAPE WITH STACKED GENERATIVE ADVERSARIAL NE...
PDF
IRJET- Transformation of Realistic Images and Videos into Cartoon Images and ...
PDF
IRJET- Identification of Scene Images using Convolutional Neural Networks - A...
PDF
IRJET- Handwritten Decimal Image Compression using Deep Stacked Autoencoder
PDF
IRJET- Image based Approach for Indian Fake Note Detection by Dark Channe...
PDF
IMAGE CAPTION GENERATOR USING DEEP LEARNING
PDF
Partial Object Detection in Inclined Weather Conditions
PDF
Unpaired Image Translations Using GANs: A Review
PDF
IRJET- A Study of Generative Adversarial Networks in 3D Modelling
PDF
Face Recognition Based on Image Processing in an Advanced Robotic System
PDF
IRJET- Sketch-Verse: Sketch Image Inversion using DCNN
PDF
IRJET - Wavelet based Image Fusion using FPGA for Biomedical Application
PDF
IRJET- Art Authentication System using Deep Neural Networks
PDF
Car Steering Angle Prediction Using Deep Learning
PDF
IRJET - Gender and Age Prediction using Wideresnet Architecture
PDF
IRJET- American Sign Language Classification
PDF
IRJET- Emotion and Gender Classification in Real-Time
PDF
IRJET- Image Classification – Cat and Dog Images
PDF
IRJET- An Approach to FPGA based Implementation of Image Mosaicing using Neur...
PDF
IRJET- Design and Implementation of ATM Security System using Vibration Senso...
MODELLING AND SYNTHESIZING OF 3D SHAPE WITH STACKED GENERATIVE ADVERSARIAL NE...
IRJET- Transformation of Realistic Images and Videos into Cartoon Images and ...
IRJET- Identification of Scene Images using Convolutional Neural Networks - A...
IRJET- Handwritten Decimal Image Compression using Deep Stacked Autoencoder
IRJET- Image based Approach for Indian Fake Note Detection by Dark Channe...
IMAGE CAPTION GENERATOR USING DEEP LEARNING
Partial Object Detection in Inclined Weather Conditions
Unpaired Image Translations Using GANs: A Review
IRJET- A Study of Generative Adversarial Networks in 3D Modelling
Face Recognition Based on Image Processing in an Advanced Robotic System
IRJET- Sketch-Verse: Sketch Image Inversion using DCNN
IRJET - Wavelet based Image Fusion using FPGA for Biomedical Application
IRJET- Art Authentication System using Deep Neural Networks
Car Steering Angle Prediction Using Deep Learning
IRJET - Gender and Age Prediction using Wideresnet Architecture
IRJET- American Sign Language Classification
IRJET- Emotion and Gender Classification in Real-Time
IRJET- Image Classification – Cat and Dog Images
IRJET- An Approach to FPGA based Implementation of Image Mosaicing using Neur...
IRJET- Design and Implementation of ATM Security System using Vibration Senso...
Ad

More from IRJET Journal (20)

PDF
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
PDF
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
PDF
Kiona – A Smart Society Automation Project
PDF
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
PDF
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
PDF
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
PDF
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
PDF
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
PDF
BRAIN TUMOUR DETECTION AND CLASSIFICATION
PDF
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
PDF
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
PDF
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
PDF
Breast Cancer Detection using Computer Vision
PDF
Auto-Charging E-Vehicle with its battery Management.
PDF
Analysis of high energy charge particle in the Heliosphere
PDF
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
PDF
Auto-Charging E-Vehicle with its battery Management.
PDF
Analysis of high energy charge particle in the Heliosphere
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
Kiona – A Smart Society Automation Project
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
BRAIN TUMOUR DETECTION AND CLASSIFICATION
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
Breast Cancer Detection using Computer Vision
Auto-Charging E-Vehicle with its battery Management.
Analysis of high energy charge particle in the Heliosphere
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
Auto-Charging E-Vehicle with its battery Management.
Analysis of high energy charge particle in the Heliosphere
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
Ad

Recently uploaded (20)

PDF
Automation-in-Manufacturing-Chapter-Introduction.pdf
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PDF
Digital Logic Computer Design lecture notes
PPTX
OOP with Java - Java Introduction (Basics)
PPTX
additive manufacturing of ss316l using mig welding
PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PDF
Well-logging-methods_new................
PDF
composite construction of structures.pdf
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PDF
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
PPTX
web development for engineering and engineering
PPT
introduction to datamining and warehousing
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PPTX
Geodesy 1.pptx...............................................
PDF
Embodied AI: Ushering in the Next Era of Intelligent Systems
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PPTX
Construction Project Organization Group 2.pptx
PDF
Operating System & Kernel Study Guide-1 - converted.pdf
PDF
Unit I ESSENTIAL OF DIGITAL MARKETING.pdf
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
Automation-in-Manufacturing-Chapter-Introduction.pdf
R24 SURVEYING LAB MANUAL for civil enggi
Digital Logic Computer Design lecture notes
OOP with Java - Java Introduction (Basics)
additive manufacturing of ss316l using mig welding
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
Well-logging-methods_new................
composite construction of structures.pdf
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
July 2025 - Top 10 Read Articles in International Journal of Software Enginee...
web development for engineering and engineering
introduction to datamining and warehousing
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
Geodesy 1.pptx...............................................
Embodied AI: Ushering in the Next Era of Intelligent Systems
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
Construction Project Organization Group 2.pptx
Operating System & Kernel Study Guide-1 - converted.pdf
Unit I ESSENTIAL OF DIGITAL MARKETING.pdf
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx

Creating Objects for Metaverse using GANs and Autoencoders

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 06 | June 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1595 Creating Objects for Metaverse using GANs and Autoencoders Aditya Singh1, Tejas Patil1 1Student, Dept. of Computer Engineering, Datta Meghe College of Engineering, Navi Mumbai, Maharashtra, India ---------------------------------------------------------------------***--------------------------------------------------------------------- Abstract – Metaverse refers to online spaces which allow people to interact with one other in a more captivating way than a website. This can be achieved through virtual reality (VR) and augmented reality (AR). To make the Metaverse more connected to reality, we make use of real-life objects that translate into the Metaverse, that's where the need to have state of an art model to achieve this goal. In this paper, we have proposed a way to create such objects using Generative Adversarial Networks (GANs) coupled with an Autoencoder. GANs will create a new object, the object could be anything from animals to human faces, cartoon figures etc. With the creation of such objects, the goal is basically achieved here. However, the output given by the GAN model is not of very high-quality that's where Autoencoder comes in with the use of Autoencoder the object is upscaled (super- resolution). Key Words: Metaverse, Generative Adversarial Networks (GANs), Autoencoder 1. INTRODUCTION Metaverse [1] is like ad immersive world where the user can interact with others, attend a concert etc. which is achieved with AR and VR technologies. In this VR world, we may make our model and react with such different models this paper aims to make this goal more realistic by creating objects based on real-life entities with which these models (people) can react. 1.1 Generative Adversarial Networks (GANs) Generative Adversarial Networks [2] also known as GANs consist of a pair of neural networks namely generator and discriminator that compete with each other. As the name suggests, the generator has to generate images while the discriminator has to detect whether the image generated is real or fake. The discriminator is fed at random real images from the data set it's trained on and images generated by the generator, it has to successfully identify the real image. On the other hand, the generator has to generate a better- quality image such that the discriminator fails to discriminate between the real image and generated image. The basic idea of a GAN model is to create a fake never seen image with help of existing data. The model stops training when the Nash equilibrium is achieved. The following is the general loss function of GAN called Minimax loss there are many other variations of this loss function. Ex[log(D(x))] + Ez[log(1-D(G(z)))] In the above minimax loss equation, D(x) is the discriminator’s estimate of the probability that the present data is real. Ex is the expected value over all real data instances. G(z) is the output given by the generator from noise G. D(G(z)) is the estimate of the probability that the fake instance is real 1.2 Autoencoders Autoencoder [3] is an interesting variant of a Fully connected neural network, It consists of three parts the encoder, the bottleneck and the decoder. The encoder takes the input and its compression is performed in order to store the spatial data in the bottleneck which only consists of three neurons, Therefore the bottleneck stores the spatial data in a lower dimension. this data is further fed into the decoder which is again a fully connected layer and the original image is generated. The bottleneck is forced to learn important information in order to compress the data into lower dimensions such that using the same data the original image can be generated. Autoencoder just like GAN has many variations depending upon the fully connected layer sometimes it's replaced by pooling layers. Depending upon the variation of Autoencoder the loss function may vary, mentioned below is general a loss function: min E(A, B) = min E(xt) = min ∆(A ◦ B(xt), xt) 2. PROPOSED WORK In this paper, we propose a model which will generate real- world objects including human faces with help of a Generative Adversarial Network model and upscale the pixel quality of the generated object with help of Autoencoders, this generated object will be further used in AR/VR application to serve the need of brining real-life feel in Metaverse. For the GAN model, we made use of the Deep Convolutional Generative Adversarial Network (DCGAN) [4] which consists of the deep convolutional layer in the generator and discriminator. In the case of the autoencoder, we made use of both the convolutional layer and pooling layer for both encoder and decoder. the GAN
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 06 | June 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1596 model and Autoencoder model were integrated and the output generated was ready to be fed into the AR/VR application. In the next section, we will look into this process in more detail as we explore the methodology Fig -1: Structure of the Application 3. METHODOLOGY 3.1 Generating images using GANs A custom dataset was created by us which consisted of human faces with more than 5000 images. the images were preprocessed reshaped in 28x28x1 and fed into the generator which consisted of a dense layer and three convolutional layers, Batch normalization [5] was used and leaky ReLU [6] as the activation function. In the case of the discriminator three convolutional were used and in the final layer, the sigmoid activation function was used. For the optimizer, we chose the Adam optimizer to train the discriminator. Fig -1: Architecture of DCGAN Model 3.2 Super-image resolution using Autoencoders The output generated from the DCGAN model is fed into the autoencoder model. The autoencoder model consists of three parts the encoder, the decoder and the bottleneck which learns the latent space representation of the data. For the encoder, we made use of 7 which consisted of convolutional layers, Dropout layers and Pooling layers. Meanwhile, for the decoder, we used 15 layers consisting of convolutional layers, pooling layers and upscaling layers. We made use of ReLU activation and L1 normalization. The training was done in batches. Fig -2: Output of the Autoencoder Model 4. RESULTS 4.1 Result from the DCGAN Model The following are outputs produced by the Deep Convolutional Generative Adversarial Model (DCGAN). Based on the 5000 image data the model was trained on. Fig -1: First result generated by the DCGAN model Fig -2: Second Result generated by the DCGAN model 4.2 Super Image Resolution using Autoencoders After getting outputs from the DCGAN model the results were fed into the Autoencoder model below are super image resolutions of the outputs ready to be fed into AR/VR applications
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 09 Issue: 06 | June 2022 www.irjet.net p-ISSN: 2395-0072 © 2022, IRJET | Impact Factor value: 7.529 | ISO 9001:2008 Certified Journal | Page 1597 Fig -3: First Result generated by the Autoencoder model Fig -4: Second Result generated by the Autoencoder model 5. CONCLUSION In this paper, we have proposed a way to make the metaverse more connected to the real world by including human faces the approach can however be even used to include other objects like cars, animals etc. The Deep Convolutional Generative Adversarial network was responsible for generating the images and Autoencoders was used for image supper resolution to upscale the images. The images generated can be used in AR/VR applications. REFERENCES 1) Stylianos Mystakidis, “Metaverse”, Encyclopedia, https://doi.org/10.3390/encyclopedia2010031 2) Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, Yoshua Bengio, “Generative Adversarial Networks”, arXiv:1406.2661v1 3) Dor Bank, Noam Koenigestein, Raja Giryes, “Autoencoders”, arXiv:2003.05991v2 4) Alec Radford, Luke Metz, Soumith Chintala, “Unsupervised Representation Learning with Deep Convolitional Generative Adversarial Networks”, arXiv:1511.06434v2 5) Segey Ioffe, Christian Szegedy, “Accelerating Deep Network Training by Reducing Internal Covariate Shift”, https://doi.org/10.48550/arXiv.1502.03167 6) Bing Xu, Naiyan Wang, Tianqi Chen, Mu Li, “Empirical Evaluation of Rectified Activation in Convolutional Network”, arXiv:1505.00853v2