In this chapter, we will cover the following topics:
- Introducing to Autoencoders
- Convolutional Autoencoder
- Applications of Autoencoders
- An example of compression
In this chapter, we will cover the following topics:
An autoencoder is a regular neural network, an unsupervised learning model that takes an input and produces the same input in the output layer. So, there is no associated label in the training data. Generally, an autoencoder consists of two parts:
It learns all the required features from unlabeled training data, which is known as lower dimensional feature representation. In the following figure, the input data (x) is passed through an encoder that produces a compressed representation of the input data. Mathematically, in the equation, z = h(x), z is a feature vector, and is usually a smaller dimension than x.
Then, we take these produced features from the input data and pass them through a decoder network to reconstruct the original data.
An encoder can be a fully connected neural network or a&...
A convolutional autoencoder is a neural network (a special case of an unsupervised learning model) that is trained to reproduce its input image in the output layer. An image is passed through an encoder, which is a ConvNet that produces a low-dimensional representation of the image. The decoder, which is another sample ConvNet, takes this compressed image and reconstructs the original image.
The encoder is used to compress the data and the decoder is used to reproduce the original image. Therefore, autoencoders may be used for data, compression. Compression logic is data-specific, meaning it is learned from data rather than predefined compression algorithms such as JPEG, MP3, and so on. Other applications of autoencoders can be image denoising (producing a cleaner image from a corrupted image), dimensionality reduction, and image search:
This...
Autoencoders are used for dimensionality reduction, or data compression, and image denoising. Dimensionality reduction, in turn, helps in improving runtime performance and consumes less memory. An image search can become highly efficient in low-dimension spaces.
The Network architecture comprises of an encoder network, which is a typical convolutional pyramid. Each convolutional layer is followed by a max-pooling layer; this reduces the dimensions of the layers.
The decoder converts the input from a sparse representation to a wide reconstructed image. A schematic of the network is shown here:
The encoder layer output image size is 4 x 4 x 8 = 128. The original image size was...
An autoencoder is a regular neural network, an unsupervised learning model that takes an input and produces the same input in the output layer. So, there is no associated label in the training data. Generally, an autoencoder consists of two parts:
It learns all the required features from unlabeled training data, which is known as lower dimensional feature representation. In the following figure, the input data (x) is passed through an encoder that produces a compressed representation of the input data. Mathematically, in the equation, z = h(x), z is a feature vector, and is usually a smaller dimension than x.
Then, we take these produced features from the input data and pass them through a decoder network to reconstruct the original data.
An encoder can be a fully connected neural network or a convolutional neural network (CNN). A decoder also uses the same kind of network as an encoder. Here, we've explained and implemented the encoder...