Podcast Beta
Questions and Answers
An Auto-Encoder consists of an NN Encoder, an NN Decoder, and a reconstruction ability.
True
PCA stands for Principle Component Application.
False
The objective in a Deep Auto-encoder is to maximize the bottleneck layer.
False
In an Encoder-Decoder for NLP, the Decoder generates the contextualized representation.
Signup and view all the answers
Attention Mechanism plays a crucial role in Encoder-Decoder networks by providing flexibility but not context vector importance.
Signup and view all the answers
Cross-Attention is used in neural machine translation to align and translate target sentences.
Signup and view all the answers
Auto-encoder is a supervised learning method that creates a compact representation of input objects.
Signup and view all the answers
The encoder and decoder in an auto-encoder neural network work independently to reconstruct the original input object.
Signup and view all the answers
Deep auto-encoders always require a symmetric layer structure for more complex representations.
Signup and view all the answers
Denoising auto-encoders are used to add noise to input data for better reconstruction.
Signup and view all the answers
Auto-encoders can be applied in Convolutional Neural Networks (CNNs) for tasks like pooling and activation.
Signup and view all the answers
Encoder and Decoder in basic RNN-based Encoder-Decoder Networks have different internal structures.
Signup and view all the answers
The context vector 'c' from the Encoder carries essential input information to the Decoder.
Signup and view all the answers
In Encoder-Decoder Attention models, the Attention Mechanism enables the Decoder to focus only on the final state of the Encoder.
Signup and view all the answers
Attention scores determine the relevance of each Encoder hidden state to the current Decoder state.
Signup and view all the answers
Transformers are a type of neural network architecture that require sequential processing of input data.
Signup and view all the answers
In Transformers, the Multi-head Attention mechanism helps capture relationships between different parts of the input sequence.
Signup and view all the answers
Self-Attention in Transformers is used to find relevant vectors within the output sequence.
Signup and view all the answers