4. Deep Learning and Variants_Session 4_20240127 - Artificial Neural Networks Learning Methods
30 Questions
1 Views
3.3 Stars

4. Deep Learning and Variants_Session 4_20240127 - Artificial Neural Networks Learning Methods

Created by
@PalatialRelativity

Questions and Answers

What defines 'Deep' in the context of neural networks?

Having more than two layers in an architecture

In Deep Learning, what is the main benefit of using multiple layers?

Ability to represent complex non-linear functions compactly

Which theorem states that an artificial neural network with a single hidden layer can approximate any continuous function on compact subsets of Rn with infinite neurons?

Universal approximation theorem

Why might visualization become difficult in a shallow network when dealing with unstructured data like an image dataset?

<p>Lack of hidden layers for feature extraction</p> Signup and view all the answers

What is a potential issue with using a very large number of neurons in just one hidden layer of a neural network?

<p>Overfitting</p> Signup and view all the answers

How does having a single hidden layer network with a large number of neurons lead to 0% training error?

<p>By memorizing the training data</p> Signup and view all the answers

Which type of problem can benefit from using deep neural networks instead of shallow networks?

<p>High-dimensional unstructured data problems</p> Signup and view all the answers

What role do hidden layers play in deep learning neural networks?

<p>Extract and transform features hierarchically</p> Signup and view all the answers

What is one of the issues faced when using more layers in deep networks?

<p>Vanishing gradients</p> Signup and view all the answers

Why does a (k−1)-layer network require an exponentially large number of hidden units to represent a function that a k-layer network can represent compactly?

<p>To achieve the same level of representational capacity</p> Signup and view all the answers

What is one of the central problems in Machine Learning according to the text?

<p>Overfitting</p> Signup and view all the answers

Why does gradient descent become very inefficient in deeper networks?

<p>Complex structures in the parameter space being optimized</p> Signup and view all the answers

What happens to the gradients as more hidden layers are added to a deep network?

<p>They vanish and become small relative to the weights</p> Signup and view all the answers

Why does adding more layers to a network not always lead to better performance?

<p>Because vanishing gradients occur</p> Signup and view all the answers

What is a major issue with deep learning highlighted in the text?

<p>Vanishing gradients</p> Signup and view all the answers

What is the derivative of a sigmoid function highlighted as in the text?

<p>Less than 0.3</p> Signup and view all the answers

What percentage of training accuracy is highlighted in the text?

<p>100%</p> Signup and view all the answers

Which issue makes optimization difficult for deep learning networks?

<p>Vanishing gradients</p> Signup and view all the answers

What type of points are mentioned as problematic in high-dimensional spaces?

<p>Saddle points</p> Signup and view all the answers

What combination is described as deadly for neural networks in the text?

<p>Vanishing gradients with overfitting to local minima or saddle points</p> Signup and view all the answers

Why is finding a minimum over the surface difficult in deep learning optimization?

<p>Due to the presence of saddle points and steep gradients</p> Signup and view all the answers

What is a challenge that arises from each sigmoid derivative being less than 0.3?

<p>Vanishing gradients</p> Signup and view all the answers

What is the purpose of back-propagation in Artificial Neural Networks (ANN)?

<p>To compute the sensitivity of the error to the change in weights</p> Signup and view all the answers

What does the term 'Back-propagation' stand for in Artificial Neural Networks?

<p>Backward propagation of Errors</p> Signup and view all the answers

Who popularized the Back-propagation method in Neural Networks?

<p>Geoffery Hinton</p> Signup and view all the answers

What problem can occur due to vanishing gradients in a Neural Network with many layers?

<p>There will be no learning because back-propagation error to initial layers is almost zero</p> Signup and view all the answers

In simple cases, what action should be taken if the output of an Artificial Neural Network is too high?

<p>Decrease the weights attached to high inputs</p> Signup and view all the answers

What does ANN stand for in the context of Artificial Neural Networks?

<p>Artificial Neural Network</p> Signup and view all the answers

'Backpropagation' led to a renaissance in which area?

<p>Neural Networks</p> Signup and view all the answers

'𝑤𝑡+1 𝜕𝐸 = 𝑤𝑡 − 𝛼 𝜕𝑊' represents which concept in Artificial Neural Networks?

<p>Weight adjustment based on error sensitivity</p> Signup and view all the answers

More Quizzes Like This

Use Quizgecko on...
Browser
Browser