Artificial Neurons and Weighted Inputs

Choose a study mode

Play Quiz
Study Flashcards
Spaced Repetition
Chat to Lesson

Podcast

Play an AI-generated podcast conversation about this lesson
Download our mobile app to listen on the go
Get App

Questions and Answers

Why are different weights assigned to the inputs in an artificial neuron?

  • To emphasize the relative importance of the different inputs (correct)
  • To ensure all inputs are equally important in calculating the output
  • To strengthen the less important inputs
  • To weaken the more important inputs

What is the mathematical form of the operation performed by an artificial neuron?

  • $x_1w_1 + x_2w_2 + x_3w_3$
  • $x_1w_1 + x_2w_2 + x_3w_3 + b$
  • $f(x_1w_1 + x_2w_2 + x_3w_3 + b)$ (correct)
  • All of the above

What is the key difference between an artificial neuron and linear regression?

  • Artificial neurons use weights and bias, while linear regression uses only coefficients
  • Artificial neurons can model non-linear relationships, while linear regression can only model linear relationships
  • Artificial neurons apply an activation function, while linear regression does not
  • Both b and c (correct)

How are artificial neurons arranged in an Artificial Neural Network (ANN)?

<p>In multiple layers, with connections only between adjacent layers (D)</p>
Signup and view all the answers

What is the role of the bias term in an artificial neuron?

<p>It is used to shift the activation function to the left or right (D)</p>
Signup and view all the answers

What is the purpose of the activation function in an artificial neuron?

<p>All of the above (D)</p>
Signup and view all the answers

What determines the number of neurons in the output layer in a neural network?

<p>The number and type of problem to be solved (C)</p>
Signup and view all the answers

In binary classification, how many neurons are typically present in the output layer?

<p>One (D)</p>
Signup and view all the answers

What is the purpose of introducing non-linearity in neural networks?

<p>To introduce a nonlinear transformation to learn complex patterns (A)</p>
Signup and view all the answers

Which function scales the input value between 0 and 1 in neural networks?

<p>Sigmoid function (D)</p>
Signup and view all the answers

What is the range of values output by the Tanh function in neural networks?

<p>-1 to 1 (D)</p>
Signup and view all the answers

In a regression problem, how many neurons are typically found in the output layer of a neural network?

<p>One (A)</p>
Signup and view all the answers

What is the function of a synapse in a neuron?

<p>It is where information is transmitted between neurons. (B)</p>
Signup and view all the answers

How are inputs to a neuron weighted before being summed in the cell body?

<p>They are strengthened or weakened based on their importance. (C)</p>
Signup and view all the answers

What is the role of the soma in a neuron?

<p>Processing summed inputs and sending them through axons. (A)</p>
Signup and view all the answers

In artificial neurons, what happens to the inputs received before they are processed?

<p>They are weighted based on their importance. (C)</p>
Signup and view all the answers

Which of the following best describes the activation function of a neuron?

<p>It introduces non-linearity into the network. (B)</p>
Signup and view all the answers

What distinguishes neurons in artificial neural networks from linear regression models?

<p>Neurons in artificial neural networks can handle non-linear relationships, unlike linear regression models. (B)</p>
Signup and view all the answers

Artificial neurons are arranged by layers.

<p>True (A)</p>
Signup and view all the answers

Nuerons in the same layer do not have any connections.

<p>True (A)</p>
Signup and view all the answers

which one is not a typical ANN layers?

<p>Forward Layer (C)</p>
Signup and view all the answers

The number of neurons in the input layer is the number of inputs we feed to the network

<p>True (A)</p>
Signup and view all the answers

There is no computaion in the input layer. it is just used for passing information from the outside world to the network.

<p>True (A)</p>
Signup and view all the answers

Any layer between the inpout layer and the output layer is called ___________.

<p>Hidden Layer</p>
Signup and view all the answers

The input layer identifies the pattern in the dataset.

<p>False (B)</p>
Signup and view all the answers

__________ identifies the pattern in dataset and is responsible for deriving complex relationships between input and output.

<p>Hidden Layer</p>
Signup and view all the answers

The network is called a ___________________ when we have many hidden layers.

<p>Deep Nueral Network (DNN)</p>
Signup and view all the answers

The number of neurons in the output layer is based on the number and the types of the problems to be solved.

<p>True (A)</p>
Signup and view all the answers

_______________ is used to introduce non-linearity in neural networks.

<p>Activation Function</p>
Signup and view all the answers

The aim of the activation function is to introduce a nonlinear transformation to learn the complex underlying patterns in data.

<p>True (A)</p>
Signup and view all the answers

The _____________ scales the input values between 0 and 1.

<p>Sigmoid Function</p>
Signup and view all the answers

Sigmoid Function = 1 / (1+ exp(-x))

<p>True (A)</p>
Signup and view all the answers

The ____________ outputs the values between -1 and +1

<p>Hyperbolic tangent (tanh) Function</p>
Signup and view all the answers

____________ outputs the values from 0 yo infinity.

<p>ReLU (Rectified Linear Unit) Function</p>
Signup and view all the answers

the sang for being zero for all negatives values is a problem called ________ and a neuron is said to be dead if it always outputs zero.

<p>dying ReLU</p>
Signup and view all the answers

_____________ is a variant of the ReLU function that solves the dying ReLU problem. The value of alpha typically se to 0.01

<p>Leaky ,ReLU Function</p>
Signup and view all the answers

ELU (Exponential Linear Unit) is like Leaky ReLU has a small slope for negative values.

<p>True (A)</p>
Signup and view all the answers

__________ is basically the generalization of the sigmoid functions.

<p>The Softmax Function</p>
Signup and view all the answers

The Softmax Functions is usually applied to the final layer of the network and while performing multi-class classification tasks. It gives the probabilities of each class for being output and thus , the sum of softmax values will always equal 1.

<p>True (A)</p>
Signup and view all the answers

softmax function = exp(x) / sum(exp(x)) [0,j]

<p>True (A)</p>
Signup and view all the answers

Flashcards are hidden until you start studying

More Like This

Use Quizgecko on...
Browser
Browser