1⟩ Explain Artificial Neural Networks (ANNs)?
A computing system made up of a number of simple, highly interconnected processing elements, which process information by their dynamic state response to external inputs.
“Artificial Intelligence Neural Networks frequently Asked Questions in various AI Neural Networks job Interviews by interviewer. Get preparation of Artificial Intelligence Neural Networks job interview”
A computing system made up of a number of simple, highly interconnected processing elements, which process information by their dynamic state response to external inputs.
A neural network can be defined as a model of reasioning
based on the human brain.
The human brain incorporates nearly 10 billion neurons and
60 trillion connections,Synapses,between them.
By using multiple neurons simultaneously,the brain can
perform its functions much faster than the faster computers.
Although a single neuron has a very simple structure,an
army of such elements constitutes a tremendous processing
power.
The network which represents the connections among several
neurons is called a neural network.
Explain your self
For the sake of trying to produce intelligent behavior however really all that's being done is work with artificial neural networks where each cell is a very simple processor and the goal is to try and make them work together to solve some problem. That's all that gets covered in this book. Many people are skeptical that artificial neural networks can produce human levels of performance because they are so much simpler than the biological neural networks.
a) True
Explanation:
Yes the perceptron works like that.
d) All of the mentioned
a) It can explain result
Explanation:
The artificial Neural Network (ANN) cannot explain result.
c) (i) and (ii) are true
Explanation:
The training time depends on the size of the network; the number of neuron is greater and therefore the number of possible 'states' is increased. Neural networks can be simulated on a conventional computer but the main advantage of neural networks - parallel execution - is lost. Artificial neurons are not identical in operation to the biological ones.
b) Because they are the only class of problem that Perceptron can solve successfully
Explanation:
Linearly separable problems of interest of neural network researchers because they are the only class of problem that Perceptron can solve successfully
c) It is the transmission of error back through the network to allow weights to be adjusted so that the network can learn.
Explanation:
Back propagation is the transmission of error back through the network to allow weights to be adjusted so that the network can learn.
d) All of the mentioned
Explanation:
Neural networks learn by example. They are more fault tolerant because they are always able to respond and small changes in input do not normally cause a change in output. Because of their parallel architecture, high computational rates are achieved.
d) Because it is the simplest linearly inseparable problem that exists.
b) a neural network that contains feedback
Explanation:
An auto-associative network is equivalent to a neural network that contains feedback. The number of feedback paths(loops) does not have to be one.
a) 238
Explanation:
The output is found by multiplying the weights with their respective inputs, summing the results and multiplying with the transfer function. Therefore:
Output = 2 * (1*4 + 2*10 + 3*5 + 4*20) = 238.
a) All of the mentioned are true
Explanation:
Neural networks have higher computational rates than conventional computers because a lot of the operation is done in parallel. That is not the case when the neural network is simulated on a computer. The idea behind neural nets is based on the way the human brain works. Neural nets cannot be programmed, they cam only learn by examples.
a) a single layer feed-forward neural network with pre-processing
Explanation:
The perceptron is a single layer feed-forward neural network. It is not an auto-associative network because it has no feedback and is not a multiple layer neural network because the pre-processing stage is not made of neurons.
c) True - perceptrons can do this but are unable to learn to do it - they have to be explicitly hand-coded
c) 000 or 010 or 110 or 100
Explanation:
The truth table before generalization is:
Inputs Output
000 $
001 $
010 $
011 $
100 $
101 $
110 0
111 1
where $ represents don't know cases and the output is random.
After generalization, the truth table becomes:
Inputs Output
000 0
001 1
010 0
011 1
100 0
101 1
110 0
111 1
b) Heaviside function
Explanation:
Also known as the step function - so answer 1 is also right. It is a hard thresholding function, either on or off with no in-between.
d) All of the mentioned
Explanation:
All mentioned are the characteristics of neural network.