A computing system made up of a number of simple, highly interconnected processing elements, which process information by their dynamic state response to external inputs.
AI Neural Networks
Home Artificial intelligence (AI) AI Neural Networks
“Artificial Intelligence Neural Networks frequently Asked Questions in various AI Neural Networks job Interviews by interviewer. Get preparation of Artificial Intelligence Neural Networks job interview”
26 AI Neural Networks Questions And Answers
A neural network can be defined as a model of reasioning
based on the human brain.
The human brain incorporates nearly 10 billion neurons and
60 trillion connections,Synapses,between them.
By using multiple neurons simultaneously,the brain can
perform its functions much faster than the faster computers.
Although a single neuron has a very simple structure,an
army of such elements constitutes a tremendous processing
The network which represents the connections among several
neurons is called a neural network.
3⟩ Suppose we need a MATLAB source code to recognize different regular geometric shapessuch as squares,rectangles,triangles,circles and ellipses in different sizes using neural network.All of the images containing these shapes should be in binary format with the size of 300*400 pixels.Give us a MATLAB code to detect these geometric shapes?
Explain your self
For the sake of trying to produce intelligent behavior however really all that's being done is work with artificial neural networks where each cell is a very simple processor and the goal is to try and make them work together to solve some problem. That's all that gets covered in this book. Many people are skeptical that artificial neural networks can produce human levels of performance because they are so much simpler than the biological neural networks.
5⟩ A perceptron adds up all the weighted inputs it receives, and if it exceeds a certain value, it outputs a 1, otherwise it just outputs a 0. a) True b) False c) Sometimes - it can also output intermediate values as well d) Can't say
Yes the perceptron works like that.
6⟩ Which of the following is an application of NN (Neural Network)? a) Sales forecasting b) Data validation c) Risk management d) All of the mentioned
d) All of the mentioned
7⟩ Which of the following is not the promise of artificial neural network? a) It can explain result b) It can survive the failure of some nodes c) It has inherent parallelism d) It can handle noise
a) It can explain result
The artificial Neural Network (ANN) cannot explain result.
8⟩ Which of the following is true for neural networks? (i) The training time depends on the size of the network. (ii) Neural networks can be simulated on a conventional computer. (iii) Artificial neurons are identical in operation to biological ones. a) All of the mentioned b) (ii) is true c) (i) and (ii) are true d) None of the mentioned
c) (i) and (ii) are true
The training time depends on the size of the network; the number of neuron is greater and therefore the number of possible 'states' is increased. Neural networks can be simulated on a conventional computer but the main advantage of neural networks - parallel execution - is lost. Artificial neurons are not identical in operation to the biological ones.
9⟩ Why are linearly separable problems of interest of neural network researchers? a) Because they are the only class of problem that network can solve successfully b) Because they are the only class of problem that Perceptron can solve successfully c) Because they are the only mathematical functions that are continue d) Because they are the only mathematical functions you can draw
b) Because they are the only class of problem that Perceptron can solve successfully
Linearly separable problems of interest of neural network researchers because they are the only class of problem that Perceptron can solve successfully
10⟩ What is back propagation? a) It is another name given to the curvy function in the perceptron b) It is the transmission of error back through the network to adjust the inputs c) It is the transmission of error back through the network to allow weights to be adjusted so that the network can learn. d) None of the mentioned
c) It is the transmission of error back through the network to allow weights to be adjusted so that the network can learn.
Back propagation is the transmission of error back through the network to allow weights to be adjusted so that the network can learn.
11⟩ What are the advantages of neural networks over conventional computers? (i) They have the ability to learn by example (ii) They are more fault tolerant (iii)They are more suited for real time operation due to their high 'computational' rates a) (i) and (ii) are true b) (i) and (iii) are true c) Only (i) d) All of the mentioned
d) All of the mentioned
Neural networks learn by example. They are more fault tolerant because they are always able to respond and small changes in input do not normally cause a change in output. Because of their parallel architecture, high computational rates are achieved.
12⟩ Why is the XOR problem exceptionally interesting to neural network researchers? a) Because it can be expressed in a way that allows you to use a neural network b) Because it is complex binary operation that cannot be solved using neural networks c) Because it can be solved by a single layer perceptron d) Because it is the simplest linearly inseparable problem that exists.
d) Because it is the simplest linearly inseparable problem that exists.
13⟩ An auto-associative network is a) a neural network that contains no loops b) a neural network that contains feedback c) a neural network that has only one loop d) a single layer feed-forward neural network with pre-processing
b) a neural network that contains feedback
An auto-associative network is equivalent to a neural network that contains feedback. The number of feedback paths(loops) does not have to be one.
14⟩ A 4-input neuron has weights 1, 2, 3 and 4. The transfer function is linear with the constant of proportionality being equal to 2. The inputs are 4, 10, 5 and 20 respectively. The output will be a) 238 b) 76 c) 119 d) 123
The output is found by multiplying the weights with their respective inputs, summing the results and multiplying with the transfer function. Therefore:
Output = 2 * (1*4 + 2*10 + 3*5 + 4*20) = 238.
15⟩ Which of the following is true? (i) On average, neural networks have higher computational rates than conventional computers. (ii) Neural networks learn by example. (iii) Neural networks mimic the way the human brain works. a) All of the mentioned are true b) (ii) and (iii) are true c) (i), (ii) and (iii) are true d) None of the mentioned
a) All of the mentioned are true
Neural networks have higher computational rates than conventional computers because a lot of the operation is done in parallel. That is not the case when the neural network is simulated on a computer. The idea behind neural nets is based on the way the human brain works. Neural nets cannot be programmed, they cam only learn by examples.
16⟩ A perceptron is a) a single layer feed-forward neural network with pre-processing b) an auto-associative neural network c) a double layer auto-associative neural network d) a neural network that contains feedback
a) a single layer feed-forward neural network with pre-processing
The perceptron is a single layer feed-forward neural network. It is not an auto-associative network because it has no feedback and is not a multiple layer neural network because the pre-processing stage is not made of neurons.
17⟩ Having multiple perceptrons can actually solve the XOR problem satisfactorily this is because each perceptron can partition off a linear part of the space itself, and they can then combine their results. a) True - this works always, and these multiple perceptrons learn to classify even complex problems. b) False - perceptrons are mathematically incapable of solving linearly inseparable functions, no matter what you do c) True - perceptrons can do this but are unable to learn to do it - they have to be explicitly hand-coded d) False - just having a single perceptron is enough
c) True - perceptrons can do this but are unable to learn to do it - they have to be explicitly hand-coded
18⟩ A 3-input neuron is trained to output a zero when the input is 110 and a one when the input is 111. After generalization, the output will be zero when and only when the input is a) 000 or 110 or 011 or 101 b) 010 or 100 or 110 or 101 c) 000 or 010 or 110 or 100 d) 100 or 111 or 101 or 001
c) 000 or 010 or 110 or 100
The truth table before generalization is:
where $ represents don't know cases and the output is random.
After generalization, the truth table becomes:
19⟩ The name for the function in question 16 is a) Step function b) Heaviside function c) Logistic function d) Perceptron function
b) Heaviside function
Also known as the step function - so answer 1 is also right. It is a hard thresholding function, either on or off with no in-between.
20⟩ Which is true for neural networks? a) It has set of nodes and connections b) Each node computes it's weighted input c) Node could be in excited state or non-excited state d) All of the mentioned
d) All of the mentioned
All mentioned are the characteristics of neural network.