A perceptron is:
a) a single layer feed-forward neural network with pre-processing
b) an auto-associative neural network
c) a double layer auto-associative neural network
d) a neural network that contains feedback
The network that involves backward links from output to the input and hidden layers is called as ____. a) Self organizing maps b) Perceptrons c) Recurrent neural network d) Multi layered perceptron
What is artificial intelligence neural networks?
Explain Generative Adversarial Network.
How to avoid overflow in the logistic function?
Which of the following is true for neural networks? (i) The training time depends on the size of the network. (ii) Neural networks can be simulated on a conventional computer. (iii) Artificial neurons are identical in operation to biological ones. a) All of the mentioned b) (ii) is true c) (i) and (ii) are true d) None of the mentioned
Which of the following is not the promise of artificial neural network? a) It can explain result b) It can survive the failure of some nodes c) It has inherent parallelism d) It can handle noise
What learning rate should be used for backprop?
How many kinds of nns exist?
How many kinds of kohonen networks exist?
A 3-input neuron is trained to output a zero when the input is 110 and a one when the input is 111. After generalization, the output will be zero when and only when the input is: a) 000 or 110 or 011 or 101 b) 010 or 100 or 110 or 101 c) 000 or 010 or 110 or 100 d) 100 or 111 or 101 or 001
What are the advantages of neural networks over conventional computers? (i) They have the ability to learn by example (ii) They are more fault tolerant (iii)They are more suited for real time operation due to their high ‘computational’ rates a) (i) and (ii) are true b) (i) and (iii) are true c) Only (i) d) All of the mentioned
Having multiple perceptrons can actually solve the XOR problem satisfactorily: this is because each perceptron can partition off a linear part of the space itself, and they can then combine their results. a) True – this works always, and these multiple perceptrons learn to classify even complex problems. b) False – perceptrons are mathematically incapable of solving linearly inseparable functions, no matter what you do c) True – perceptrons can do this but are unable to learn to do it – they have to be explicitly hand-coded d) False – just having a single perceptron is enough
AI Algorithms (74)
AI Natural Language Processing (96)
AI Knowledge Representation Reasoning (12)
AI Robotics (183)
AI Computer Vision (13)
AI Neural Networks (66)
AI Fuzzy Logic (31)
AI Games (8)
AI Languages (141)
AI Tools (11)
AI Machine Learning (659)
Data Science (671)
Data Mining (120)
AI Deep Learning (111)
Generative AI (153)
AI Frameworks Libraries (197)
AI Ethics Safety (100)
AI Applications (427)
AI General (197)
AI AllOther (6)