The MLP architecture is a layered feedforward neural network, in which the nonlinear elements (neurons) are arranged in successive layers, and the information flows unidirectionally, from input layer to output layer, through the hidden layer(s) (Figure 1). Nodes from one layer are connected (using interconnections or links) to all nodes in the adjacent layer(s), but no lateral connection between nodes within one layer, or feedback connection is possible. The number ofinput and output units depends on the representations of the input and the output objects, respectively. The hidden layer(s) is(are) an important parameter(s) in the network. The MLPs with an arbitrary number of hidden units have been shown to be universal approximators for continuous maps to implement any function.
Was this article helpful?
Do You Want To Learn More About Green Living That Can Save You Money? Discover How To Create A Worm Farm From Scratch! Recycling has caught on with a more people as the years go by. Well, now theres another way to recycle that may seem unconventional at first, but it can save you money down the road.