Time Delay Neural Networks

Time-delay is frequently observed in ecological processes and is efficiently accommodated in temporal networks. A typical application of time-delay neural networks (TDNNs) is speech recognition. In learning the time series data, the time-delayed output could be reused as input. To pattern relationships between different time events of variable changes, initially a simple multilayer perceptron with a backpropagation algorithm is used as a nonlinear predictor (see Multilayer Perceptron). The architecture consists of well-known static multilayers; however, input and output data were provided using a time delay. The input vector is defined in terms of past samples, X(t — 1), X(t — 2), ..., X(t — q), where q, prediction order, is the number oftotal delays. The current data, X(t), is given as matching output. With each delay, input nodes are correspondingly added.

The input layer is subsequently interconnected to the hidden layer. The internal state of the network, NETpj is obtained by linear summation of products of weights and output values of nodes in the hidden layer over time. Subsequently, these values are adjusted in a nonlinear fashion, logistic function in this case, to produce the outputs, Y(t)pj as follows:


where Yp j is activation of neuron j for pattern p xpj is output value of the neuron i of the previous layer for pattern pwpj is weight of the connection between the neuron i of the previous layer and the neuron j of the current layer for pattern p; and A is activation function coefficient.

The output Y(t) of the multilayer perceptron is produced in response to the input vector, and is equivalent to the one-step prediction for the future development (see Multilayer Perceptron). Actual data at time t, X(t), are provided as the target. Subsequently, the difference between Y(t) and X(t) is measured and propagated backward to adjust weights in the usual manner of the backpropagation algorithm. Weights at output neurons are updated as follows:

Awfj, (t + 1) — r\bfj Yfj + aAwfj, (t) Wfji (t + 1) — Wfji (t) + AWfji (t + 1)

where dpj is the desired output of node j for pattern p, q is the training rate coefficient, and a is the momentum coefficient. Weight updating for the hidden layers is similar to the processes occurring at the neurons in the output layer.

Was this article helpful?

0 0
10 Ways To Fight Off Cancer

10 Ways To Fight Off Cancer

Learning About 10 Ways Fight Off Cancer Can Have Amazing Benefits For Your Life The Best Tips On How To Keep This Killer At Bay Discovering that you or a loved one has cancer can be utterly terrifying. All the same, once you comprehend the causes of cancer and learn how to reverse those causes, you or your loved one may have more than a fighting chance of beating out cancer.

Get My Free Ebook

Post a comment