Also called
Feedback Signals.
Feedback describes the use of output signals, which are used as input signals. These are usually used in tandem with other input signals, and cause Turing-indeterminate responses of the systems that employ them.
These do not necessarily describe direct feedback, from the output of a neuron to its input. It may also describe indirect feedback, such as when the output of a given neuron is connected to
pre-synaptic neuron(s) which eventually are connected (possibly through other neurons) back to the original neuron's input.
Finally, and possibly most importantly, the output of a given neuron or set of neurons may produce signals that have an effect on the external environment. The changes in the external environment are sensed by input sensors which, in turn, produce signals that are eventually part of the input to the neurons that produced the initial output.
. . . . . . .
In Neural Networks
Many traditional neural network
learning algorithms restrict network topologies to forward signalling only. That is, they do not allow
feedback in the design of the neural network. This leads to topological layers in these designs, which are sometimes referred to as Multilayer Perceptrons, or
MLPs.
Netlab introduces a new learning algorithm, called
influence learning, which is based on attraction to influence exerted over other neurons during
signal propagation. This algorithm is completely feedback-tolerant.