Neural Network Toolbox | ![]() ![]() |
Linearly Dependent Vectors
Normally it is a straightforward job to determine whether or not a linear network can solve a problem. Commonly, if a linear network has at least as many degrees of freedom (S*R+S = number of weights and biases) as constraints (Q = pairs of input/target vectors), then the network can solve the problem. This is true except when the input vectors are linearly dependent and they are applied to a network without biases. In this case, as shown with demonstration script demolin6
, the network cannot solve the problem with zero error. You might want to try demolin6
.
Too Large a Learning Rate
A linear network can always be trained with the Widrow-Hoff rule to find the minimum error solution for its weights and biases, as long as the learning rate is small enough. Demonstration script demolin7
shows what happens when a neuron with one input and a bias is trained with a learning rate larger than that recommended by maxlinlr
. The network is trained with two different learning rates to show the results of using too large a learning rate.
![]() | Limitations and Cautions | Summary | ![]() |