The document summarizes research on improving the training of multilayer perceptron (MLP) neural networks. It proposes using multiple optimal learning factors (MOLF) during training, which is shown to be equivalent to optimally transforming the net function vector in the MLP. For large networks, the MOLF Hessian matrix can become large, so the paper develops a method to compress the matrix into a smaller, well-conditioned form. Simulation results show the proposed algorithm performs almost as well as Levenberg-Marquardt but with the computational complexity of a first-order method.