This document proposes using meta-learning and an LSTM model to learn an optimization algorithm for few-shot learning. The model, called a meta-learner, is trained on multiple datasets to learn how to efficiently train a learner network on new small datasets. The meta-learner LSTM models the parameter updates of the learner network during training, learning an initialization and update rule. The inputs to the meta-learner are the loss, parameters, and gradient, and it outputs updated parameters. This learned update rule can then be used to train the learner network on new small datasets, enabling few-shot learning using only a small amount of labeled data.