This document presents a benchmark for deep learning algorithms developed by identifying basic operations that account for most CPU usage. Three algorithms were implemented - sparse autoencoder, convolutional neural network, and FISTA optimization. The operations were abstracted into an API for easier optimization. Results showed the Theano GPU implementation was 3-15 times faster than Numpy. Challenges included choosing array dimensions and memory allocation to optimize performance. Convolution was identified as the most expensive operation for CNNs in terms of CPU usage.