The document provides an overview of Music Information Retrieval (MIR) techniques for analyzing music with computers. It discusses common MIR tasks like genre/mood classification, beat tracking, and music similarity. Recent approaches to music auto-tagging using deep learning are highlighted, such as using neural networks to learn features directly from audio rather than relying on hand-designed features. Recurrent neural networks are presented as a way to model temporal dependencies in music for applications like onset detection. As an example, the document describes a system for live drum transcription that uses onset detection, spectrogram slicing, and non-negative matrix factorization for source separation to detect drum activations in real-time performance audio.