This document provides an overview of information theory and coding concepts including:
1) Definitions of information, entropy, joint entropy, conditional entropy, and mutual information are introduced along with examples of calculating these quantities for discrete memoryless sources and channels.
2) Shannon's theorem for channel capacity is discussed and the channel capacity of a discrete memoryless channel is defined as the maximum mutual information over all possible input distributions.
3) Properties of entropy such as it being a measure of uncertainty, having a minimum of 0 and maximum of log2K, and being maximized when probabilities are equal are proven.