This document covers the principles of vector semantics and word embeddings in natural language processing (NLP). It discusses the distributional hypothesis, which posits that words appearing in similar contexts have similar meanings, and explores various methodologies for capturing word semantics using vector representations. Additionally, it highlights the importance of co-occurrence statistics and different ways to measure word similarity through vector distances and positive pointwise mutual information.