The document compares CBOW, Skip-gram, and Skip-gram with subword information regarding their training data, model settings, and performance on tasks such as related concepts and similarity. It highlights the limitations of plain word embeddings in capturing fine-grained semantic similarities and their dependency on training data quality. Visual comparisons and metrics such as precision, recall, and F-score are also discussed.