Lazy learning methods store training data and wait until test data is received to perform classification, taking less time to train but more time to predict. Eager learning methods construct a classification model during training. Lazy methods like k-nearest neighbors use a richer hypothesis space while eager methods commit to a single hypothesis. The k-nearest neighbor algorithm classifies new examples based on the labels of its k closest training examples. Case-based reasoning uses a symbolic case database for classification while genetic algorithms evolve rule populations through crossover and mutation to classify data.