Token, Pattern and Lexeme defines some key concepts in lexical analysis:
Tokens are valid sequences of characters that can be identified as keywords, constants, identifiers, numbers, operators or punctuation. A lexeme is the sequence of characters that matches a token pattern. Patterns are defined by regular expressions or grammar rules to identify lexemes as specific tokens. The lexical analyzer collects attributes like values for number tokens and symbol table entries for identifiers and passes the tokens and attributes to the parser. Lexical errors occur if a character sequence cannot be scanned as a valid token. Error recovery strategies include deleting or inserting characters to allow tokenization to continue.