From the course: Introduction to Transformer Models for NLP

Unlock this course with a free trial

Join today to access over 24,800 courses taught by industry experts.

How language models look at text

How language models look at text

- Section 1.4, "How Language Models Look At Text." Back in the beginning, I mentioned how in 2001, a language modeling task was solved for the first time using a deep learning feedforward architecture, but I didn't really talk about what a language modeling task is. Now, it's actually important we take a look at that because language models and language modeling tasks are the core tenant of how transformers and transformer-based architectures learn language, and more specifically, language rules, how words are used in sentences, how sentences are treated in a larger corpora. A language modeling task, a model is trained to predict a missing word, or token, in a sequence of words, or tokens. Now, in general, there's two kinds of language models, auto-regressive and auto-encoding. Consider the following example, "If you don't," blank, "at the sign, you will get a ticket." Now, if you're watching this at home, you're probably already filling in the blank yourself, but this is actually a…

Contents