Large Language Models: Text Classification for NLP using BERT
.MP4, AVC, 1280x720, 30 fps | English, AAC, 2 Ch | 55m | 136 MB
Instructor: Jonathan Fernandes
.MP4, AVC, 1280x720, 30 fps | English, AAC, 2 Ch | 55m | 136 MB
Instructor: Jonathan Fernandes
Transformers are taking the natural language processing (NLP) world by storm. In this course, instructor Jonathan Fernandes teaches you all about this go-to architecture for NLP and computer vision tasks and must-have skill in your Artificial Intelligence toolkit. Jonathan uses a hands-on approach to show you the basics of working with transformers in NLP and production. He goes over BERT model sizes, bias in BERT, and how BERT was trained. Jonathan explores transfer learning, shows you how to use the BERT model and tokenization, and covers text classification. After thoroughly explaining the transformer model architecture, he finishes up with some additional training runs.
Learning objectives
- Identify important transformer releases and recognize their differences from each other.
- Identify the two datasets BERT was trained on.
- Explain the application differences between encoder-only, decoder-only, and encoder-decoder transformers.
- Describe how BERT begins to tokenize words that are not in its vocabulary.
- Describe how BERT using `checkpoint = 'bert-base-uncased'` handles tokenizing sentences.
- Identify the three vectors used to calculate attention weights.
- Explain how to fine-tune BERT for text classification.