Member-only story
NLP with Transformers: Introduction — Part 1
13 min readJun 27, 2023
The first section of the thorough summary of the book Natural Language Processing with Transformers.
If you can afford to buy and read the book [1,] I strongly advise you to do so rather than reading my notes. But if you can’t… let’s get started :)
This series will include 11 sections, to teach you:
- Learn how to build, debug, and optimize transformer models for core NLP tasks such as text categorization, named entity identification, and question answering.
- Learn how to use transformers for cross-lingual transfer learning.
In real-world circumstances where labeled data is rare, use transformers. - Make transformer models more deployable by employing approaches like distillation, pruning, and quantization.
- Learn how to train transformers from the ground up and scale them across many GPUs and distributed environments.
The Transformer architecture is so good at capturing patterns in long sequences of data and coping with large datasets that it is already being used for applications other than NLP, such as image processing.