Yıldırım S., Asgari-Chenaghlu M.
Mastering transformers. The Journey from BERT to large
language models and stable diffusion. Second edition
Packt Publishing, 2024
Transformer-based language models such as BERT, T5, GPT, DALL-E, and ChatGPT
have dominated NLP studies and become a new paradigm. Thanks to their accurate
and fast fine-tuning capabilities, transformer-based language models have been able
to outperform traditional machine learning-based approaches for many challenging
natural language understanding (NLU) problems.
Aside from NLP, a fast-growing area in multimodal learning and generative AI has
recently been established, showing promising results. Mastering Transformers will
help you understand and implement multimodal solutions, including text-to-image.
Computer vision solutions that are based on transformers are also explained in the
book. You’ll get started by understanding various transformer models before learning
how to train different autoregressive language models such as GPT and XLNet. The
book will also get you up to speed with boosting model performance, as well as
tracking model training using the TensorBoard toolkit. In the later chapters, you’ll
focus on using vision transformers to solve computer vision problems. Finally, you’ll
discover how to harness the power of transformers to model time series data and for
predicting.
By the end of this transformers book, you’ll have an understanding of transformer
models and how to use them to solve challenges in NLP and CV.
What you will learn:
Focus on solving simple-to-complex NLP problems with Python
Discover how to solve classification/regression problems with traditional NLP
approaches
Train a language model and explore how to fine-tune models to the downstream
tasks
Understand how to use transformers for generative AI and computer vision
tasks
Build transformer-based NLP apps with the Python transformers library
Focus on language generation such as machine translation and conversational
AI in any language
Speed up transformer model inference to reduce latency
Who this book is for:
This book is for deep learning researchers, hands-on practitioners, and ML/NLP
researchers. Educators, as well as students who have a good command of
programming subjects, knowledge in the field of machine learning and artificial
intelligence, and who want to develop apps in the field of NLP as well as multimodal
tasks will also benefit from this book’s hands-on approach. Knowledge of Python (or
any programming language) and machine learning literature, as well as a basic
understanding of computer science, are required.
Source: amazon.com