English | 2024 | ISBN: 979-8836693312 | 229 Pages | PDF | 10 MB
This book is a concise and illustrated guide for anyone who wants to understand the inner workings of large language models in the context of interviews, projects or to satisfy their own curiosity.
It is divided into 5 parts:
- Foundations: primer on neural networks and important deep learning concepts for training and evaluation
- Embeddings: tokenization algorithms, word-embeddings (word2vec) and sentence embeddings (RNN, LSTM, GRU)
- Transformers: motivation behind its self-attention mechanism, detailed overview on the encoder-decoder architecture and related variations such as BERT, GPT and T5, along with tips and tricks on how to speed up computations
- Large language models: main techniques to tune Transformer-based models, such as prompt engineering, (parameter efficient) finetuning and preference tuning
- Applications: most common problems including sentiment extraction, machine translation, retrieval-augmented generation and many more
Download from free file storage
Resolve the captcha to access the links!