This book will teach you the underlying concepts of large language models (LLMs), as well as the technologies associated with them. The book starts with an introduction to the rise of conversational AIs such as ChatGPT, and how they are related to the broader spectrum of large language models. From there, you will learn about natural language processing (NLP), its core concepts, and how it has led to the rise of LLMs. Next, you will gain insight into transformers and how their characteristics, such as self-attention, enhance the capabilities of language modeling, along with the unique capabilities of LLMs. The book concludes with an exploration of the architectures of various LLMs and the opportunities presented by their ever-increasing capabilities―as well as the dangers of their misuse. After completing this book, you will have a thorough understanding of LLMs and will be ready to take your first steps in implementing them into your own projects. What You Will Learn Who This Book Is For Anyone interested in learning the foundational concepts of NLP, LLMs, and recent advancements of deep learning
Software Architect at Pearson | AI Practitioner | Inventor | Author
An inventor. A software architect with over 10 years of industry experience. A practitioner and a researcher in AI & machine learning in education and computer vision domains.
Has a Masters in Computer Science with a Bachelors in IT. Has filed 3 patents to date, in the fields of dynamic neural networks and semantics for online learning platforms. Author of 2 books in deep learning & AI.
Quite a good general overview. As someone with a long background in development (but not artificial intelligence) some of the details were hard to follow, but the general information is useful and to the point.
As a primer and jumping off point for further learning, works well.