My Translation "Illustrated Large Models - Principles and Practice of Generative AI" is Coming Soon
My translation “Illustrated Large Models - Principles and Practice of Generative AI” (Hands-On Large Language Models) has finally gone to print and will be available in mid-May.
Praise for the Book (Chinese Edition)
Many thanks to Yuan Jinhui, founder of Silicon-Based Mobility, Zhou Lidong, director of Microsoft Research Asia, Lin Junyang, head of algorithms at Alibaba Qwen, Li Guohao, founder of CAMEL-AI.org community, and Zhong Tai, founder of AgentUniverse, for their strong recommendations!
Translator’s Preface
The development of large models is rapid, as the saying goes, “One day in AI, one year in the human world.” Many people are lost in the flourishing garden of models, unsure of which model to use for their application scenarios and unable to predict the development direction of models in the coming year, often feeling anxious. In fact, almost all large models today are based on the Transformer architecture, remaining fundamentally unchanged.
The book “Illustrated Large Models” is an excellent resource to help you systematically understand the basic principles and capability boundaries of Transformers and large models. When Turing Company approached me to translate this book, I immediately agreed upon seeing the author’s name, as it was Jay Alammar’s blog post “The Illustrated Transformer” that truly helped me understand Transformers (Chapter 3 of this book is an expansion of that blog post). Although there are countless books and articles explaining large models on the market, the exquisite illustrations and the depth and simplicity of the explanations in this book are rare. The book starts with tokens and embeddings, not limited to generative models, but also includes representation models that many overlook. Additionally, the book covers practical content such as text classification, text clustering, prompt engineering, RAG, and model fine-tuning.
I am very honored to be the translator of this book, working with editor Liu Meiying to bring this book to Chinese readers.
Take some time to read this book and systematically understand the basic principles and capability boundaries of Transformers and large models, just like having a map and compass on an adventure in the world of large models. This way, we won’t worry about new models rendering long-term engineering accumulation useless overnight, and we can develop products for future models. Once the model capabilities are ready, the product can scale up immediately.
I hope this book can become a sightseeing bus in the garden of large models, allowing more people to see the panorama of large models. Thus, the ever-expanding capability boundaries of large models become a visual feast rather than a monster devouring everything; we have the opportunity to stand at the forefront of AI, realize more dreams, and gain more freedom.