Target New ArrivalsGift Ideas for MomClothing, Shoes & AccessoriesHome & DecorKitchen & DiningOutdoor Living & GardenGroceryHousehold EssentialsBabyBeautyPersonal CareHealthWellnessLuggageSports & OutdoorsToysElectronicsVideo GamesMovies, Music & BooksSchool & Office SuppliesParty SuppliesGift IdeasGift CardsPetsUlta Beauty at TargetShop by CommunityTarget OpticalDealsClearanceTarget New ArrivalsSpring OutfitsGift Ideas for MomWomen’s Festival OutfitsTop DealsTarget Circle DealsWeekly AdShop Order PickupShop Same Day DeliveryRegistryRedCardTarget CircleFind Stores
Building Large Language Models from Scratch - by  Dilyan Grigorov (Paperback) - 1 of 1

Building Large Language Models from Scratch - by Dilyan Grigorov (Paperback)

$59.99

Pre-order

Free & easy returns

Free & easy returns

Return this item by mail or in store within 90 days for a full refund.
Eligible for registries and wish lists

About this item

Highlights

  • This book is a complete, hands-on guide to designing, training, and deploying your own Large Language Models (LLMs)--from the foundations of tokenization to the advanced stages of fine-tuning and reinforcement learning.
  • About the Author: Dilyan Grigorov is a software developer with a passion for Python software development, generative deep learning & machine learning, data structures, and algorithms.
  • 530 Pages
  • Computers + Internet, Intelligence (AI) & Semantics

Description



Book Synopsis



This book is a complete, hands-on guide to designing, training, and deploying your own Large Language Models (LLMs)--from the foundations of tokenization to the advanced stages of fine-tuning and reinforcement learning. Written for developers, data scientists, and AI practitioners, it bridges core principles and state-of-the-art techniques, offering a rare, transparent look at how modern transformers truly work beneath the surface.

Starting from the essentials, you'll learn how to set up your environment with Python and PyTorch, manage datasets, and implement critical fundamentals such as tensors, embeddings, and gradient descent. You'll then progress through the architectural heart of modern models, covering RMS normalization, rotary positional embeddings (RoPE), scaled dot-product attention, Grouped Query Attention (GQA), Mixture of Experts (MoE), and SwiGLU activations, each explored in depth and built step by step in code. As you advance, the book introduces custom CUDA kernel integration, teaching you how to optimize key components for speed and memory efficiency at the GPU level--an essential skill for scaling real-world LLMs. You'll also gain mastery over the phases of training that define today's leading models:

    Pretraining - Building general linguistic and semantic understanding. Midtraining - Expanding domain-specific capabilities and adaptability. Supervised Fine-Tuning (SFT) - Aligning behavior with curated, task-driven data. Reinforcement Learning from Human Feedback (RLHF) - Refining responses through reward-based optimization for human alignment.
The final chapters guide you through dataset preparation, filtering, deduplication, and training optimization, culminating in model evaluation and real-world prompting with a custom TokenGenerator for text generation and inference.

By the end of this book, you'll have the knowledge and confidence to architect, train, and deploy your own transformer-based models, equipped with both the theoretical depth and practical expertise to innovate in the rapidly evolving world of AI.

What You'll Learn

    How to configure and optimize your development environment using PyTorch The mechanics of tokenization, embeddings, normalization, and attention mechanisms. How to implement transformer components like RMSNorm, RoPE, GQA, MoE, and SwiGLU from scratch. How to integrate custom CUDA kernels to accelerate transformer computations. The full LLM training pipeline: pretraining, midtraining, supervised fine-tuning, and RLHF. Techniques for dataset preparation, deduplication, model debugging, and GPU memory management. How to train, evaluate, and deploy a complete GPT-like architecture for real-world tasks.
Who this book is for:

Software developers, data scientists, machine learning engineers and AI enthusiasts looking to build their models from scratch.



From the Back Cover



This book is a complete, hands-on guide to designing, training, and deploying your own Large Language Models (LLMs)--from the foundations of tokenization to the advanced stages of fine-tuning and reinforcement learning. Written for developers, data scientists, and AI practitioners, it bridges core principles and state-of-the-art techniques, offering a rare, transparent look at how modern transformers truly work beneath the surface.

Starting from the essentials, you'll learn how to set up your environment with Python and PyTorch, manage datasets, and implement critical fundamentals such as tensors, embeddings, and gradient descent. You'll then progress through the architectural heart of modern models, covering RMS normalization, rotary positional embeddings (RoPE), scaled dot-product attention, Grouped Query Attention (GQA), Mixture of Experts (MoE), and SwiGLU activations, each explored in depth and built step by step in code. As you advance, the book introduces custom CUDA kernel integration, teaching you how to optimize key components for speed and memory efficiency at the GPU level--an essential skill for scaling real-world LLMs. You'll also gain mastery over the phases of training that define today's leading models:

  • Pretraining - Building general linguistic and semantic understanding.
  • Midtraining - Expanding domain-specific capabilities and adaptability.
  • Supervised Fine-Tuning (SFT) - Aligning behavior with curated, task-driven data.
  • Reinforcement Learning from Human Feedback (RLHF) - Refining responses through reward-based optimization for human alignment.

The final chapters guide you through dataset preparation, filtering, deduplication, and training optimization, culminating in model evaluation and real-world prompting with a custom TokenGenerator for text generation and inference.

By the end of this book, you'll have the knowledge and confidence to architect, train, and deploy your own transformer-based models, equipped with both the theoretical depth and practical expertise to innovate in the rapidly evolving world of AI.

What You'll Learn

  • How to configure and optimize your development environment using PyTorch
  • The mechanics of tokenization, embeddings, normalization, and attention mechanisms.
  • How to implement transformer components like RMSNorm, RoPE, GQA, MoE, and SwiGLU from scratch.
  • How to integrate custom CUDA kernels to accelerate transformer computations.
  • The full LLM training pipeline: pretraining, midtraining, supervised fine-tuning, and RLHF.
  • Techniques for dataset preparation, deduplication, model debugging, and GPU memory management.
  • How to train, evaluate, and deploy a complete GPT-like architecture for real-world tasks.



About the Author



Dilyan Grigorov is a software developer with a passion for Python software development, generative deep learning & machine learning, data structures, and algorithms. He is an advocate for open source and the Python language itself. He has 16 years of industry experience programming in Python and has spent 5 of those years researching and testing Generative AI solutions. His passion for them stems from his background as an SEO specialist dealing with search engine algorithms daily. He enjoys engaging with the software community, often giving talks at local meetups and larger conferences. In his spare time, he enjoys reading books, hiking in the mountains, taking long walks, playing with his son, and playing the piano.

Dimensions (Overall): 10.0 Inches (H) x 7.01 Inches (W)
Suggested Age: 22 Years and Up
Number of Pages: 530
Genre: Computers + Internet
Sub-Genre: Intelligence (AI) & Semantics
Publisher: Apress
Format: Paperback
Author: Dilyan Grigorov
Language: English
Street Date: May 5, 2026
TCIN: 1008647337
UPC: 9798868822964
Item Number (DPCI): 247-19-3821
Origin: Made in the USA or Imported
If the item details aren’t accurate or complete, we want to know about it.

Shipping details

Estimated ship dimensions: 1 inches length x 7.01 inches width x 10 inches height
Estimated ship weight: 1 pounds
We regret that this item cannot be shipped to PO Boxes.
This item cannot be shipped to the following locations: American Samoa (see also separate entry under AS), Guam (see also separate entry under GU), Northern Mariana Islands, Puerto Rico (see also separate entry under PR), United States Minor Outlying Islands, Virgin Islands, U.S., APO/FPO, Alaska, Hawaii

Return details

This item can be returned to any Target store or Target.com.
This item must be returned within 90 days of the date it was purchased in store, delivered to the guest, delivered by a Shipt shopper, or picked up by the guest.
See the return policy for complete information.

Q: Who is the author of this book?

submitted by AI Shopping Assistant - 1 month ago
  • A: The book is authored by Dilyan Grigorov, a software developer specializing in Python and generative deep learning.

    submitted byAI Shopping Assistant - 1 month ago
    Ai generated

Q: What key concepts are explored in this book?

submitted by AI Shopping Assistant - 1 month ago
  • A: Key concepts include tokenization, embeddings, normalization, attention mechanisms, and transformer architectures.

    submitted byAI Shopping Assistant - 1 month ago
    Ai generated

Q: What is the main focus of the book?

submitted by AI Shopping Assistant - 1 month ago
  • A: The book focuses on designing, training, and deploying large language models, covering both foundational and advanced techniques.

    submitted byAI Shopping Assistant - 1 month ago
    Ai generated

Q: What is the target audience for this book?

submitted by AI Shopping Assistant - 1 month ago
  • A: The book targets software developers, data scientists, machine learning engineers, and AI enthusiasts.

    submitted byAI Shopping Assistant - 1 month ago
    Ai generated

Q: What programming language is emphasized in the book?

submitted by AI Shopping Assistant - 1 month ago
  • A: The book emphasizes Python, especially in the context of developing large language models using PyTorch.

    submitted byAI Shopping Assistant - 1 month ago
    Ai generated

Additional product information and recommendations

Discover more options

Get top deals, latest trends, and more.

Privacy policy