- Home
- Computers
- Systems Architecture
- Build an Advanced RAG Application (From Scratch)
Build an Advanced RAG Application (From Scratch)
List Price:
$59.99
| Expected release date is Sep 29th 2026 |
- Availability: Confirm prior to ordering
- Branding: minimum 50 pieces (add’l costs below)
- Check Freight Rates (branded products only)
Branding Options (v), Availability & Lead Times
- 1-Color Imprint: $2.00 ea.
- Promo-Page Insert: $2.50 ea. (full-color printed, single-sided page)
- Belly-Band Wrap: $2.50 ea. (full-color printed)
- Set-Up Charge: $45 per decoration
- Availability: Product availability changes daily, so please confirm your quantity is available prior to placing an order.
- Branded Products: allow 10 business days from proof approval for production. Branding options may be limited or unavailable based on product design or cover artwork.
- Unbranded Products: allow 3-5 business days for shipping. All Unbranded items receive FREE ground shipping in the US. Inquire for international shipping.
- RETURNS/CANCELLATIONS: All orders, branded or unbranded, are NON-CANCELLABLE and NON-RETURNABLE once a purchase order has been received.
Product Details
Author:
Hamza Farooq
Format:
Paperback
Pages:
325
Publisher:
Manning (September 29, 2026)
Imprint:
Manning
Release Date:
September 29, 2026
Language:
English
ISBN-13:
9781633436527
ISBN-10:
1633436527
Weight:
13.73oz
Dimensions:
7.375" x 9.25"
File:
Eloquence-SimonSchuster_04272026_P10007149_onix30-20260426.xml
Folder:
Eloquence
List Price:
$59.99
Pub Discount:
37
Series:
From Scratch
As low as:
$56.99
Publisher Identifier:
P-SS
Discount Code:
H
Overview
Get the eBook free when you register your print book at Manning.
Retrieval Augmented Generation—RAG—is now the standard way to improve LLM accuracy and relevance. But building production-grade RAG systems requires far more than connecting an LLM to a vector database. In this book, you’ll learn RAG from first principles by creating a complete portfolio of end-to-end applications. You’ll build each component of the pipeline, ensuring full control over every part of the stack.
Written by former Google research scientist Hamza Farooq, this hands-on guide takes you from LLM and transformer fundamentals through keyword search and semantic retrieval to production RAG systems. You’ll build a hotel search engine with semantic ranking, implement semantic caching for cost-effective production deployments, develop autonomous AI agents powered by RAG context, and deploy optimized open-source LLMs. Through under-the-hood experience, you’ll master embeddings, chunking, reranking, vector databases, evaluation frameworks, fine-tuning, and more.
What's inside
• Design and implement efficient search algorithms for LLM applications
• Master deep customization techniques for every RAG pipeline component
• Model fine-tuning techniques for task-specific and domain adaptation
• Deploy quantized versions of open-source LLMs using vLLMs and Ollama
About the reader
For Python developers with NLP basics, who are ready to move beyond framework abstractions and build RAG systems optimized for their specific constraints.
About the author
Hamza Farooq is the founder and CEO of Traversaal.ai, and he's a seasoned AI expert. His experience includes roles as both a research scientist at Google and a distinguished adjunct professor at leading institutions like Stanford UCLA and University of Minnesota.
Retrieval Augmented Generation—RAG—is now the standard way to improve LLM accuracy and relevance. But building production-grade RAG systems requires far more than connecting an LLM to a vector database. In this book, you’ll learn RAG from first principles by creating a complete portfolio of end-to-end applications. You’ll build each component of the pipeline, ensuring full control over every part of the stack.
Written by former Google research scientist Hamza Farooq, this hands-on guide takes you from LLM and transformer fundamentals through keyword search and semantic retrieval to production RAG systems. You’ll build a hotel search engine with semantic ranking, implement semantic caching for cost-effective production deployments, develop autonomous AI agents powered by RAG context, and deploy optimized open-source LLMs. Through under-the-hood experience, you’ll master embeddings, chunking, reranking, vector databases, evaluation frameworks, fine-tuning, and more.
What's inside
• Design and implement efficient search algorithms for LLM applications
• Master deep customization techniques for every RAG pipeline component
• Model fine-tuning techniques for task-specific and domain adaptation
• Deploy quantized versions of open-source LLMs using vLLMs and Ollama
About the reader
For Python developers with NLP basics, who are ready to move beyond framework abstractions and build RAG systems optimized for their specific constraints.
About the author
Hamza Farooq is the founder and CEO of Traversaal.ai, and he's a seasoned AI expert. His experience includes roles as both a research scientist at Google and a distinguished adjunct professor at leading institutions like Stanford UCLA and University of Minnesota.









