The era of cloud-dependent AI is over. Today's developers can run state-of-the-art language models on their own hardware-from laptops to GPU clusters-without ever sending data to a third party. But the gap between downloading a model and deploying it efficiently is filled with questions about quantization, memory bandwidth, batching strategies, and tool selection. This book is your guide through that gap, showing you how to build scalable, cost-effective inference systems using the three pillars of open-source AI: Ollama, llama.cpp, and vLLM. AI Inference with Ollama, llama.cpp, and vLLM takes you from running your first local model in minutes to optimizing production deployments serving thousands of requests per second. You'll learn when to use each tool, how to navigate the memory wall that bottlenecks LLM performance, and how to choose the right hardware and quantization strategy for your use case. Whether you're building RAG systems, deploying chatbots, or scaling inference across GPU clusters, this book gives you the practical knowledge to move from experimentation to production with confidence. About the Author GK Marballi has spent 20+ years turning data into competitive advantage for global brands from Priceline to S&P Global and Barnes & Noble. He has led high-impact product and analytics teams, and navigated the front lines of the AI revolution. He is based in New York City and holds an MBA from Harvard Business School.
ThriftBooks sells millions of used books at the lowest
everyday prices. We personally assess every book's quality and offer rare, out-of-print treasures. We
deliver the joy of reading in recyclable packaging with free standard shipping on US orders over $15.
ThriftBooks.com. Read more. Spend less.