Muutke küpsiste eelistusi

Operational AI with Docker: Deploy, Scale and Operate Agentic AI services with Docker and Kubernetes [Pehme köide]

  • Formaat: Paperback / softback, kõrgus x laius: 235x191 mm
  • Ilmumisaeg: 29-Apr-2026
  • Kirjastus: Packt Publishing Limited
  • ISBN-10: 1807301095
  • ISBN-13: 9781807301095
Teised raamatud teemal:
  • Pehme köide
  • Hind: 63,89 €
  • See raamat ei ole veel ilmunud. Raamatu kohalejõudmiseks kulub orienteeruvalt 3-4 nädalat peale raamatu väljaandmist.
  • Kogus:
  • Lisa ostukorvi
  • Tasuta tarne
  • Tellimisaeg 2-4 nädalat
  • Lisa soovinimekirja
  • Formaat: Paperback / softback, kõrgus x laius: 235x191 mm
  • Ilmumisaeg: 29-Apr-2026
  • Kirjastus: Packt Publishing Limited
  • ISBN-10: 1807301095
  • ISBN-13: 9781807301095
Teised raamatud teemal:
Run production-grade GenAI workloads by containerizing, serving, and scaling LLMs, agents, and multi-model pipelines with Docker, MCP, and Kubernetes for cloud platforms

Key Features

Deploy and operate local and edge-friendly LLM inference using Docker Model Runner and an OpenAI-compatible API Orchestrate multi-model and multi-agent workloads with Docker Compose and Kubernetes patterns used by platform teams Purchase of the print or Kindle book includes a free PDF eBook

Book DescriptionThe book blends hands-on Docker and Kubernetes fundamentals with practical AI/ML deployment workflows, showing not just how containers work but why they are essential for modern machine learning pipelines. Its unique selling point (USP) is that it focuses on Docker as the central model runner demonstrating end-to-end examples from building and containerizing models to orchestrating large-scale inference and training workloads, all with real, reproducible demos. What you will learn

Containerize GenAI services using Docker images, registries, and Compose-based deployment stacks Package and distribute models as OCI artifacts for repeatable builds and controlled promotions across environments Choose GGUF quantization levels to balance cost, latency, and accuracy for cloud and hybrid runtimes Serve LLMs via Docker Model Runner with an OpenAI-compatible API suitable for internal platforms Integrate tools and data securely using MCP and Docker MCP Gateway with least-privilege access patterns

Who this book is forCloud engineers, DevOps engineers, SREs, and platform engineers who need to deploy, operate, and scale GenAI workloads using Docker and Kubernetes on cloud, hybrid, or edge environments. You should be comfortable with the command line and basic service operations; prior Docker or Kubernetes exposure is helpful but not required.
Table of Contents

Introduction to Containerisation for AI
Understanding AI Models in Docker
Model Serving with Docker Model Runner
Docker Offload
Running ML container models on Kubernetes
Protocol-based AI Integration with MCP
Building Autonomous AI Agents
Multi-Model and Multi-Agent Architectures
Advanced Agents Orchestration
Ajeet Singh Raina is a Developer Advocate at Docker and an early Docker adopter who has authored 600+ blogs on containerization, cloud-native technologies, and DevOps. He leads a large Docker community ecosystem and organizes initiatives such as Kubetools, sharing practical guidance across Docker, Kubernetes, IoT, and AI/ML operations. Harsh Manvar is a Senior Software Engineer with over a decade of experience in softwareengineering and DevOps. A Docker Captain, Google Developer Expert, CNCF Ambassador, and Google Champion Innovator, he focuses on building scalable, reliable cloud-native systems and is a top contributor in the Kubernetes space on Stack Overflow.