SFr. 80.00
€ 86.40


bestellen

Artikel-Nr. 42374715


Diesen Artikel in meine
Wunschliste
Diesen Artikel
weiterempfehlen
Diesen Preis
beobachten

Weitersagen:



Autor(en): 
  • Roland Huß
  • Daniele Zonca
  • Generative AI on Kubernetes: Operationalizing Large Language Models 
     

    (Buch)
    Dieser Artikel gilt, aufgrund seiner Grösse, beim Versand als 3 Artikel!


    Übersicht

    Auf mobile öffnen
     
    Lieferstatus:   Auf Bestellung (Lieferzeit unbekannt)
    Veröffentlichung:  März 2026  
    Genre:  EDV / Informatik 
     
    Artificial Intelligence / Artificial Intelligence (AI) / COMPUTERS / Artificial Intelligence / General / COMPUTERS / Distributed Systems / Cloud Computing / Generative artificial intelligence / generative AI / Virtualization
    ISBN:  9781098171926 
    EAN-Code: 
    9781098171926 
    Verlag:  O'Reilly 
    Einband:  Kartoniert  
    Sprache:  English  
    Dimensionen:  H 232 mm / B 178 mm / D 
    Bewertung: Titel bewerten / Meinung schreiben
    Inhalt:

    Generative AI is revolutionizing industries, and Kubernetes has fast become the backbone for deploying and managing these resource-intensive workloads. This book serves as a practical, hands-on guide for MLOps engineers, software developers, Kubernetes administrators, and AI professionals ready to combine AI innovation with the power of cloud native infrastructure. Authors Roland Huss and Daniele Zonca provide a clear road map for training, fine-tuning, deploying, and scaling GenAI models on Kubernetes, addressing challenges like resource optimization, automation, and security along the way.

    With actionable insights with real-world examples, readers will learn to tackle the opportunities and complexities of managing GenAI applications in production environments. Whether you're experimenting with large-scale language models or facing the nuances of AI deployment at scale, you'll uncover expertise you need to operationalize this exciting technology effectively.

    • Learn how to deploy LLMs more efficiently with optimized inference runtimes
    • Get hands-on with GPU scheduling, including hardware detection and multinode scaling
    • Monitor and understand LLM-specific metrics like Time to First Token and token throughput
    • Know when to fine-tune a model or when retrieval augmentation is the better choice
    • Discover how to evaluate models with standardized benchmarks before committing GPU resources
    • Learn to run agentic applications with secure tool integration, identity management, and persistent state

      



    Wird aktuell angeschaut...
     

    Zurück zur letzten Ansicht


    AGB | Datenschutzerklärung | Mein Konto | Impressum | Partnerprogramm
    Newsletter | 1Advd.ch RSS News-Feed Newsfeed | 1Advd.ch Facebook-Page Facebook | 1Advd.ch Twitter-Page Twitter
    Forbidden Planet AG © 1999-2026
    Alle Angaben ohne Gewähr
     
    SUCHEN

     
     Kategorien
    Im Sortiment stöbern
    Genres
    Hörbücher
    Aktionen
     Infos
    Mein Konto
    Warenkorb
    Meine Wunschliste
     Kundenservice
    Recherchedienst
    Fragen / AGB / Kontakt
    Partnerprogramm
    Impressum
    © by Forbidden Planet AG 1999-2026