Why and how to run NVIDIA NIM on Amazon EKS
Автор: AWS Events
Загружено: 2024-10-15
Просмотров: 1496
Описание:
Discover how to efficiently deploy and scale large language models like Llama3/Mistral7b on Kubernetes using NVIDIA Inference Microservices (NIM) on Amazon EKS. This video guides you through:
• Setting up GPU-ready EKS clusters
• Scaling with Kubernetes
• Leveraging NVIDIA's NIM Helm chart
• Real-time benchmarking with GenAIPerf
• Monitoring costs and performance
Perfect for ML engineers and cloud architects looking to optimize their AI infrastructure on AWS. Watch our live demo and gain practical insights for cost-effective LLM deployment in production!
Follow AWS OnAir:
LinkedIn: https://bit.ly/AWSOnAir-LinkedIn
Twitch: https://bit.ly/Twitch-AWS-OnAir
ABOUT AWS
Amazon Web Services (AWS) hosts events, both online and in-person, bringing the cloud computing community together to connect, collaborate, and learn from AWS experts. AWS is the world’s most comprehensive and broadly adopted cloud platform, offering over 200 fully featured services from data centers globally. Millions of customers—including the fastest-growing startups, largest enterprises, and leading government agencies—are using AWS to lower costs, become more agile, and innovate faster.
#AWS #AWSpartner #nvidia #AI #NIM
Повторяем попытку...
Доступные форматы для скачивания:
Скачать видео
-
Информация по загрузке: