Data Driven

Ronen Dar on GPU Orchestration for Building ML Models

Informações:

Sinopsis

In this episode, our Andy Leonard and Frank La Vigne sit down with Ronen Dar, the co-founder and CTO of Run AI, to explore the world of artificial intelligence and GPU orchestration for machine learning models.Ronen shares insights into the challenges of utilizing GPUs in AI research and how Run AI's platform addresses these issues by optimizing GPU usage and providing tools for easier and faster model training and deployment. The conversation delves into the concept of fractional GPU usage, allowing multiple workloads to run on a single GPU, making expensive GPUs more accessible and cost-effective for organizations.LinksRun AI https://www.run.ai/Acquired Podcast Episode on Nvidia https://www.acquired.fm/episodes/nvidia-the-machine-learning-company-2006-2022Show Notes04:40 GPU technology enabled for cloud AI workloads.07:00 RunAI enables sharing expensive GPU resources for all.11:59 As enterprise AI matures, organizations become more savvy.15:35 Deep learning, GPUs for speed, CPUs backup.16:54 LLMs running on