Home / Companies / Koyeb / Blog / Post Details
Content Deep Dive

Serverless GPUs in Private Preview: L4, L40S, V100, and more

Blog post from Koyeb

Post Details
Company
Date Published
Author
Yann Léger
Word Count
555
Language
English
Hacker News Points
-
Summary

Koyeb has launched Serverless GPUs in public preview, offering an innovative solution for AI inference workloads through their platform. These GPUs, designed to handle both generative AI and computer vision models, feature high-end specifications such as up to 48GB of vRAM, 733 TFLOPS, and 900GB/s memory bandwidth, with pricing starting at $0.50 per hour. The platform supports seamless deployment with features like one-click Docker container deployment, built-in load balancing, horizontal autoscaling, zero downtime, auto-healing, and real-time monitoring. Users can pause GPU Instances to optimize costs and deploy services using the Koyeb CLI or Dashboard, either by using pre-built containers or linking to a GitHub repository. Koyeb is actively working with early users to expand their offerings, including more GPUs and accelerators, and invites feedback from users requiring specific configurations. The company fosters community engagement through their serverless community and social media presence.