Company
Date Published
Author
Yann Léger
Word count
555
Language
English
Hacker News points
None

Summary

Koyeb has launched Serverless GPUs in public preview, offering an innovative solution for AI inference workloads through their platform. These GPUs, designed to handle both generative AI and computer vision models, feature high-end specifications such as up to 48GB of vRAM, 733 TFLOPS, and 900GB/s memory bandwidth, with pricing starting at $0.50 per hour. The platform supports seamless deployment with features like one-click Docker container deployment, built-in load balancing, horizontal autoscaling, zero downtime, auto-healing, and real-time monitoring. Users can pause GPU Instances to optimize costs and deploy services using the Koyeb CLI or Dashboard, either by using pre-built containers or linking to a GitHub repository. Koyeb is actively working with early users to expand their offerings, including more GPUs and accelerators, and invites feedback from users requiring specific configurations. The company fosters community engagement through their serverless community and social media presence.