Company
Date Published
Author
Gilles Closset, Fabien Ric, and Elias Tourneux
Word count
788
Language
-
Hacker News points
None

Summary

OVHcloud has become an Inference Provider on the Hugging Face Hub, enhancing the platform's capabilities for serverless inference and allowing users to easily access and utilize popular open-weight models like gpt-oss and Llama directly from Hugging Face. The integration is seamless with Hugging Face's client SDKs for both JavaScript and Python, facilitating the use of various models through preferred providers. OVHcloud AI Endpoints offer a fully managed, serverless service with competitive pricing and are designed for production use, delivering fast response times suitable for interactive applications. The service is hosted in European data centers, ensuring data sovereignty and low latency for European users, and supports advanced features for text and image processing. Users can set their API keys for different providers in their account settings, and billing is handled according to whether requests are made directly through providers or routed via Hugging Face. Hugging Face's PRO plan offers additional benefits, including Inference credits and higher usage limits, encouraging users to provide feedback through its community platform.