Home / Companies / HuggingFace / Blog / Post Details
Content Deep Dive

New in llama.cpp: Anthropic Messages API

Blog post from HuggingFace

Post Details
Company
Date Published
Author
Xuan-Son Nguyen and Victor Mustar
Word Count
541
Language
-
Hacker News Points
-
Summary

Llama.cpp has introduced support for the Anthropic Messages API, allowing Claude-compatible clients to interact with locally-running models via a lightweight, OpenAI-compatible HTTP server. This update enables applications like Claude Code to work with local models by converting Anthropic's format to OpenAI's internally, utilizing the existing inference pipeline. Users can direct their Anthropic clients to the /v1/messages endpoint to facilitate chat completions with streaming support, token counting, and tool use. The update also includes features for handling image inputs and reasoning models, as well as compatibility with specialized coding models for agentic workloads. This enhancement was contributed by noname22 and addresses a popular community request, reflecting the ongoing development and community engagement around the llama.cpp project.