At OpenPipe, a fully-managed fine-tuning platform for developers is now available, allowing users to replace their existing prompts with fine-tuned models in just a few minutes. The platform captures existing prompts and completions, synthesizes them into a dataset, and fine-tunes models that are a drop-in replacement for the prompt. Starting today, OpenPipe also automates the process of evaluating model performance by using GPT-4 to compare output across multiple models on a test set, allowing users to view results, see which model won, and access custom instructions and reasoning. This feature is designed to help users build and improve their fine-tuned models with ease.