Bento has launched an open-source tool called llm-optimizer to assist AI teams in optimizing and benchmarking large language model (LLM) inference across various frameworks. This tool allows users to set specific performance constraints and quickly identify configurations that meet these requirements without extensive trial and error. Accompanying the tool is the LLM Performance Explorer, a website that displays benchmark results for easy comparison without the need for personal experimentation. The llm-optimizer supports systematic testing of server and client parameters, offering detailed metrics and constraint filtering to streamline performance tuning. Additionally, it features performance estimation capabilities for theoretical analysis and provides an interactive user interface for visualizing results. This initiative aims to simplify and speed up the process of LLM optimization, allowing engineers to focus more on development rather than extensive testing and configuration.