SPADE (System for Prompt Analysis and Delta-based Evaluation) is a tool developed by researchers at UC Berkeley in collaboration with LangChain to enhance the evaluation of Large Language Model (LLM) chains by leveraging prompt refinement history. The tool suggests Python-based evaluation functions that can assess the quality and reliability of LLM outputs by identifying changes in prompt versions and categorizing them based on a developed taxonomy. SPADE aims to address the challenges of prompt engineering and monitoring in LLM deployments by offering automated evaluation functions that can verify the adherence to constraints and guardrails encoded in prompt refinements. The prototype suggests evaluation functions by analyzing the differences between prompt versions, which are useful in ensuring LLM outputs meet specified criteria, such as excluding certain items in response to a specific context. Despite being in a preliminary stage, SPADE offers potential improvements in LLM deployment reliability and invites feedback and collaboration from developers interested in this research area.