The LangChain team has introduced a new extraction dataset designed to evaluate the ability of Large Language Models (LLMs) to extract structured information from chat logs, which aims to address common challenges in LLM application development such as classifying unstructured text and reasoning over multi-task scenarios. The dataset's schema is crafted to extract structured insights from chatbot interactions and has been tested using various LLMs, including closed-source models like GPT-4 and Claude-2, as well as open-source models like Llama 2 and models from Nous Research. While GPT-4 generally outperforms others in structured output, open-source models like Llama 2 show varying degrees of success based on model size and fine-tuning. The study also examines the effectiveness of different prompting strategies and structured decoding techniques, revealing that while structured decoding guarantees schema compliance, it does not necessarily improve the quality of the extracted values. The findings highlight the challenges in achieving consistent and accurate structured information extraction from chat data, suggesting a need for further refinement in model training and prompting strategies.