Sloptimism is breaking any system built on human validation
Blog post from Bugcrowd
In March 2026, Bugcrowd implemented changes to its submission pipeline due to a surge in low-quality submissions, a phenomenon dubbed "sloptimism," where reports are generated quickly with more reliance on AI than on evidence. This issue, affecting sectors like security, academia, and law, arises from AI making content cheap to produce but costly to validate, challenging the manual triage systems. Bugcrowd's response included measures such as permanent bans and identity verification to manage spam. This problem mirrors the email spam issue of the late 1990s, requiring systemic adjustments like identity, reputation, and rate limits to manage submissions effectively. The underlying challenge is that AI-generated reports, while appearing legitimate, often lack substance, shifting the bottleneck from generation to validation. As AI continues to evolve, the focus must be on balancing the benefits of AI tools with the need to filter out insubstantial content without banning AI altogether, emphasizing the importance of responsible disclosure and thorough validation in technical submissions.