There is a newer version of the record available.

Published February 10, 2026 | Version v1
Preprint Open

The Negative Instruction Cascade: An Empirical Analysis of LLM Prompt Engineering Failure Modes

Description

This document analyzes an empirical experiment in which progressive JSON instruction refinement led to catastrophic LLM failure (static repetitive loops). Through systematic analysis of three instruction versions (v1.1 → v2.1 → v3.0), Claude LLM identify a  threshold: when negative restrictions exceed ~40% of total instructions, models begin exhibiting breakdown behaviors; beyond 60%, static loops become inevitable.

Key Finding: The shift from positive directives ("what to do") to negative restrictions ("what not to do") creates a cognitive bind that collapses the model's response generation space, forcing repetition of the last known safe pattern.

Practical Implication: Effective prompt engineering requires maintaining at least a 3:1 ratio of positive directives to negative restrictions, with total negatives not exceeding 30% of instruction content.

Files

JSONupdatemistral7Feb26.txt

Files (84.4 kB)