The Probability Pixie Will Ruin Your Day
The Probability Pixie is the monster of non-deterministic output. The same prompt produces different results each time, and the developer who got a great answer on Tuesday can't reproduce it on Wednesday. Consistency becomes a matter of luck.
This isn't a bug in the model — it's a feature. Language models sample from probability distributions. They're designed to produce varied output. That's what makes them creative and useful. But it's also what makes them unreliable as components of a deterministic system like a software build pipeline.
The Probability Pixie shows up when teams start treating AI output as reproducible. They write a prompt, get a result, and assume that the same prompt will produce the same result next time. It won't. And the failure mode is subtle: the output is usually similar, which makes the difference hard to spot until it matters.
This is an excerpt from The AI Developer's Field Guide.