Evidence suggests reliability decreases as agent autonomy increases with current LLMs. If this tradeoff proves non-negotiable, we're optimizing for the wrong metric. Or maybe the tension dissolves as models improve. Which assumption shapes your architecture today?