Traditional ML compares outputs against correct answers. But subjective decisions have no single right answer. That's exactly where AI promises value and where defining correctness gets messy. How does the system measure quality when ground truth is fuzzy or contested?