AI has gotten weirdly good at making something look reasonable before anyone has really pressure-tested it. That feels useful right up until a draft, spec, or answer sounds polished enough that people stop asking the annoying questions.
I keep wondering whether that’s a real work quality problem or just a speed bump we’ll all get used to. Have you seen AI make it easier to miss the assumptions hiding under the surface?