Stop proxy buffering from breaking AI response streams

Useful breakdown of why “streaming” AI responses still show up as a delayed text dump: proxies and observability layers often buffer tokens, and the fix is.

Yoshiii

The missing caveat is that even with proxy buffering off, gzip, HTTP/2 coalescing, or framework middleware can still batch chunks and make SSE look haunted.

Arthur