Working on AI agent infrastructure, and the biggest unsung problem is observability. When a traditional app breaks, you get stack traces, logs, metrics. When an agent decides to take a weird reasoning path, you get… nothing useful. We’ve tried embedding structured logging into every agent step, but the volume is insane. One conversation can generate 10k+ decision points. Who actually reviews that? Curious what others are doing. Are you building observability into your agents, or just hoping for the best? submitted by /u/No-Contract9167
Originally posted by u/No-Contract9167 on r/ArtificialInteligence
You must log in or # to comment.
