Hi all, I’m researching a gap in multi agent observability and evals.
In orgs where different teams each own their own LangSmith project, it seems like traces, evals, and debugging can become fragmented. This can make root cause analysis slower, especially when another team’s runs are not immediately visible or end-to-end evals stop at the project boundary.
I’d love to hear from teams who’ve run into this in production or late-stage development.
A few things I’m curious about:
-
How do you debug failures that cross team or project boundaries?
-
How do you have confidence in the output from different project boundaries?
-
Has this ever slowed incident resolution or delayed release confidence?
Just trying to understand whether this is a real pain point and how people are handling it today.