Hey everyone,
Does anyone know how to control or limit the reasoning depth in GPT-5 series models when using LangChain with OpenAI?
I’d like to use the latest models, but I’ve noticed that extended reasoning increases response latency. My goal is to reduce unnecessary thinking time while still maintaining good output quality.
Is there a recommended way to manage or restrict reasoning behavior when working with these models in LangChain?
Would appreciate any guidance. Thanks!