From 90731653605065c1a7e965a3af80b5af2b559364 Mon Sep 17 00:00:00 2001 From: Pulkit Thukral <85827394+PulkitT01@users.noreply.github.com> Date: Tue, 16 Sep 2025 10:39:31 +0200 Subject: [PATCH] Fix token limit check to use compression model --- src/open_deep_research/deep_researcher.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/open_deep_research/deep_researcher.py b/src/open_deep_research/deep_researcher.py index 279dbffd9..e4e5d4b34 100644 --- a/src/open_deep_research/deep_researcher.py +++ b/src/open_deep_research/deep_researcher.py @@ -566,7 +566,7 @@ async def compress_research(state: ResearcherState, config: RunnableConfig): synthesis_attempts += 1 # Handle token limit exceeded by removing older messages - if is_token_limit_exceeded(e, configurable.research_model): + if is_token_limit_exceeded(e, configurable.compression_model): researcher_messages = remove_up_to_last_ai_message(researcher_messages) continue @@ -716,4 +716,4 @@ async def final_report_generation(state: AgentState, config: RunnableConfig): deep_researcher_builder.add_edge("final_report_generation", END) # Final exit point # Compile the complete deep researcher workflow -deep_researcher = deep_researcher_builder.compile() \ No newline at end of file +deep_researcher = deep_researcher_builder.compile()