From 96d02b5224fc225d761eca5e7768badb6b2160da Mon Sep 17 00:00:00 2001 From: vincent Date: Sun, 26 Jan 2025 00:01:16 +0800 Subject: [PATCH] fix cut message --- src/agent/custom_massage_manager.py | 2 +- tests/test_browser_use.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/src/agent/custom_massage_manager.py b/src/agent/custom_massage_manager.py index 075a675..c9ca432 100644 --- a/src/agent/custom_massage_manager.py +++ b/src/agent/custom_massage_manager.py @@ -95,7 +95,7 @@ class CustomMassageManager(MessageManager): """Get current message list, potentially trimmed to max tokens""" diff = self.history.total_tokens - self.max_input_tokens i = 1 # start from 1 to keep system message in history - while diff > 0: + while diff > 0 and i < len(self.history.messages): self.history.remove_message(i) diff = self.history.total_tokens - self.max_input_tokens i += 1 diff --git a/tests/test_browser_use.py b/tests/test_browser_use.py index 7dba56d..7f4a65e 100644 --- a/tests/test_browser_use.py +++ b/tests/test_browser_use.py @@ -249,7 +249,7 @@ async def test_browser_use_custom_v2(): llm = utils.get_llm_model( provider="deepseek", - model_name="deepseek-chat", + model_name="deepseek-reasoner", temperature=0.8 )