From b8db82a2f8c362307f025119923751219961dfec Mon Sep 17 00:00:00 2001 From: thinkwee Date: Fri, 8 Sep 2023 10:28:31 +0800 Subject: [PATCH] fix multi-turn token exceed problem --- camel/model_backend.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/camel/model_backend.py b/camel/model_backend.py index 6d95dc56..5d895b44 100644 --- a/camel/model_backend.py +++ b/camel/model_backend.py @@ -51,7 +51,7 @@ class OpenAIModel(ModelBackend): string = "\n".join([message["content"] for message in kwargs["messages"]]) encoding = tiktoken.encoding_for_model(self.model_type.value) num_prompt_tokens = len(encoding.encode(string)) - gap_between_send_receive = 50 # known issue + gap_between_send_receive = 15 * len(kwargs["messages"]) num_prompt_tokens += gap_between_send_receive num_max_token_map = {