diff --git a/api/core/model_runtime/model_providers/__base/large_language_model.py b/api/core/model_runtime/model_providers/__base/large_language_model.py index b45b2ca025..9c7465536e 100644 --- a/api/core/model_runtime/model_providers/__base/large_language_model.py +++ b/api/core/model_runtime/model_providers/__base/large_language_model.py @@ -537,7 +537,6 @@ if you are not sure about the structure. """ raise NotImplementedError - @abstractmethod def get_num_tokens( self, model: str, @@ -554,7 +553,8 @@ if you are not sure about the structure. :param tools: tools for tool calling :return: """ - raise NotImplementedError + # Disable the token count in LLMs for profermance testing. + return 0 def enforce_stop_tokens(self, text: str, stop: list[str]) -> str: """Cut off the text as soon as any stop words occur.""" diff --git a/api/core/model_runtime/model_providers/__base/tokenizers/gpt2_tokenzier.py b/api/core/model_runtime/model_providers/__base/tokenizers/gpt2_tokenzier.py index 5aeb25c6f2..2f6f4fbbef 100644 --- a/api/core/model_runtime/model_providers/__base/tokenizers/gpt2_tokenzier.py +++ b/api/core/model_runtime/model_providers/__base/tokenizers/gpt2_tokenzier.py @@ -25,8 +25,7 @@ class GPT2Tokenizer: # future = _executor.submit(GPT2Tokenizer._get_num_tokens_by_gpt2, text) # result = future.result() # return cast(int, result) - # return GPT2Tokenizer._get_num_tokens_by_gpt2(text) - return 0 + return GPT2Tokenizer._get_num_tokens_by_gpt2(text) @staticmethod def get_encoder() -> Any: