From a47cb9cdea32572eb9bd59058514ba625c26af52 Mon Sep 17 00:00:00 2001 From: Paul Gauthier Date: Thu, 29 Aug 2024 06:36:22 -0700 Subject: [PATCH] refactor: conditionally output cached tokens message based on verbosity --- aider/coders/base_coder.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/aider/coders/base_coder.py b/aider/coders/base_coder.py index 73f87b49f..569e4b1de 100755 --- a/aider/coders/base_coder.py +++ b/aider/coders/base_coder.py @@ -1040,8 +1040,8 @@ class Coder: completion.usage, "prompt_cache_hit_tokens", 0 ) or getattr(completion.usage, "cache_read_input_tokens", 0) - # if self.verbose: - self.io.tool_output(f"Warmed {format_tokens(cache_hit_tokens)} cached tokens.") + if self.verbose: + self.io.tool_output(f"Warmed {format_tokens(cache_hit_tokens)} cached tokens.") self.cache_warming_thread = threading.Timer(0, warm_cache_worker) self.cache_warming_thread.daemon = True