From ddec8325e75ec5ed2a89bd842ee83cff39b4a8e0 Mon Sep 17 00:00:00 2001 From: Paul Gauthier Date: Tue, 4 Feb 2025 11:45:04 -0800 Subject: [PATCH] fix: Simplify temperature handling in model completion call --- aider/coders/base_coder.py | 8 +------- 1 file changed, 1 insertion(+), 7 deletions(-) diff --git a/aider/coders/base_coder.py b/aider/coders/base_coder.py index 1885ccd8e..89780381a 100755 --- a/aider/coders/base_coder.py +++ b/aider/coders/base_coder.py @@ -1612,19 +1612,13 @@ See https://aider.chat/docs/llms/ollama.html#setting-the-context-window-size self.io.log_llm_history("TO LLM", format_messages(messages)) - if self.main_model.use_temperature: - temp = self.temperature - else: - temp = None - completion = None try: hash_object, completion = model.send_completion( messages, functions, self.stream, - temp, - extra_params=model.extra_params, + self.temperature, ) self.chat_completion_call_hashes.append(hash_object.hexdigest())