fix: Use temperature setting from model configuration

This commit is contained in:
Paul Gauthier 2024-09-12 14:24:21 -07:00 committed by Paul Gauthier (aider)
parent 8aee4d25ed
commit 1755d2e0f4
3 changed files with 11 additions and 2 deletions

View file

@ -1381,6 +1381,11 @@ class Coder:
self.io.log_llm_history("TO LLM", format_messages(messages))
if self.main_model.use_temperature:
temp = self.temperature
else:
temp = None
completion = None
try:
hash_object, completion = send_completion(
@ -1388,7 +1393,7 @@ class Coder:
messages,
functions,
self.stream,
self.temperature,
temp,
extra_headers=model.extra_headers,
max_tokens=model.max_tokens,
)