mirror of
https://github.com/Aider-AI/aider.git
synced 2025-06-01 02:05:00 +00:00
refactor: Update temperature handling in Model class
This commit is contained in:
parent
56eb1d106f
commit
e07fddb20b
1 changed files with 11 additions and 4 deletions
|
@ -102,7 +102,7 @@ class ModelSettings:
|
||||||
cache_control: bool = False
|
cache_control: bool = False
|
||||||
caches_by_default: bool = False
|
caches_by_default: bool = False
|
||||||
use_system_prompt: bool = True
|
use_system_prompt: bool = True
|
||||||
use_temperature: bool = True
|
use_temperature: bool = True # how can i make this a bool or a float? ai!
|
||||||
streaming: bool = True
|
streaming: bool = True
|
||||||
editor_model_name: Optional[str] = None
|
editor_model_name: Optional[str] = None
|
||||||
editor_edit_format: Optional[str] = None
|
editor_edit_format: Optional[str] = None
|
||||||
|
@ -536,7 +536,7 @@ class Model(ModelSettings):
|
||||||
def is_ollama(self):
|
def is_ollama(self):
|
||||||
return self.name.startswith("ollama/") or self.name.startswith("ollama_chat/")
|
return self.name.startswith("ollama/") or self.name.startswith("ollama_chat/")
|
||||||
|
|
||||||
def send_completion(self, messages, functions, stream, temperature=0):
|
def send_completion(self, messages, functions, stream, temperature=None):
|
||||||
if os.environ.get("AIDER_SANITY_CHECK_TURNS"):
|
if os.environ.get("AIDER_SANITY_CHECK_TURNS"):
|
||||||
sanity_check_messages(messages)
|
sanity_check_messages(messages)
|
||||||
|
|
||||||
|
@ -549,7 +549,13 @@ class Model(ModelSettings):
|
||||||
stream=stream,
|
stream=stream,
|
||||||
)
|
)
|
||||||
|
|
||||||
if self.use_temperature:
|
if self.use_temperature is not False:
|
||||||
|
if temperature is None:
|
||||||
|
if self.use_temperature in (True, None):
|
||||||
|
temperature = 0
|
||||||
|
else:
|
||||||
|
temperature = float(self.use_temperature)
|
||||||
|
|
||||||
kwargs["temperature"] = temperature
|
kwargs["temperature"] = temperature
|
||||||
|
|
||||||
if functions is not None:
|
if functions is not None:
|
||||||
|
@ -563,6 +569,8 @@ class Model(ModelSettings):
|
||||||
kwargs["num_ctx"] = num_ctx
|
kwargs["num_ctx"] = num_ctx
|
||||||
key = json.dumps(kwargs, sort_keys=True).encode()
|
key = json.dumps(kwargs, sort_keys=True).encode()
|
||||||
|
|
||||||
|
dump(kwargs)
|
||||||
|
|
||||||
hash_object = hashlib.sha1(key)
|
hash_object = hashlib.sha1(key)
|
||||||
res = litellm.completion(**kwargs)
|
res = litellm.completion(**kwargs)
|
||||||
return hash_object, res
|
return hash_object, res
|
||||||
|
@ -588,7 +596,6 @@ class Model(ModelSettings):
|
||||||
"messages": messages,
|
"messages": messages,
|
||||||
"functions": None,
|
"functions": None,
|
||||||
"stream": False,
|
"stream": False,
|
||||||
"temperature": 0,
|
|
||||||
}
|
}
|
||||||
_hash, response = self.send_completion(**kwargs)
|
_hash, response = self.send_completion(**kwargs)
|
||||||
if not response or not hasattr(response, "choices") or not response.choices:
|
if not response or not hasattr(response, "choices") or not response.choices:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue