mirror of
https://github.com/Aider-AI/aider.git
synced 2025-06-13 08:05:01 +00:00
more deepseek config
This commit is contained in:
parent
e7ec80f58a
commit
2968087d37
3 changed files with 37 additions and 2 deletions
|
@ -76,6 +76,7 @@ MODEL_ALIASES = {
|
|||
"3": "gpt-3.5-turbo",
|
||||
# Other models
|
||||
"deepseek": "deepseek/deepseek-chat",
|
||||
"r1": "deepseek/deepseek-reasoner",
|
||||
"flash": "gemini/gemini-2.0-flash-exp",
|
||||
}
|
||||
|
||||
|
@ -622,6 +623,21 @@ MODEL_SETTINGS = [
|
|||
use_repo_map=True,
|
||||
send_undo_reply=False,
|
||||
),
|
||||
ModelSettings(
|
||||
"openrouter/deepseek/deepseek-r1",
|
||||
"diff",
|
||||
weak_model_name="openrouter/deepseek/deepseek-chat",
|
||||
editor_model_name="openrouter/deepseek/deepseek-chat",
|
||||
editor_edit_format="editor-diff",
|
||||
use_repo_map=True,
|
||||
examples_as_sys_msg=True,
|
||||
use_temperature=False,
|
||||
reminder="user",
|
||||
caches_by_default=True,
|
||||
extra_params={
|
||||
"max_tokens": 8192,
|
||||
},
|
||||
),
|
||||
ModelSettings(
|
||||
"deepseek/deepseek-reasoner",
|
||||
"diff",
|
||||
|
|
|
@ -1,7 +1,23 @@
|
|||
{
|
||||
"deepseek-reasoner": {
|
||||
"max_tokens": 8192,
|
||||
"max_input_tokens": 128000,
|
||||
"max_input_tokens": 64000,
|
||||
"max_output_tokens": 8192,
|
||||
"input_cost_per_token": 0.00000055,
|
||||
"input_cost_per_token_cache_hit": 0.00000014,
|
||||
"cache_read_input_token_cost": 0.00000014,
|
||||
"cache_creation_input_token_cost": 0.0,
|
||||
"output_cost_per_token": 0.00000219,
|
||||
"litellm_provider": "deepseek",
|
||||
"mode": "chat",
|
||||
//"supports_function_calling": true,
|
||||
"supports_assistant_prefill": true,
|
||||
//"supports_tool_choice": true,
|
||||
"supports_prompt_caching": true
|
||||
},
|
||||
"openrouter/deepseek/deepseek-r1": {
|
||||
"max_tokens": 8192,
|
||||
"max_input_tokens": 64000,
|
||||
"max_output_tokens": 8192,
|
||||
"input_cost_per_token": 0.00000055,
|
||||
"input_cost_per_token_cache_hit": 0.00000014,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue