more deepseek config

This commit is contained in:
Paul Gauthier 2025-01-20 10:09:48 -08:00
parent e7ec80f58a
commit 2968087d37
3 changed files with 37 additions and 2 deletions

View file

@ -2,6 +2,9 @@
### main branch
- Support for DeepSeek R1.
- Use shortcut: `--model r1`
- Also via OpenRouter: `--model openrouter/deepseek/deepseek-r1`
- Added `--line-endings` for file writing, by Titusz Pan.
- Added examples_as_sys_msg=True for GPT-4o models, improves benchmark scores.
- Bumped all dependencies, to pick up litellm support for o1 system messages.
@ -10,7 +13,7 @@
### Aider v0.71.1
- Fix permissions issue in Docker images.
- Added read-only file announcements to chat.
- Added read-only file announcements.
- Bugfix: ASCII fallback for unicode errors.
- Bugfix: integer indices for list slicing in repomap calculations.

View file

@ -76,6 +76,7 @@ MODEL_ALIASES = {
"3": "gpt-3.5-turbo",
# Other models
"deepseek": "deepseek/deepseek-chat",
"r1": "deepseek/deepseek-reasoner",
"flash": "gemini/gemini-2.0-flash-exp",
}
@ -622,6 +623,21 @@ MODEL_SETTINGS = [
use_repo_map=True,
send_undo_reply=False,
),
ModelSettings(
"openrouter/deepseek/deepseek-r1",
"diff",
weak_model_name="openrouter/deepseek/deepseek-chat",
editor_model_name="openrouter/deepseek/deepseek-chat",
editor_edit_format="editor-diff",
use_repo_map=True,
examples_as_sys_msg=True,
use_temperature=False,
reminder="user",
caches_by_default=True,
extra_params={
"max_tokens": 8192,
},
),
ModelSettings(
"deepseek/deepseek-reasoner",
"diff",

View file

@ -1,7 +1,23 @@
{
"deepseek-reasoner": {
"max_tokens": 8192,
"max_input_tokens": 128000,
"max_input_tokens": 64000,
"max_output_tokens": 8192,
"input_cost_per_token": 0.00000055,
"input_cost_per_token_cache_hit": 0.00000014,
"cache_read_input_token_cost": 0.00000014,
"cache_creation_input_token_cost": 0.0,
"output_cost_per_token": 0.00000219,
"litellm_provider": "deepseek",
"mode": "chat",
//"supports_function_calling": true,
"supports_assistant_prefill": true,
//"supports_tool_choice": true,
"supports_prompt_caching": true
},
"openrouter/deepseek/deepseek-r1": {
"max_tokens": 8192,
"max_input_tokens": 64000,
"max_output_tokens": 8192,
"input_cost_per_token": 0.00000055,
"input_cost_per_token_cache_hit": 0.00000014,