mirror of
https://github.com/Aider-AI/aider.git
synced 2025-06-04 11:45:00 +00:00
copy
This commit is contained in:
parent
e17c29c258
commit
384ff3484c
5 changed files with 35 additions and 29 deletions
|
@ -13,7 +13,7 @@ Model aliases allow you to create shorthand names for models you frequently use.
|
||||||
You can define aliases when launching aider using the `--alias` option:
|
You can define aliases when launching aider using the `--alias` option:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
aider --alias "fast:gpt-3.5-turbo" --alias "smart:gpt-4"
|
aider --alias "fast:gpt-4o-mini" --alias "smart:o3-mini"
|
||||||
```
|
```
|
||||||
|
|
||||||
Multiple aliases can be defined by using the `--alias` option multiple times. Each alias definition should be in the format `alias:model-name`.
|
Multiple aliases can be defined by using the `--alias` option multiple times. Each alias definition should be in the format `alias:model-name`.
|
||||||
|
@ -24,8 +24,8 @@ You can also define aliases in your [`.aider.conf.yml` file](https://aider.chat/
|
||||||
|
|
||||||
```yaml
|
```yaml
|
||||||
alias:
|
alias:
|
||||||
- "fast:gpt-3.5-turbo"
|
- "fast:gpt-4o-mini"
|
||||||
- "smart:gpt-4"
|
- "smart:o3-mini"
|
||||||
- "hacker:claude-3-sonnet-20240229"
|
- "hacker:claude-3-sonnet-20240229"
|
||||||
```
|
```
|
||||||
|
|
||||||
|
@ -34,8 +34,8 @@ alias:
|
||||||
Once defined, you can use the alias instead of the full model name:
|
Once defined, you can use the alias instead of the full model name:
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
aider --model fast # Uses gpt-3.5-turbo
|
aider --model fast # Uses gpt-4o-mini
|
||||||
aider --model smart # Uses gpt-4
|
aider --model smart # Uses o3-mini
|
||||||
```
|
```
|
||||||
|
|
||||||
## Built-in Aliases
|
## Built-in Aliases
|
||||||
|
|
|
@ -8,7 +8,7 @@ nav_order: 100
|
||||||
To work with OpenAI's models, you need to provide your
|
To work with OpenAI's models, you need to provide your
|
||||||
[OpenAI API key](https://help.openai.com/en/articles/4936850-where-do-i-find-my-secret-api-key)
|
[OpenAI API key](https://help.openai.com/en/articles/4936850-where-do-i-find-my-secret-api-key)
|
||||||
either in the `OPENAI_API_KEY` environment variable or
|
either in the `OPENAI_API_KEY` environment variable or
|
||||||
via the `--openai-api-key` command line switch.
|
via the `--api-key openai=<key>` command line switch.
|
||||||
|
|
||||||
Aider has some built in shortcuts for the most popular OpenAI models and
|
Aider has some built in shortcuts for the most popular OpenAI models and
|
||||||
has been tested and benchmarked to work well with them:
|
has been tested and benchmarked to work well with them:
|
||||||
|
@ -16,28 +16,34 @@ has been tested and benchmarked to work well with them:
|
||||||
```
|
```
|
||||||
python -m pip install -U aider-chat
|
python -m pip install -U aider-chat
|
||||||
|
|
||||||
export OPENAI_API_KEY=<key> # Mac/Linux
|
# o3-mini
|
||||||
setx OPENAI_API_KEY <key> # Windows, restart shell after setx
|
aider --model o3-mini --api-key openai=<key>
|
||||||
|
|
||||||
# Aider uses gpt-4o by default (or use --4o)
|
|
||||||
aider
|
|
||||||
|
|
||||||
# GPT-4o
|
|
||||||
aider --4o
|
|
||||||
|
|
||||||
# GPT-3.5 Turbo
|
|
||||||
aider --35-turbo
|
|
||||||
|
|
||||||
# o1-mini
|
# o1-mini
|
||||||
aider --model o1-mini
|
aider --model o1-mini --api-key openai=<key>
|
||||||
|
|
||||||
# o1-preview
|
# GPT-4o
|
||||||
aider --model o1-preview
|
aider --4o --api-key openai=<key>
|
||||||
|
|
||||||
# List models available from OpenAI
|
# List models available from OpenAI
|
||||||
aider --list-models openai/
|
aider --list-models openai/
|
||||||
|
|
||||||
|
# You can also store you API key in environment variables (or .env)
|
||||||
|
export OPENAI_API_KEY=<key> # Mac/Linux
|
||||||
|
setx OPENAI_API_KEY <key> # Windows, restart shell after setx
|
||||||
```
|
```
|
||||||
|
|
||||||
You can use `aider --model <model-name>` to use any other OpenAI model.
|
You can use `aider --model <model-name>` to use any other OpenAI model.
|
||||||
For example, if you want to use a specific version of GPT-4 Turbo
|
For example, if you want to use a specific version of GPT-4 Turbo
|
||||||
you could do `aider --model gpt-4-0125-preview`.
|
you could do `aider --model gpt-4-0125-preview`.
|
||||||
|
|
||||||
|
## o1 models from other providers
|
||||||
|
|
||||||
|
Many of OpenAI's o1
|
||||||
|
"reasoning" models have restrictions on streaming and setting the temperature parameter.
|
||||||
|
Aider is configured to work properly with these models
|
||||||
|
when served through major provider APIs.
|
||||||
|
|
||||||
|
You may need to [configure reasoning model settings](/docs/config/reasoning.html)
|
||||||
|
if you are using them through another provider
|
||||||
|
and see errors related to temperature or system prompt.
|
||||||
|
|
|
@ -16,13 +16,13 @@ aider --model deepseek --api-key deepseek=your-key-goes-here
|
||||||
# Work with Claude 3.5 Sonnet via Anthropic's API
|
# Work with Claude 3.5 Sonnet via Anthropic's API
|
||||||
aider --model sonnet --api-key anthropic=your-key-goes-here
|
aider --model sonnet --api-key anthropic=your-key-goes-here
|
||||||
|
|
||||||
# Work with GPT-4o via OpenAI's API
|
# Work with o3-mini via OpenAI's API
|
||||||
aider --model gpt-4o --api-key openai=your-key-goes-here
|
aider --model o3-mini --api-key openai=your-key-goes-here
|
||||||
|
|
||||||
# Work with Sonnet via OpenRouter's API
|
# Work with Sonnet via OpenRouter's API
|
||||||
aider --model openrouter/anthropic/claude-3.5-sonnet --api-key openrouter=your-key-goes-here
|
aider --model openrouter/anthropic/claude-3.5-sonnet --api-key openrouter=your-key-goes-here
|
||||||
|
|
||||||
# Work with DeepSeek via OpenRouter's API
|
# Work with DeepSeek Chat V3 via OpenRouter's API
|
||||||
aider --model openrouter/deepseek/deepseek-chat --api-key openrouter=your-key-goes-here
|
aider --model openrouter/deepseek/deepseek-chat --api-key openrouter=your-key-goes-here
|
||||||
```
|
```
|
||||||
|
|
||||||
|
|
|
@ -29,7 +29,7 @@ Total tokens: 4864 of 16385
|
||||||
To reduce output tokens:
|
To reduce output tokens:
|
||||||
- Ask for smaller changes in each request.
|
- Ask for smaller changes in each request.
|
||||||
- Break your code into smaller source files.
|
- Break your code into smaller source files.
|
||||||
- Try using a stronger model like gpt-4o or opus that can return diffs.
|
- Try using a stronger model like DeepSeek V3 or Sonnet that can return diffs.
|
||||||
|
|
||||||
For more info: https://aider.chat/docs/token-limits.html
|
For more info: https://aider.chat/docs/token-limits.html
|
||||||
```
|
```
|
||||||
|
@ -47,7 +47,7 @@ overflowing its context window.
|
||||||
Technically you can exhaust the context window if the input is
|
Technically you can exhaust the context window if the input is
|
||||||
too large or if the input plus output are too large.
|
too large or if the input plus output are too large.
|
||||||
|
|
||||||
Strong models like GPT-4o and Opus have quite
|
Strong models like GPT-4o and Sonnet have quite
|
||||||
large context windows, so this sort of error is
|
large context windows, so this sort of error is
|
||||||
typically only an issue when working with weaker models.
|
typically only an issue when working with weaker models.
|
||||||
|
|
||||||
|
@ -73,7 +73,7 @@ To avoid hitting output token limits:
|
||||||
|
|
||||||
- Ask for smaller changes in each request.
|
- Ask for smaller changes in each request.
|
||||||
- Break your code into smaller source files.
|
- Break your code into smaller source files.
|
||||||
- Use a strong model like gpt-4o, sonnet or opus that can return diffs.
|
- Use a strong model like gpt-4o, sonnet or DeepSeek V3 that can return diffs.
|
||||||
- Use a model that supports [infinite output](/docs/more/infinite-output.html).
|
- Use a model that supports [infinite output](/docs/more/infinite-output.html).
|
||||||
|
|
||||||
## Other causes
|
## Other causes
|
||||||
|
|
|
@ -68,11 +68,11 @@ relevant context from the rest of your repo.
|
||||||
{% include works-best.md %}
|
{% include works-best.md %}
|
||||||
|
|
||||||
```
|
```
|
||||||
# GPT-4o
|
# o3-mini
|
||||||
$ aider --4o
|
$ aider --model o3-mini --api-key openai=<key>
|
||||||
|
|
||||||
# Claude 3.5 Sonnet
|
# Claude 3.5 Sonnet
|
||||||
$ aider --sonnet
|
$ aider --model sonnet --api-key anthropic=<key>
|
||||||
```
|
```
|
||||||
|
|
||||||
Or you can run `aider --model XXX` to launch aider with
|
Or you can run `aider --model XXX` to launch aider with
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue