This commit is contained in:
Paul Gauthier 2024-04-19 16:42:25 -07:00
parent 08f8c8c0f5
commit 0969380ce9
3 changed files with 21 additions and 34 deletions

View file

@ -9,7 +9,8 @@ with sensible commit messages.
You can start a new project or work with an existing git repo.
Aider is unique in that it lets you ask for changes to [pre-existing, larger codebases](https://aider.chat/docs/repomap.html).
Aider works well with GPT 3.5, GPT-4, GPT-4 Turbo with Vision,
and Claude 3 Opus; it also has support for [connecting to almost any LLM](https://aider.chat/docs/llms.html).
and Claude 3 Opus.
It also supports [connecting to almost any LLM](https://aider.chat/docs/llms.html).
<p align="center">
<img src="assets/screencast.svg" alt="aider screencast">
@ -74,7 +75,7 @@ You can find more chat transcripts on the [examples page](https://aider.chat/exa
* Chat with aider about your code by launching `aider` from the command line with set of source files to discuss and edit together. Aider lets the LLM see and edit the content of those files.
* Aider can write and edit code in most popular languages: python, javascript, typescript, php, html, css, etc.
* Aider works well with GPT 3.5, GPT-4, GPT-4 Turbo with Vision,
and Claude 3 Opus; it also has support for [connecting to almost any LLM](https://aider.chat/docs/llms.html).
and Claude 3 Opus. It also supports [connecting to almost any LLM](https://aider.chat/docs/llms.html).
* Request new features, changes, improvements, or bug fixes to your code. Ask for new test cases, updated documentation or code refactors.
* Aider will apply the edits suggested by the LLM directly to your source files.
* Aider will [automatically commit each changeset to your local git repo](https://aider.chat/docs/faq.html#how-does-aider-use-git) with a descriptive commit message. These frequent, automatic commits provide a safety net. It's easy to undo changes or use standard git workflows to manage longer sequences of changes.

View file

@ -379,7 +379,7 @@ class Commands:
self.io.tool_error(
f"Cannot add image file {matched_file} as the"
f" {self.coder.main_model.name} does not support image.\nYou can run `aider"
" --4turbo` to use GPT-4 Turbo with Vision."
" --4-turbo-vision` to use GPT-4 Turbo with Vision."
)
continue
content = self.io.read_text(abs_file_path)

View file

@ -6,11 +6,11 @@ Anthropic's Claude 3 Opus and Sonnet.
GPT-4 Turbo and Claude 3 Opus are recommended for the best results.
Aider also has support for connecting to almost any LLM, but it may not work as well
depending on the capabilities of the model.
For context, GPT-3.5 is just barely capable of *editing code* to provide aider's
Aider also supports connecting to almost any LLM,
but it may not work well with some models depending on their capabilities.
For example, GPT-3.5 is just barely capable of reliably *editing code* to provide aider's
interactive "pair programming" style workflow.
Models that are less capable than GPT-3.5 may struggle to perform well with aider.
So you should expect that models which are less capable than GPT-3.5 may struggle to perform well with aider.
- [OpenAI](#openai)
- [Anthropic](#anthropic)
@ -29,9 +29,9 @@ via the `--openai-api-key` command line switch.
Aider has some built in shortcuts for the most popular OpenAI models and
has been tested and benchmarked to work well with them:
- GPT-4 Turbo: running `aider` with no arguments uses GPT-4 Turbo by default.
- GPT-4 Turbo with Vision: run `aider --4-turbo-vision` to use this vision capable model, allowing you to share images with GPT by adding them to the chat with `/add` or by naming them on the command line.
- GPT-3.5 Turbo: Run `aider --35-turbo`.
- **GPT-4 Turbo**: running `aider` with no arguments uses GPT-4 Turbo by default.
- **GPT-4 Turbo with Vision**: run `aider --4-turbo-vision` to use this vision capable model, allowing you to share images with GPT by adding them to the chat with `/add` or by naming them on the command line.
- **GPT-3.5 Turbo**: `aider --35-turbo`.
You can use `aider --model <model-name>` to use any other OpenAI model.
For example, if you want to use a specific version of GPT-4 Turbo
@ -47,8 +47,8 @@ via the `--anthropic-api-key` command line switch.
Aider has some built in shortcuts for the most popular Anthropic models and
has been tested and benchmarked to work well with them:
- Claude 3 Opus: `aider --opus`
- Claude 3 Sonnet: `aider --sonnet`
- **Claude 3 Opus**: `aider --opus`
- **Claude 3 Sonnet**: `aider --sonnet`
You can use `aider --model <model-name>` to use any other Anthropic model.
For example, if you want to use a specific version of Opus
@ -57,36 +57,22 @@ you could do `aider --model claude-3-opus-20240229`.
## Azure
Aider can be configured to connect to the OpenAI models on Azure.
You can run aider with the following arguments to connect to Azure:
Run aider with the following arguments to connect to Azure:
```
$ aider \
--openai-api-type azure \
--openai-api-key your-key-goes-here \
--openai-api-base https://example-endpoint.openai.azure.com \
--openai-api-version 2023-05-15 \
--openai-api-deployment-id deployment-name \
...
--model azure/<your_deployment_name>
--openai-api-key your-key-goes-here
--openai-api-base https://example-endpoint.openai.azure.com
--openai-api-version 2023-05-15
```
You could also store those values in an `.aider.conf.yml` file in your home directory:
Or you can run `aider --model azure/<your_deployment_name>` and
populate the following environment variables instead:
```
openai-api-type: azure
openai-api-key: your-key-goes-here
openai-api-base: https://example-endpoint.openai.azure.com
openai-api-version: 2023-05-15
openai-api-deployment-id: deployment-name
```
Or you can populate the following environment variables instead:
```
OPENAI_API_TYPE=azure
OPENAI_API_KEY=your-key-goes-here
OPENAI_API_BASE=https://example-endpoint.openai.azure.com
OPENAI_API_VERSION=2023-05-15
OPENAI_API_DEPLOYMENT_ID=deployment-name
```
See the
@ -95,7 +81,7 @@ for more information on how to populate the above configuration values.
## OpenAI compatible APIs
If you can make an LLM accessible via an OpenAI compatible API endpoint,
If your LLM is accessible via an OpenAI compatible API endpoint,
you can use `--openai-api-base` to have aider connect to it.
You might need to use `--no-require-model-info` if aider doesn't