mirror of
https://github.com/Aider-AI/aider.git
synced 2025-05-30 09:14:59 +00:00
Merge branch 'main' of github.com:Aider-AI/aider
This commit is contained in:
commit
2f307e46e2
9 changed files with 36 additions and 15 deletions
|
@ -237,8 +237,8 @@ def get_parser(default_config_files, git_root):
|
||||||
type=int,
|
type=int,
|
||||||
default=None,
|
default=None,
|
||||||
help=(
|
help=(
|
||||||
"Maximum number of tokens to use for chat history. If not specified, uses the model's"
|
"Soft limit on tokens for chat history, after which summarization begins."
|
||||||
" max_chat_history_tokens."
|
" If unspecified, defaults to the model's max_chat_history_tokens."
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
# This is a duplicate of the argument in the preparser and is a no-op by this time of
|
# This is a duplicate of the argument in the preparser and is a no-op by this time of
|
||||||
|
@ -719,7 +719,6 @@ def get_md_help():
|
||||||
parser.formatter_class = MarkdownHelpFormatter
|
parser.formatter_class = MarkdownHelpFormatter
|
||||||
|
|
||||||
return argparse.ArgumentParser.format_help(parser)
|
return argparse.ArgumentParser.format_help(parser)
|
||||||
return parser.format_help()
|
|
||||||
|
|
||||||
|
|
||||||
def get_sample_yaml():
|
def get_sample_yaml():
|
||||||
|
@ -733,7 +732,6 @@ def get_sample_yaml():
|
||||||
parser.formatter_class = YamlHelpFormatter
|
parser.formatter_class = YamlHelpFormatter
|
||||||
|
|
||||||
return argparse.ArgumentParser.format_help(parser)
|
return argparse.ArgumentParser.format_help(parser)
|
||||||
return parser.format_help()
|
|
||||||
|
|
||||||
|
|
||||||
def get_sample_dotenv():
|
def get_sample_dotenv():
|
||||||
|
@ -747,7 +745,6 @@ def get_sample_dotenv():
|
||||||
parser.formatter_class = DotEnvFormatter
|
parser.formatter_class = DotEnvFormatter
|
||||||
|
|
||||||
return argparse.ArgumentParser.format_help(parser)
|
return argparse.ArgumentParser.format_help(parser)
|
||||||
return parser.format_help()
|
|
||||||
|
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
|
|
|
@ -319,7 +319,7 @@ MODEL_SETTINGS = [
|
||||||
ModelSettings(
|
ModelSettings(
|
||||||
"openrouter/anthropic/claude-3.5-sonnet",
|
"openrouter/anthropic/claude-3.5-sonnet",
|
||||||
"diff",
|
"diff",
|
||||||
weak_model_name="openrouter/anthropic/claude-3-haiku-20240307",
|
weak_model_name="openrouter/anthropic/claude-3-haiku",
|
||||||
editor_model_name="openrouter/anthropic/claude-3.5-sonnet",
|
editor_model_name="openrouter/anthropic/claude-3.5-sonnet",
|
||||||
editor_edit_format="editor-diff",
|
editor_edit_format="editor-diff",
|
||||||
use_repo_map=True,
|
use_repo_map=True,
|
||||||
|
@ -334,7 +334,7 @@ MODEL_SETTINGS = [
|
||||||
ModelSettings(
|
ModelSettings(
|
||||||
"openrouter/anthropic/claude-3.5-sonnet:beta",
|
"openrouter/anthropic/claude-3.5-sonnet:beta",
|
||||||
"diff",
|
"diff",
|
||||||
weak_model_name="openrouter/anthropic/claude-3-haiku-20240307",
|
weak_model_name="openrouter/anthropic/claude-3-haiku:beta",
|
||||||
editor_model_name="openrouter/anthropic/claude-3.5-sonnet:beta",
|
editor_model_name="openrouter/anthropic/claude-3.5-sonnet:beta",
|
||||||
editor_edit_format="editor-diff",
|
editor_edit_format="editor-diff",
|
||||||
use_repo_map=True,
|
use_repo_map=True,
|
||||||
|
@ -513,6 +513,18 @@ MODEL_SETTINGS = [
|
||||||
use_temperature=False,
|
use_temperature=False,
|
||||||
streaming=False,
|
streaming=False,
|
||||||
),
|
),
|
||||||
|
ModelSettings(
|
||||||
|
"azure/o1-mini",
|
||||||
|
"whole",
|
||||||
|
weak_model_name="azure/gpt-4o-mini",
|
||||||
|
editor_model_name="azure/gpt-4o",
|
||||||
|
editor_edit_format="editor-diff",
|
||||||
|
use_repo_map=True,
|
||||||
|
reminder="user",
|
||||||
|
use_system_prompt=False,
|
||||||
|
use_temperature=False,
|
||||||
|
streaming=False,
|
||||||
|
),
|
||||||
ModelSettings(
|
ModelSettings(
|
||||||
"o1-mini",
|
"o1-mini",
|
||||||
"whole",
|
"whole",
|
||||||
|
@ -537,6 +549,18 @@ MODEL_SETTINGS = [
|
||||||
use_temperature=False,
|
use_temperature=False,
|
||||||
streaming=False,
|
streaming=False,
|
||||||
),
|
),
|
||||||
|
ModelSettings(
|
||||||
|
"azure/o1-preview",
|
||||||
|
"diff",
|
||||||
|
weak_model_name="azure/gpt-4o-mini",
|
||||||
|
editor_model_name="azure/gpt-4o",
|
||||||
|
editor_edit_format="editor-diff",
|
||||||
|
use_repo_map=True,
|
||||||
|
reminder="user",
|
||||||
|
use_system_prompt=False,
|
||||||
|
use_temperature=False,
|
||||||
|
streaming=False,
|
||||||
|
),
|
||||||
ModelSettings(
|
ModelSettings(
|
||||||
"o1-preview",
|
"o1-preview",
|
||||||
"architect",
|
"architect",
|
||||||
|
|
|
@ -2,4 +2,4 @@ You can send long, multi-line messages in the chat in a few ways:
|
||||||
- Paste a multi-line message directly into the chat.
|
- Paste a multi-line message directly into the chat.
|
||||||
- Enter `{` alone on the first line to start a multiline message and `}` alone on the last line to end it.
|
- Enter `{` alone on the first line to start a multiline message and `}` alone on the last line to end it.
|
||||||
- Use Meta-ENTER to start a new line without sending the message (Esc+ENTER in some environments).
|
- Use Meta-ENTER to start a new line without sending the message (Esc+ENTER in some environments).
|
||||||
- Use `/clipboard` to paste text from the clipboard into the chat.
|
- Use `/paste` to paste text from the clipboard into the chat.
|
||||||
|
|
|
@ -104,7 +104,7 @@
|
||||||
## Only work with models that have meta-data available (default: True)
|
## Only work with models that have meta-data available (default: True)
|
||||||
#show-model-warnings: true
|
#show-model-warnings: true
|
||||||
|
|
||||||
## Maximum number of tokens to use for chat history. If not specified, uses the model's max_chat_history_tokens.
|
## Soft limit on tokens for chat history, after which summarization begins. If unspecified, defaults to the model's max_chat_history_tokens.
|
||||||
#max-chat-history-tokens: xxx
|
#max-chat-history-tokens: xxx
|
||||||
|
|
||||||
## Specify the .env file to load (default: .env in git root)
|
## Specify the .env file to load (default: .env in git root)
|
||||||
|
|
|
@ -108,7 +108,7 @@
|
||||||
## Only work with models that have meta-data available (default: True)
|
## Only work with models that have meta-data available (default: True)
|
||||||
#AIDER_SHOW_MODEL_WARNINGS=true
|
#AIDER_SHOW_MODEL_WARNINGS=true
|
||||||
|
|
||||||
## Maximum number of tokens to use for chat history. If not specified, uses the model's max_chat_history_tokens.
|
## Soft limit on tokens for chat history, after which summarization begins. If unspecified, defaults to the model's max_chat_history_tokens.
|
||||||
#AIDER_MAX_CHAT_HISTORY_TOKENS=
|
#AIDER_MAX_CHAT_HISTORY_TOKENS=
|
||||||
|
|
||||||
## Specify the .env file to load (default: .env in git root)
|
## Specify the .env file to load (default: .env in git root)
|
||||||
|
|
|
@ -550,7 +550,7 @@ cog.out("```\n")
|
||||||
use_repo_map: true
|
use_repo_map: true
|
||||||
use_system_prompt: true
|
use_system_prompt: true
|
||||||
use_temperature: true
|
use_temperature: true
|
||||||
weak_model_name: openrouter/anthropic/claude-3-haiku-20240307
|
weak_model_name: openrouter/anthropic/claude-3-haiku
|
||||||
- accepts_images: true
|
- accepts_images: true
|
||||||
cache_control: true
|
cache_control: true
|
||||||
caches_by_default: false
|
caches_by_default: false
|
||||||
|
@ -568,7 +568,7 @@ cog.out("```\n")
|
||||||
use_repo_map: true
|
use_repo_map: true
|
||||||
use_system_prompt: true
|
use_system_prompt: true
|
||||||
use_temperature: true
|
use_temperature: true
|
||||||
weak_model_name: openrouter/anthropic/claude-3-haiku-20240307
|
weak_model_name: openrouter/anthropic/claude-3-haiku:beta
|
||||||
- accepts_images: true
|
- accepts_images: true
|
||||||
cache_control: false
|
cache_control: false
|
||||||
caches_by_default: false
|
caches_by_default: false
|
||||||
|
|
|
@ -160,7 +160,7 @@ cog.outl("```")
|
||||||
## Only work with models that have meta-data available (default: True)
|
## Only work with models that have meta-data available (default: True)
|
||||||
#show-model-warnings: true
|
#show-model-warnings: true
|
||||||
|
|
||||||
## Maximum number of tokens to use for chat history. If not specified, uses the model's max_chat_history_tokens.
|
## Soft limit on tokens for chat history, after which summarization begins. If unspecified, defaults to the model's max_chat_history_tokens.
|
||||||
#max-chat-history-tokens: xxx
|
#max-chat-history-tokens: xxx
|
||||||
|
|
||||||
## Specify the .env file to load (default: .env in git root)
|
## Specify the .env file to load (default: .env in git root)
|
||||||
|
|
|
@ -150,7 +150,7 @@ cog.outl("```")
|
||||||
## Only work with models that have meta-data available (default: True)
|
## Only work with models that have meta-data available (default: True)
|
||||||
#AIDER_SHOW_MODEL_WARNINGS=true
|
#AIDER_SHOW_MODEL_WARNINGS=true
|
||||||
|
|
||||||
## Maximum number of tokens to use for chat history. If not specified, uses the model's max_chat_history_tokens.
|
## Soft limit on tokens for chat history, after which summarization begins. If unspecified, defaults to the model's max_chat_history_tokens.
|
||||||
#AIDER_MAX_CHAT_HISTORY_TOKENS=
|
#AIDER_MAX_CHAT_HISTORY_TOKENS=
|
||||||
|
|
||||||
## Specify the .env file to load (default: .env in git root)
|
## Specify the .env file to load (default: .env in git root)
|
||||||
|
|
|
@ -223,7 +223,7 @@ Aliases:
|
||||||
- `--no-show-model-warnings`
|
- `--no-show-model-warnings`
|
||||||
|
|
||||||
### `--max-chat-history-tokens VALUE`
|
### `--max-chat-history-tokens VALUE`
|
||||||
Maximum number of tokens to use for chat history. If not specified, uses the model's max_chat_history_tokens.
|
Soft limit on tokens for chat history, after which summarization begins. If unspecified, defaults to the model's max_chat_history_tokens.
|
||||||
Environment variable: `AIDER_MAX_CHAT_HISTORY_TOKENS`
|
Environment variable: `AIDER_MAX_CHAT_HISTORY_TOKENS`
|
||||||
|
|
||||||
### `--env-file ENV_FILE`
|
### `--env-file ENV_FILE`
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue