mirror of
https://github.com/Aider-AI/aider.git
synced 2025-05-31 09:44:59 +00:00
cleanup
This commit is contained in:
parent
1dd1f7eb22
commit
97a70830e9
2 changed files with 12 additions and 12 deletions
|
@ -22,7 +22,7 @@ import git
|
||||||
from rich.console import Console, Text
|
from rich.console import Console, Text
|
||||||
from rich.markdown import Markdown
|
from rich.markdown import Markdown
|
||||||
|
|
||||||
from aider import __version__, models, prompts, urls
|
from aider import __version__, models, prompts, urls, utils
|
||||||
from aider.commands import Commands
|
from aider.commands import Commands
|
||||||
from aider.history import ChatSummary
|
from aider.history import ChatSummary
|
||||||
from aider.io import ConfirmGroup, InputOutput
|
from aider.io import ConfirmGroup, InputOutput
|
||||||
|
@ -32,13 +32,7 @@ from aider.mdstream import MarkdownStream
|
||||||
from aider.repo import GitRepo
|
from aider.repo import GitRepo
|
||||||
from aider.repomap import RepoMap
|
from aider.repomap import RepoMap
|
||||||
from aider.sendchat import retry_exceptions, send_completion
|
from aider.sendchat import retry_exceptions, send_completion
|
||||||
from aider.utils import (
|
from aider.utils import format_content, format_messages, format_tokens, is_image_file
|
||||||
format_content,
|
|
||||||
format_messages,
|
|
||||||
format_tokens,
|
|
||||||
is_image_file,
|
|
||||||
utils,
|
|
||||||
)
|
|
||||||
|
|
||||||
from ..dump import dump # noqa: F401
|
from ..dump import dump # noqa: F401
|
||||||
from .chat_chunks import ChatChunks
|
from .chat_chunks import ChatChunks
|
||||||
|
@ -1003,7 +997,7 @@ class Coder:
|
||||||
|
|
||||||
def warm_cache_worker():
|
def warm_cache_worker():
|
||||||
for i in range(self.num_cache_warming_pings):
|
for i in range(self.num_cache_warming_pings):
|
||||||
time.sleep(10) # 290 == 4 minutes and 50 seconds
|
time.sleep(20) # 290 == 4 minutes and 50 seconds
|
||||||
try:
|
try:
|
||||||
completion = litellm.completion(
|
completion = litellm.completion(
|
||||||
model=self.main_model.name,
|
model=self.main_model.name,
|
||||||
|
@ -1021,8 +1015,8 @@ class Coder:
|
||||||
) or getattr(completion.usage, "cache_read_input_tokens", 0)
|
) or getattr(completion.usage, "cache_read_input_tokens", 0)
|
||||||
|
|
||||||
self.io.tool_output(
|
self.io.tool_output(
|
||||||
f"Warmed {utils.format_tokens(cache_hit_tokens)} cached tokens."
|
f"Warmed {format_tokens(cache_hit_tokens)} cached tokens."
|
||||||
f" ({i+1}/{self.num_cache_warming_pings})"
|
f" ({i + 1}/{self.num_cache_warming_pings})"
|
||||||
)
|
)
|
||||||
|
|
||||||
self.io.tool_output("Stopped warming.")
|
self.io.tool_output("Stopped warming.")
|
||||||
|
|
|
@ -47,7 +47,13 @@ def lazy_litellm_retry_decorator(func):
|
||||||
|
|
||||||
|
|
||||||
def send_completion(
|
def send_completion(
|
||||||
model_name, messages, functions, stream, temperature=0, extra_headers=None, max_tokens=None
|
model_name,
|
||||||
|
messages,
|
||||||
|
functions,
|
||||||
|
stream,
|
||||||
|
temperature=0,
|
||||||
|
extra_headers=None,
|
||||||
|
max_tokens=None,
|
||||||
):
|
):
|
||||||
from aider.llm import litellm
|
from aider.llm import litellm
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue