mirror of
https://github.com/Aider-AI/aider.git
synced 2025-05-30 09:14:59 +00:00
cleanup
This commit is contained in:
parent
1dd1f7eb22
commit
97a70830e9
2 changed files with 12 additions and 12 deletions
|
@ -22,7 +22,7 @@ import git
|
|||
from rich.console import Console, Text
|
||||
from rich.markdown import Markdown
|
||||
|
||||
from aider import __version__, models, prompts, urls
|
||||
from aider import __version__, models, prompts, urls, utils
|
||||
from aider.commands import Commands
|
||||
from aider.history import ChatSummary
|
||||
from aider.io import ConfirmGroup, InputOutput
|
||||
|
@ -32,13 +32,7 @@ from aider.mdstream import MarkdownStream
|
|||
from aider.repo import GitRepo
|
||||
from aider.repomap import RepoMap
|
||||
from aider.sendchat import retry_exceptions, send_completion
|
||||
from aider.utils import (
|
||||
format_content,
|
||||
format_messages,
|
||||
format_tokens,
|
||||
is_image_file,
|
||||
utils,
|
||||
)
|
||||
from aider.utils import format_content, format_messages, format_tokens, is_image_file
|
||||
|
||||
from ..dump import dump # noqa: F401
|
||||
from .chat_chunks import ChatChunks
|
||||
|
@ -1003,7 +997,7 @@ class Coder:
|
|||
|
||||
def warm_cache_worker():
|
||||
for i in range(self.num_cache_warming_pings):
|
||||
time.sleep(10) # 290 == 4 minutes and 50 seconds
|
||||
time.sleep(20) # 290 == 4 minutes and 50 seconds
|
||||
try:
|
||||
completion = litellm.completion(
|
||||
model=self.main_model.name,
|
||||
|
@ -1021,8 +1015,8 @@ class Coder:
|
|||
) or getattr(completion.usage, "cache_read_input_tokens", 0)
|
||||
|
||||
self.io.tool_output(
|
||||
f"Warmed {utils.format_tokens(cache_hit_tokens)} cached tokens."
|
||||
f" ({i+1}/{self.num_cache_warming_pings})"
|
||||
f"Warmed {format_tokens(cache_hit_tokens)} cached tokens."
|
||||
f" ({i + 1}/{self.num_cache_warming_pings})"
|
||||
)
|
||||
|
||||
self.io.tool_output("Stopped warming.")
|
||||
|
|
|
@ -47,7 +47,13 @@ def lazy_litellm_retry_decorator(func):
|
|||
|
||||
|
||||
def send_completion(
|
||||
model_name, messages, functions, stream, temperature=0, extra_headers=None, max_tokens=None
|
||||
model_name,
|
||||
messages,
|
||||
functions,
|
||||
stream,
|
||||
temperature=0,
|
||||
extra_headers=None,
|
||||
max_tokens=None,
|
||||
):
|
||||
from aider.llm import litellm
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue