fixed tests

This commit is contained in:
Paul Gauthier 2024-05-09 14:35:57 -07:00
parent 728a629789
commit bb93c2a903
2 changed files with 6 additions and 4 deletions

View file

@ -7,7 +7,8 @@
- Refactored prompting to unify some phrasing across edit formats. - Refactored prompting to unify some phrasing across edit formats.
- Enhanced the canned assistant responses used in prompts. - Enhanced the canned assistant responses used in prompts.
- Added explicit model settings for `openrouter/anthropic/claude-3-opus`, `gpt-3.5-turbo` - Added explicit model settings for `openrouter/anthropic/claude-3-opus`, `gpt-3.5-turbo`
- Added `-show-prompts` debug switch. - Added `--show-prompts` debug switch.
- Bugfix: catch and retry on all litellm exceptions.
### v0.33.0 ### v0.33.0

View file

@ -2,8 +2,8 @@ import unittest
from unittest.mock import MagicMock, patch from unittest.mock import MagicMock, patch
import httpx import httpx
import openai
from aider.litellm import litellm
from aider.sendchat import send_with_retries from aider.sendchat import send_with_retries
@ -20,10 +20,11 @@ class TestSendChat(unittest.TestCase):
# Set up the mock to raise # Set up the mock to raise
mock_completion.side_effect = [ mock_completion.side_effect = [
openai.RateLimitError( litellm.exceptions.RateLimitError(
"rate limit exceeded", "rate limit exceeded",
response=mock, response=mock,
body=None, llm_provider="llm_provider",
model="model",
), ),
None, None,
] ]