From 5e99c51d936d89a5c3df331f8bdd6dcb55a9d200 Mon Sep 17 00:00:00 2001 From: Paul Gauthier Date: Tue, 26 Nov 2024 06:01:46 -0800 Subject: [PATCH] copy --- aider/website/_posts/2024-11-21-quantization.md | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/aider/website/_posts/2024-11-21-quantization.md b/aider/website/_posts/2024-11-21-quantization.md index efba3066b..303e01b9e 100644 --- a/aider/website/_posts/2024-11-21-quantization.md +++ b/aider/website/_posts/2024-11-21-quantization.md @@ -45,6 +45,11 @@ and silently discards data that exceeds it. - Output token limits -- Open source models are often served with wildly differing output token limits. This has a direct impact on how much code the model can write or edit in a response. +- Buggy cloud providers -- Between Qwen and DeepSeep, there were +multiple cloud providers with broken or buggy API endpoints that seemed +to be returning result different from expected based on the advertised +quantization and context sizes. + ### Sections {: .no_toc }