mirror of
https://github.com/Aider-AI/aider.git
synced 2025-06-01 10:14:59 +00:00
copy
This commit is contained in:
parent
757eac0579
commit
73de0ea8be
2 changed files with 24 additions and 2 deletions
|
@ -181,4 +181,27 @@
|
||||||
date: 2024-11-24
|
date: 2024-11-24
|
||||||
versions: 0.64.2.dev
|
versions: 0.64.2.dev
|
||||||
seconds_per_case: 17.5
|
seconds_per_case: 17.5
|
||||||
total_cost: 0.0000
|
total_cost: 0.0000
|
||||||
|
|
||||||
|
- dirname: 2024-11-24-04-12-58--fireworks-qwen-diff
|
||||||
|
test_cases: 133
|
||||||
|
model: Fireworks (unknown quant)
|
||||||
|
edit_format: diff
|
||||||
|
commit_hash: 757eac0
|
||||||
|
pass_rate_1: 57.9
|
||||||
|
pass_rate_2: 72.2
|
||||||
|
percent_cases_well_formed: 94.0
|
||||||
|
error_outputs: 23
|
||||||
|
num_malformed_responses: 19
|
||||||
|
num_with_malformed_responses: 8
|
||||||
|
user_asks: 8
|
||||||
|
lazy_comments: 0
|
||||||
|
syntax_errors: 6
|
||||||
|
indentation_errors: 0
|
||||||
|
exhausted_context_windows: 4
|
||||||
|
test_timeouts: 1
|
||||||
|
command: aider --model fireworks_ai/accounts/fireworks/models/qwen2p5-coder-32b-instruct
|
||||||
|
date: 2024-11-24
|
||||||
|
versions: 0.64.2.dev
|
||||||
|
seconds_per_case: 10.4
|
||||||
|
total_cost: 0.5759
|
|
@ -28,7 +28,6 @@ The graph above compares different versions of the Qwen 2.5 Coder 32B Instruct m
|
||||||
served both locally and from cloud providers.
|
served both locally and from cloud providers.
|
||||||
|
|
||||||
- The [HuggingFace BF16 weights](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) served via [glhf.chat](https://glhf.chat).
|
- The [HuggingFace BF16 weights](https://huggingface.co/Qwen/Qwen2.5-Coder-32B-Instruct) served via [glhf.chat](https://glhf.chat).
|
||||||
- Hyperbolic labs API for [qwen2-5-coder-32b-instruct](https://app.hyperbolic.xyz/models/qwen2-5-coder-32b-instruct), which is using BF16. This result is probably within the expected variance of the HF result.
|
|
||||||
- [4bit and 8bit quants for mlx](https://t.co/cwX3DYX35D).
|
- [4bit and 8bit quants for mlx](https://t.co/cwX3DYX35D).
|
||||||
- The results from [OpenRouter's mix of providers](https://openrouter.ai/qwen/qwen-2.5-coder-32b-instruct/providers) which serve the model with different levels of quantization.
|
- The results from [OpenRouter's mix of providers](https://openrouter.ai/qwen/qwen-2.5-coder-32b-instruct/providers) which serve the model with different levels of quantization.
|
||||||
- Ollama locally serving different quantizations from the [Ollama model library](https://ollama.com/library/qwen2.5-coder:32b-instruct-q4_K_M).
|
- Ollama locally serving different quantizations from the [Ollama model library](https://ollama.com/library/qwen2.5-coder:32b-instruct-q4_K_M).
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue