From e9254070b182343a421e1dc07f7962e8c07e1bd3 Mon Sep 17 00:00:00 2001 From: paul-gauthier <69695708+paul-gauthier@users.noreply.github.com> Date: Mon, 6 Nov 2023 21:05:12 -0800 Subject: [PATCH] Update benchmarks-1106.md --- docs/benchmarks-1106.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/benchmarks-1106.md b/docs/benchmarks-1106.md index 9e6df71a7..48614e1bb 100644 --- a/docs/benchmarks-1106.md +++ b/docs/benchmarks-1106.md @@ -33,7 +33,7 @@ and a test suite to evaluate whether the coder has correctly solved the problem. The benchmark gives aider two tries to complete the task: -1. On the first try, aider gives GPT the stub code file to edit and the natural language instructions that describe the problem. +1. On the first try, aider gives GPT the stub code file to edit and the natural language instructions that describe the problem. This reflects how you code with aider. You add your source code files to the chat and ask for changes, which are automatically applied. 2. If the test suite fails after the first try, aider gives GPT the test error output and asks it to fix the code. Aider supports this sort of interaction when chatting with GPT using a command like `/run pytest` to run and share pytest results in the chat with GPT. You can `/run` whatever tests/linters/etc make sense for your language/framework/situation. ## Benchmark results on the new "1106" models