From 0062f69d9e1d22afcbf98dde283b76576273210b Mon Sep 17 00:00:00 2001 From: paul-gauthier <69695708+paul-gauthier@users.noreply.github.com> Date: Fri, 30 Jun 2023 17:06:24 -0700 Subject: [PATCH] Update benchmarks.md --- docs/benchmarks.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/benchmarks.md b/docs/benchmarks.md index 4039f9d7a..72e9c0939 100644 --- a/docs/benchmarks.md +++ b/docs/benchmarks.md @@ -40,7 +40,7 @@ This produced some interesting observations: - The new June (`0613`) versions of `gpt-3.5-turbo` are worse at code editing than the older Feb (`0301`) version. This was unexpected. - The GPT-4 models are much better at code editing than the GPT-3.5 models. This was expected. -These results agree with an intuition that I've been +The quantitative benchmark results agree with an intuition that I've been developing about how to prompt GPT for complex tasks like coding. You want to minimize the "cognitive overhead" of formatting the response, so that GPT can focus on the task at hand.