Skip to the content.

GPT code editing benchmarks

benchmark results

Aider is an open source command line chat tool that lets you work with GPT to edit code in your local git repo. To do this, aider needs to be able to reliably recognize when GPT wants to edit local files, determine which files it wants to modify and what changes to save. Such automated code editing hinges on using the system prompt to tell GPT how to structure code edits in its responses.

Aider currently asks GPT to use simple text based “edit formats”, but OpenAI’s new function calling API looks like a promising way to create more structured edit formats. After implementing a couple of function based edit formats, I wanted to measure the potential benefits of switching aider to use them by default.

With this in mind, I developed a benchmark based on the Exercism python coding exercises. This benchmark evaluates how effectively aider and GPT can translate a natural language coding request into executable code saved into files that pass unit tests. It provides an end-to-end evaluation of not just GPT’s coding ability, but also its capacity to edit existing code and format those code edits so that aider can save the edits to the local source files.

I ran the benchmark on all the ChatGPT models (except gpt-4-32k), using a variety of edit formats. The results were interesting:

The quantitative benchmark results agree with my intuitions about prompting GPT for complex tasks like coding. It’s beneficial to minimize the “cognitive overhead” of formatting the response, allowing GPT to concentrate on the coding task at hand.

As a thought experiment, imagine a slack conversation with a junior developer where you ask them to write the code to add some new feature to your app. They’re going to type the response back to you by hand in the chat. Should they type out the code and wrap it in a normal markdown code block? Or should they type up a properly escaped and syntactically correct json data structure that contains the text of the new code?

Using more complex output formats with GPT seems to cause two issues:

I was expecting to start using function call based edits in aider for both GPT-3.5 and GPT-4. But given these benchmark results, I won’t be adopting the functions API at this time. I will certainly plan to benchmark functions again with future versions of the models.

More details on the benchmark, edit formats and results are discussed below.

The benchmark

The benchmark uses 133 practice exercises from the Exercism python repository. These exercises were designed to help individuals learn Python and hone their coding skills.

Each exercise includes:

The goal is for GPT to read the instructions, implement the provided function/class skeletons and pass all the unit tests. The benchmark measures what percentage of the 133 exercises are completed successfully, causing all the associated unit tests to pass.

To start each exercise, aider sends GPT the initial contents of the implementation file, the Exercism instructions and a final instruction:

Use the above instructions to modify the supplied files: <implementation file>
Keep and implement the existing function or class stubs, they will be called from unit tests.
Only use standard python libraries, don't suggest installing any packages.

Aider updates the implementation file based on GPT’s reply and runs the unit tests. If all tests pass, the exercise is considered complete. If some tests fail, aider sends GPT a second message with the test error output. It only sends the first 50 lines of test errors to try and avoid exceeding the context window of the smaller models. Aider also includes this final instruction:

See the testing errors above.
The tests are correct.
Fix the code in <implementation file> to resolve the errors.

Requiring GPT to fix its first implementation in response to test failures is another way in which this benchmark stresses code editing skill. This second chance is also important because it gives GPT the opportunity to adjust if the instructions were imprecise with respect to the specific requirements of the unit tests. Many of the exercises have multiple paragraphs of instructions, and most human coders would likely fail some tests on their first try.

The bars in the graph show the percent of exercises that were completed by each model and edit format combination. The full bar height represents the final outcome following both coding attempts. Each bar also has a horizontal mark that shows the intermediate performance after the first coding attempt, without the benefit of the second try that includes the test error output.

It’s worth noting that GPT never gets to see the source code of the unit tests during the benchmark. It only sees the error output from failed tests. Of course, all of this code was probably part of its original training data!

In summary, passing an exercise means GPT was able to:

Conversely, failing an exercise only requires a breakdown in one of those steps. In practice, GPT fails at different steps in different exercises. Sometimes it simply writes the wrong code. Other times, it fails to format the code edits in a way that conforms to the edit format, resulting in the code not being saved correctly.

It’s worth keeping in mind that changing the edit format often affects both aspects of GPT’s performance. Complex edit formats often lead GPT to write worse code and make it less successful at formatting the edits correctly.

Edit formats

I benchmarked 4 different edit formats, described below. Each description includes a sample response that GPT might provide to a user who requests: “Change the print from hello to goodbye.”

whole

The whole format asks GPT to return an updated copy of the entire file, including any changes. The file should be formatted with normal markdown triple-backtick fences, inlined with the rest of its response text.

This format is very similar to how ChatGPT returns code snippets during normal chats, except with the addition of a filename right before the opening triple-backticks.

Here is the updated copy of your file demo.py:

demo.py
```python
def main():
    print("goodbye")
```

diff

The diff format also asks GPT to return edits as part of the normal response text, in a simple diff format. Each edit is a fenced code block that specifies the filename and a chunk of ORIGINAL and UPDATED code. GPT provides some original lines from the file and then a new updated set of lines.

Here are the changes you requested to demo.py:

```python
demo.py
<<<<<<< ORIGINAL
    print("hello")
=======
    print("goodbye")
>>>>>>> UPDATED
```

whole-func

The whole-func format requests updated copies of whole files to be returned using the function call API.

{
    "explanation": "Changed hello to goodbye.",
    "files": [
        {
            "path": "demo.py",
            "content": "def main():\n    print(\"goodbye\")\n"
        }
}

diff-func

The diff-func format requests a list of original/updated style edits to be returned using the function call API.

{
    "explanation": "Changed hello to goodbye.",
    "edits": [
        {
            "path": "demo.py",
            "original_lines": [
                "    print(\"hello\")"
            ],
            "updated_lines": [
                "    print(\"goodbye\")"
            ],
        }
    ]
}

GPT-3.5’s performance

The 0613 models seem worse?

The GPT-3.5 benchmark results have me fairly convinced that the new gpt-3.5-turbo-0613 and gpt-3.5-16k-0613 models are a bit worse at code editing than the older gpt-3.5-turbo-0301 model.

This is visible in the “first attempt” portion of each result, before GPT gets a second chance to edit the code. Look at the horizontal white line in the middle of the first three blue bars. Performance with the whole edit format was 46% for the February model and only 39% for the June models.

But also note how much the solid green diff bars degrade between the February and June GPT-3.5 models. They drop from 30% down to about 19%.

I saw other signs of this degraded performance in earlier versions of the benchmark as well.

Pathological use of diff

When GPT-3.5 is able to correctly generate the diff edit format, it often uses it in a pathological manner. It places the entire original source file in the ORIGINAL block and the entire updated file in the UPDATED block. This is strictly worse than just using the whole edit format, as GPT is sending two full copies of the file.

Hallucinated function calls

When GPT-3.5 uses the functions API it is prone to ignoring the JSON Schema that specifies valid functions. It often returns a completely novel and semantically invalid function_call fragment with "name": "python".

The arguments attribute is supposed to be a set of key/value pairs with the arguments to the function specified in the name field. Instead, GPT-3.5 frequently just stuffs an entire python file into that field.

        "function_call": {
          "name": "python",
          "arguments": "def main():\n    print(\"hello\")\n"
        },

It seems like it might be getting confused by fine-tuning that was done for the ChatGPT code interpreter plugin?

Randomness

The benchmark attempts to be deterministic, always sending identical requests for each exercise on repeated runs. As part of this effort, when sending test error output to GPT, it removes the wall-clock timing information that is normally included by the unittest module.

The benchmark harness also logs SHA hashes of all the OpenAI API requests and replies. This makes it possible to detect randomness or nondeterminism in the benchmarking process.

It turns out that the OpenAI chat APIs are not deterministic, even at temperature=0. The same identical request will produce multiple distinct responses, usually less than 5-10 variations. This suggests that OpenAI may be load balancing their API across a number of slightly different instances of the model?

For certain exercises, some of these variable responses pass the unit tests while other variants do not. Results for exercises like this, which are “on the bubble”, are therefore a bit random, depending on which variant OpenAI returns.

Given that, it would be ideal to run all 133 exercises many times for each model/edit-format combination and report an average performance. This would average away the effect of the API variance. It would also significantly increase the cost of this sort of benchmarking. So I didn’t do that.

Benchmarking against 133 exercises already provides some robustness, since we are measuring the performance across many exercises.

But to get a sense of how much the API variance impacts the benchmark outcomes, I ran all 133 exercises 10 times each against gpt-3.5-turbo-0613 with the whole edit format. You’ll see one set of error bars in the graph, which show the range of results from those 10 runs.

The OpenAI API randomness doesn’t seem to cause a large variance in the overall benchmark results.

Conclusions

Based on these benchmark results, aider will continue to use the whole edit format for GPT-3.5, and diff for GPT-4.

GPT-4 gets comparable results with the whole and diff edit formats, but using whole significantly increases costs and latency compared to diff.

The latency of streaming back the entire updated copy of each edited file is a real challenge with the whole format. The GPT-3.5 models are quite responsive, and can stream back entire files at reasonable speed. Aider displays a progress bar and live diffs of the files as they stream in, which helps pass the time.

The GPT-4 models are much slower, and waiting for even small files to be completely “retyped” on each request is probably unacceptable.