diff --git a/docs/docs/finetuning_benchmark/index.md b/docs/docs/finetuning_benchmark/index.md
index b33deae3..6305b732 100644
--- a/docs/docs/finetuning_benchmark/index.md
+++ b/docs/docs/finetuning_benchmark/index.md
@@ -1,8 +1,8 @@
# PR-Agent Code Fine-tuning Benchmark
-On coding tasks, the gap between open-source models and top closed-source models such as GPT4 is significant.
+On coding tasks, the gap between open-source models and top closed-source models such as GPT4 is significant.
-In practice, open-source models are unsuitable for most real-world code tasks, and require further fine-tuning to produce acceptable results.
+In practice, open-source models are unsuitable for most real-world code tasks, and require further fine-tuning to produce acceptable results.
_PR-Agent fine-tuning benchmark_ aims to benchmark open-source models on their ability to be fine-tuned for a code task.
Specifically, we chose to fine-tune open-source models on the task of analyzing a pull request, and providing useful feedback and code suggestions.
@@ -11,6 +11,8 @@ Here are the results:
+**Model performance:**
+
| Model name | Model size [B] | Better than gpt-4 rate, after fine-tuning [%] |
|-----------------------------|----------------|----------------------------------------------|
| **DeepSeek 34B-instruct** | **34** | **40.7** |
@@ -55,7 +57,6 @@ Here are the results: