mirror of
https://github.com/qodo-ai/pr-agent.git
synced 2025-07-05 05:10:38 +08:00
Add validation for committable comments within PR hunks in GitHub provider
This commit is contained in:
@ -1,5 +1,8 @@
|
|||||||
|
import copy
|
||||||
|
import difflib
|
||||||
import hashlib
|
import hashlib
|
||||||
import itertools
|
import itertools
|
||||||
|
import re
|
||||||
import time
|
import time
|
||||||
import traceback
|
import traceback
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
@ -11,6 +14,7 @@ from retry import retry
|
|||||||
from starlette_context import context
|
from starlette_context import context
|
||||||
|
|
||||||
from ..algo.file_filter import filter_ignored
|
from ..algo.file_filter import filter_ignored
|
||||||
|
from ..algo.git_patch_processing import extract_hunk_headers
|
||||||
from ..algo.language_handler import is_valid_file
|
from ..algo.language_handler import is_valid_file
|
||||||
from ..algo.types import EDIT_TYPE
|
from ..algo.types import EDIT_TYPE
|
||||||
from ..algo.utils import (PRReviewHeader, Range, clip_tokens,
|
from ..algo.utils import (PRReviewHeader, Range, clip_tokens,
|
||||||
@ -415,7 +419,10 @@ class GithubProvider(GitProvider):
|
|||||||
Publishes code suggestions as comments on the PR.
|
Publishes code suggestions as comments on the PR.
|
||||||
"""
|
"""
|
||||||
post_parameters_list = []
|
post_parameters_list = []
|
||||||
for suggestion in code_suggestions:
|
|
||||||
|
code_suggestions_validated = self.validate_comments_inside_hunks(code_suggestions)
|
||||||
|
|
||||||
|
for suggestion in code_suggestions_validated:
|
||||||
body = suggestion['body']
|
body = suggestion['body']
|
||||||
relevant_file = suggestion['relevant_file']
|
relevant_file = suggestion['relevant_file']
|
||||||
relevant_lines_start = suggestion['relevant_lines_start']
|
relevant_lines_start = suggestion['relevant_lines_start']
|
||||||
@ -872,3 +879,100 @@ class GithubProvider(GitProvider):
|
|||||||
|
|
||||||
def calc_pr_statistics(self, pull_request_data: dict):
|
def calc_pr_statistics(self, pull_request_data: dict):
|
||||||
return {}
|
return {}
|
||||||
|
|
||||||
|
def validate_comments_inside_hunks(self, code_suggestions):
|
||||||
|
"""
|
||||||
|
validate that all committable comments are inside PR hunks - this is a must for committable comments in GitHub
|
||||||
|
"""
|
||||||
|
code_suggestions_copy = copy.deepcopy(code_suggestions)
|
||||||
|
diff_files = self.get_diff_files()
|
||||||
|
RE_HUNK_HEADER = re.compile(
|
||||||
|
r"^@@ -(\d+)(?:,(\d+))? \+(\d+)(?:,(\d+))? @@[ ]?(.*)")
|
||||||
|
|
||||||
|
# map file extensions to programming languages
|
||||||
|
language_extension_map_org = get_settings().language_extension_map_org
|
||||||
|
extension_to_language = {}
|
||||||
|
for language, extensions in language_extension_map_org.items():
|
||||||
|
for ext in extensions:
|
||||||
|
extension_to_language[ext] = language
|
||||||
|
for file in diff_files:
|
||||||
|
extension_s = '.' + file.filename.rsplit('.')[-1]
|
||||||
|
language_name = "txt"
|
||||||
|
if extension_s and (extension_s in extension_to_language):
|
||||||
|
language_name = extension_to_language[extension_s]
|
||||||
|
file.language = language_name.lower()
|
||||||
|
|
||||||
|
for suggestion in code_suggestions_copy:
|
||||||
|
try:
|
||||||
|
relevant_file_path = suggestion['relevant_file']
|
||||||
|
for file in diff_files:
|
||||||
|
if file.filename == relevant_file_path:
|
||||||
|
|
||||||
|
# generate on-demand the patches range for the relevant file
|
||||||
|
patch_str = file.patch
|
||||||
|
if not hasattr(file, 'patches_range'):
|
||||||
|
file.patches_range = []
|
||||||
|
patch_lines = patch_str.splitlines()
|
||||||
|
for i, line in enumerate(patch_lines):
|
||||||
|
if line.startswith('@@'):
|
||||||
|
match = RE_HUNK_HEADER.match(line)
|
||||||
|
# identify hunk header
|
||||||
|
if match:
|
||||||
|
section_header, size1, size2, start1, start2 = extract_hunk_headers(match)
|
||||||
|
file.patches_range.append({'start': start2, 'end': start2 + size2 - 1})
|
||||||
|
|
||||||
|
patches_range = file.patches_range
|
||||||
|
comment_start_line = suggestion.get('relevant_lines_start', None)
|
||||||
|
comment_end_line = suggestion.get('relevant_lines_end', None)
|
||||||
|
original_suggestion = suggestion.get('original_suggestion', None) # needed for diff code
|
||||||
|
if not comment_start_line or not comment_end_line or not original_suggestion:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# check if the comment is inside a valid hunk
|
||||||
|
is_valid_hunk = False
|
||||||
|
min_distance = float('inf')
|
||||||
|
patch_range_min = None
|
||||||
|
# find the hunk that contains the comment, or the closest one
|
||||||
|
for i, patch_range in enumerate(patches_range):
|
||||||
|
d1 = comment_start_line - patch_range['start']
|
||||||
|
d2 = patch_range['end'] - comment_end_line
|
||||||
|
if d1 >= 0 and d2 >= 0: # found a valid hunk
|
||||||
|
is_valid_hunk = True
|
||||||
|
min_distance = 0
|
||||||
|
patch_range_min = patch_range
|
||||||
|
break
|
||||||
|
elif d1 * d2 <= 0: # comment is possibly inside the hunk
|
||||||
|
d1_clip = abs(min(0, d1))
|
||||||
|
d2_clip = abs(min(0, d2))
|
||||||
|
d = max(d1_clip, d2_clip)
|
||||||
|
if d < min_distance:
|
||||||
|
patch_range_min = patch_range
|
||||||
|
min_distance = min(min_distance, d)
|
||||||
|
if not is_valid_hunk:
|
||||||
|
if min_distance < 10: # 10 lines - a reasonable distance to consider the comment inside the hunk
|
||||||
|
# make the suggestion non-committable, yet multi line
|
||||||
|
suggestion['relevant_lines_start'] = max(suggestion['relevant_lines_start'], patch_range_min['start'])
|
||||||
|
suggestion['relevant_lines_end'] = min(suggestion['relevant_lines_end'], patch_range_min['end'])
|
||||||
|
body = suggestion['body'].strip()
|
||||||
|
|
||||||
|
# present new diff code in collapsible
|
||||||
|
existing_code = original_suggestion['existing_code'].rstrip() + "\n"
|
||||||
|
improved_code = original_suggestion['improved_code'].rstrip() + "\n"
|
||||||
|
diff = difflib.unified_diff(existing_code.split('\n'),
|
||||||
|
improved_code.split('\n'), n=999)
|
||||||
|
patch_orig = "\n".join(diff)
|
||||||
|
patch = "\n".join(patch_orig.splitlines()[5:]).strip('\n')
|
||||||
|
diff_code = f"\n\n<details><summary>New proposed code:</summary>\n\n```diff\n{patch.rstrip()}\n```"
|
||||||
|
# replace ```suggestion ... ``` with diff_code, using regex:
|
||||||
|
body = re.sub(r'```suggestion.*?```', diff_code, body, flags=re.DOTALL)
|
||||||
|
body += "\n\n</details>"
|
||||||
|
suggestion['body'] = body
|
||||||
|
get_logger().info(f"Comment was moved to a valid hunk, "
|
||||||
|
f"start_line={suggestion['relevant_lines_start']}, end_line={suggestion['relevant_lines_end']}, file={file.filename}")
|
||||||
|
else:
|
||||||
|
get_logger().error(f"Comment is not inside a valid hunk, "
|
||||||
|
f"start_line={suggestion['relevant_lines_start']}, end_line={suggestion['relevant_lines_end']}, file={file.filename}")
|
||||||
|
except Exception as e:
|
||||||
|
get_logger().error(f"Failed to process patch for committable comment, error: {e}")
|
||||||
|
return code_suggestions_copy
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user