mirror of
https://github.com/codeflash-ai/codeflash.git
synced 2026-05-04 18:25:17 +00:00
Remove docstrings and all imports before testing for zero diffs.
This commit is contained in:
parent
ddb087760e
commit
07d00ae279
3 changed files with 617 additions and 12 deletions
577
code_to_optimize/bubble_sort_typed.py
Normal file
577
code_to_optimize/bubble_sort_typed.py
Normal file
|
|
@ -0,0 +1,577 @@
|
|||
from __future__ import annotations
|
||||
|
||||
from typing import Any, Callable, Iterable, NewType, Optional, Protocol, TypeVar
|
||||
|
||||
try:
|
||||
from typing import _TypingBase # type: ignore[attr-defined]
|
||||
except ImportError:
|
||||
from typing import _Final as _TypingBase # type: ignore[attr-defined]
|
||||
typing_base = _TypingBase
|
||||
|
||||
_T = TypeVar("_T")
|
||||
|
||||
|
||||
class Comparable(Protocol):
|
||||
def __lt__(self: _T, __other: _T) -> bool: ...
|
||||
|
||||
|
||||
ComparableT = TypeVar("ComparableT", bound=Comparable)
|
||||
|
||||
|
||||
def sorter(arr: list[ComparableT]) -> list[ComparableT]:
|
||||
for i in range(len(arr)):
|
||||
for j in range(len(arr) - 1):
|
||||
if arr[j] > arr[j + 1]:
|
||||
temp = arr[j]
|
||||
arr[j] = arr[j + 1]
|
||||
arr[j + 1] = temp
|
||||
return arr
|
||||
|
||||
|
||||
def sorter2(arr: list[ComparableT]) -> list[ComparableT]:
|
||||
n = len(arr)
|
||||
for i in range(n):
|
||||
swapped = False
|
||||
for j in range(n - i - 1):
|
||||
if arr[j] > arr[j + 1]:
|
||||
arr[j], arr[j + 1] = arr[j + 1], arr[j]
|
||||
swapped = True
|
||||
if not swapped:
|
||||
break
|
||||
return arr
|
||||
|
||||
|
||||
def sorter3(arr: list[ComparableT]) -> list[ComparableT]:
|
||||
arr.sort()
|
||||
return arr
|
||||
|
||||
|
||||
def is_valid_field_name(name: str) -> bool:
|
||||
return not name.startswith("_")
|
||||
|
||||
|
||||
def is_valid_field_name2(name: str) -> bool:
|
||||
return not (name and name[0] == "_")
|
||||
|
||||
|
||||
def is_self_type(tp: Any) -> bool:
|
||||
"""Check if a given class is a Self type (from `typing` or `typing_extensions`)"""
|
||||
return isinstance(tp, typing_base) and getattr(tp, "_name", None) == "Self"
|
||||
|
||||
|
||||
def is_self_type2(tp: Any) -> bool:
|
||||
"""Check if a given class is a Self type (from `typing` or `typing_extensions`)"""
|
||||
if not isinstance(tp, _TypingBase):
|
||||
return False
|
||||
return tp._name == "Self" if hasattr(tp, "_name") else False
|
||||
|
||||
|
||||
test_new_type = NewType("test_new_type", str)
|
||||
|
||||
|
||||
def is_new_type(type_: type[Any]) -> bool:
|
||||
"""Check whether type_ was created using typing.NewType.
|
||||
Can't use isinstance because it fails <3.10.
|
||||
"""
|
||||
return isinstance(type_, test_new_type.__class__) and hasattr(type_, "__supertype__") # type: ignore[arg-type]
|
||||
|
||||
|
||||
def is_new_type2(type_: type[Any]) -> bool:
|
||||
"""Check whether type_ was created using typing.NewType.
|
||||
Can't use isinstance because it fails <3.10.
|
||||
"""
|
||||
return type(type_) is type(test_new_type) and hasattr(type_, "__supertype__")
|
||||
|
||||
|
||||
def _to_str(
|
||||
size: int,
|
||||
suffixes: Iterable[str],
|
||||
base: int,
|
||||
*,
|
||||
precision: Optional[int] = 1,
|
||||
separator: Optional[str] = " ",
|
||||
) -> str:
|
||||
if size == 1:
|
||||
return "1 byte"
|
||||
elif size < base:
|
||||
return f"{size:,} bytes"
|
||||
|
||||
for i, suffix in enumerate(suffixes, 2): # noqa: B007
|
||||
unit = base**i
|
||||
if size < unit:
|
||||
break
|
||||
return "{:,.{precision}f}{separator}{}".format(
|
||||
(base * size / unit),
|
||||
suffix,
|
||||
precision=precision,
|
||||
separator=separator,
|
||||
)
|
||||
|
||||
|
||||
# Given: (size=-1, suffixes=(), base=-1, precision=0, separator=None),
|
||||
# code_to_optimize.bubble_sort_typed._to_str : raises UnboundLocalError("cannot access local variable 'unit' where it is not associated with a value")
|
||||
# code_to_optimize.bubble_sort_typed._to_str2 : raises IndexError()
|
||||
|
||||
|
||||
def _to_str2(
|
||||
size: int,
|
||||
suffixes: Iterable[str],
|
||||
base: int,
|
||||
*,
|
||||
precision: Optional[int] = 1,
|
||||
separator: Optional[str] = " ",
|
||||
) -> str:
|
||||
if size == 1:
|
||||
return "1 byte"
|
||||
elif size < base:
|
||||
return f"{size:,} bytes"
|
||||
|
||||
unit = base
|
||||
for suffix in suffixes:
|
||||
unit *= base
|
||||
if size < unit:
|
||||
return f"{size / (unit / base):,.{precision}f}{separator}{suffix}"
|
||||
|
||||
# Extra condition if size exceeds the largest unit
|
||||
return f"{size / (unit / base):,.{precision}f}{separator}{suffixes[-1]}"
|
||||
|
||||
|
||||
def find_common_tags(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = articles[0]["tags"]
|
||||
for article in articles[1:]:
|
||||
common_tags = [tag for tag in common_tags if tag in article["tags"]]
|
||||
return set(common_tags)
|
||||
|
||||
|
||||
# crosshair diffbehavior --max_uninteresting_iterations 64 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags2
|
||||
# Given: (articles=[{'tags': ['', '']}, {'tags': ['', '']}, {'tags': []}, {}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : returns set()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags2 : raises KeyError()
|
||||
|
||||
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article["tags"])
|
||||
return common_tags
|
||||
|
||||
|
||||
# Given: (articles=[{'\x00\x00\x00\x00': [], 'tags': ['']}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {'tags': ['']}, {}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : raises KeyError()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags2_1 : returns set()
|
||||
|
||||
|
||||
def find_common_tags2_1(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0].get("tags", []))
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article.get("tags", []))
|
||||
return common_tags
|
||||
|
||||
|
||||
# % crosshair diffbehavior --max_uninteresting_iterations 64 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags2_2
|
||||
# Given: (articles=[{'\x00\x00\x00\x00': [''], 'tags': ['']}, {'\x00\x00\x00\x00': [''], 'tags': ['']}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {'\x00\x00\x00\x00': [], '': []}, {'\x00\x00\x00\x00': [], 'tags': ['']}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : raises KeyError()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags2_2 : returns set()
|
||||
# (codeflash312) renaud@Renauds-Laptop codeflash %
|
||||
|
||||
|
||||
def find_common_tags2_2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
if not common_tags:
|
||||
break
|
||||
common_tags.intersection_update(article["tags"])
|
||||
return common_tags
|
||||
|
||||
|
||||
# % crosshair diffbehavior --max_uninteresting_iterations 128 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags2_3
|
||||
# Given: (articles=[{'tags': ['', '']}, {'tags': ['', '']}, {'tags': []}, {}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : returns set()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags2_3 : raises KeyError()
|
||||
# Given: (articles=[{'\x00\x00\x00\x00': [], 'tags': []}, {'\x00\x00\x00\x00': [], 'tags': []}, {'\x00\x00\x00\x00': [], 'tags': []}, {'\x00\x00\x00\x00': []}, {}, {}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : returns set()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags2_3 : raises KeyError()
|
||||
|
||||
|
||||
def find_common_tags2_3(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
article_tags = article["tags"] # Access 'tags' key to match KeyError behavior
|
||||
if not common_tags:
|
||||
continue # Skip intersection but maintain KeyError on missing 'tags'
|
||||
common_tags.intersection_update(article_tags)
|
||||
return common_tags
|
||||
|
||||
|
||||
def find_common_tags2_4(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
if common_tags:
|
||||
article_tags = article["tags"] # Access 'tags' only if common_tags is not empty
|
||||
common_tags.intersection_update(article_tags)
|
||||
else:
|
||||
# Do not access article["tags"]; no KeyError is raised
|
||||
pass
|
||||
return common_tags
|
||||
|
||||
|
||||
def find_common_tags2_5(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
# Initialize with the first article's tags, defaulting to an empty list if "tags" is missing
|
||||
common_tags = set(articles[0].get("tags", []))
|
||||
|
||||
for article in articles[1:]:
|
||||
# Use .get("tags", []) to safely access tags, defaulting to an empty list if missing
|
||||
common_tags.intersection_update(article.get("tags", []))
|
||||
|
||||
# Early exit if there are no common tags left
|
||||
if not common_tags:
|
||||
break
|
||||
|
||||
return common_tags
|
||||
|
||||
|
||||
def find_common_tags2_6(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
# Initialize with the first article's tags
|
||||
common_tags = set(articles[0]["tags"]) # Raises KeyError if "tags" is missing
|
||||
|
||||
for article in articles[1:]:
|
||||
# Directly access "tags", maintaining behavior
|
||||
common_tags.intersection_update(article["tags"])
|
||||
|
||||
# Early exit if no common tags remain
|
||||
if not common_tags:
|
||||
break
|
||||
|
||||
return common_tags
|
||||
|
||||
|
||||
def find_common_tags2_7(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
# Initialize with the first article's tags (raises KeyError if "tags" is missing)
|
||||
common_tags = set(articles[0]["tags"])
|
||||
|
||||
for article in articles[1:]:
|
||||
if not common_tags:
|
||||
# If no common tags remain, no need to process further
|
||||
break
|
||||
|
||||
# Access "tags" directly, maintaining original behavior (raises KeyError if missing)
|
||||
common_tags.intersection_update(article["tags"])
|
||||
|
||||
return common_tags
|
||||
|
||||
|
||||
def find_common_tags2_8(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
# Initialize with the first article's tags (raises KeyError if "tags" is missing)
|
||||
try:
|
||||
common_tags = set(articles[0]["tags"])
|
||||
except KeyError:
|
||||
raise KeyError("The first article is missing the 'tags' key.")
|
||||
|
||||
for index, article in enumerate(articles[1:], start=2):
|
||||
try:
|
||||
tags = article["tags"]
|
||||
except KeyError:
|
||||
raise KeyError(f"Article at position {index} is missing the 'tags' key.")
|
||||
|
||||
# Perform intersection with the current article's tags
|
||||
common_tags.intersection_update(tags)
|
||||
|
||||
return common_tags
|
||||
|
||||
|
||||
def find_common_tags2_9(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
# Initialize with the first article's tags (raises KeyError if "tags" is missing)
|
||||
common_tags = set(articles[0]["tags"])
|
||||
|
||||
for article in articles[1:]:
|
||||
if not common_tags:
|
||||
# If no common tags remain, no need to process further
|
||||
break
|
||||
# Directly access "tags", allowing KeyError to propagate naturally
|
||||
common_tags.intersection_update(article["tags"])
|
||||
|
||||
return common_tags
|
||||
|
||||
|
||||
# crosshair diffbehavior --max_uninteresting_iterations 64 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags3
|
||||
# Given: (articles=[{'tags': ['', '', '', '']}, {'tags': ['', '', '', '']}, {'tags': ['', '', '']}, {'tags': ['', '', '', '']}, {'tags': ['', '', '']}, {}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : raises KeyError()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags3 : returns set()
|
||||
# Given: (articles=[{'\x00\x00\x00\x00': ['', ''], 'tags': [], '': []}, {}, {'\x00\x00\x00\x00': ['', ''], '': []}, {'': []}, {'\x00\x00\x00\x00': ['', ''], 'tags': [], '': []}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : returns set()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags3 : raises KeyError()
|
||||
|
||||
|
||||
def find_common_tags3(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article["tags"])
|
||||
if not common_tags:
|
||||
break
|
||||
return common_tags
|
||||
|
||||
|
||||
# % crosshair diffbehavior --max_uninteresting_iterations 64 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags4
|
||||
# Given: (articles=[{'\x00\x00\x00\x00': ['', ''], 'tags': [], '': []}, {}, {'\x00\x00\x00\x00': ['', ''], '': []}, {'': []}, {'\x00\x00\x00\x00': ['', ''], 'tags': [], '': []}]),
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags : returns set()
|
||||
# code_to_optimize.bubble_sort_typed.find_common_tags4 : raises KeyError()
|
||||
|
||||
|
||||
def find_common_tags4(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
common_tags &= set(article["tags"])
|
||||
if not common_tags: # Early exit if no common tags.
|
||||
break
|
||||
return common_tags
|
||||
|
||||
|
||||
def with_pattern(pattern: str, regex_group_count: int | None = None) -> Callable:
|
||||
def decorator(func: Callable) -> Callable:
|
||||
func.pattern = pattern
|
||||
func.regex_group_count = regex_group_count
|
||||
return func
|
||||
|
||||
return decorator
|
||||
|
||||
|
||||
def with_pattern2(pattern: str, regex_group_count: int | None = None) -> Callable:
|
||||
return (
|
||||
lambda func: setattr(func, "pattern", pattern)
|
||||
or setattr(func, "regex_group_count", regex_group_count)
|
||||
or func
|
||||
)
|
||||
|
||||
|
||||
"""
|
||||
We have this original code:
|
||||
```
|
||||
def find_common_tags(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = articles[0]["tags"]
|
||||
for article in articles[1:]:
|
||||
common_tags = [tag for tag in common_tags if tag in article["tags"]]
|
||||
return set(common_tags)
|
||||
```
|
||||
|
||||
We generated optimized code:
|
||||
```
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article["tags"])
|
||||
return common_tags
|
||||
```
|
||||
|
||||
When analyzed the optimized code for correctness with Concolic Testing (using CrossHair DiffBehavior), we found the following counter-example:
|
||||
```
|
||||
crosshair diffbehavior --max_uninteresting_iterations 64 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags2
|
||||
|
||||
Given: (articles=[{'tags': ['', '']}, {'tags': ['', '']}, {'tags': []}, {}]), code_to_optimize.bubble_sort_typed.find_common_tags : returns set() code_to_optimize.bubble_sort_typed.find_common_tags2 : raises KeyError()
|
||||
```
|
||||
|
||||
Fix find_common_tags2 for correctness relative to find_common_tags, making sure it is still optimized for speed.
|
||||
"""
|
||||
|
||||
"""
|
||||
To fix the `KeyError` in `find_common_tags2` and ensure it behaves like the original `find_common_tags`, we'll modify the code to handle cases where the `'tags'` key might be missing in some articles. In the original code, it seems that articles might not always have the `'tags'` key, and in such cases, the code should default to an empty list.
|
||||
|
||||
Here's the corrected and optimized version:
|
||||
|
||||
```python
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0].get("tags", []))
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article.get("tags", []))
|
||||
return common_tags
|
||||
```
|
||||
|
||||
**Explanation:**
|
||||
|
||||
- **Use `get` with a default value:**
|
||||
- We replaced `article["tags"]` with `article.get("tags", [])`. This ensures that if the `'tags'` key is missing, it defaults to an empty list instead of raising a `KeyError`.
|
||||
- **Initialize `common_tags` properly:**
|
||||
- We initialize `common_tags` with `set(articles[0].get("tags", []))` to handle cases where the first article might not have `'tags'`.
|
||||
- **Maintain Performance Optimization:**
|
||||
- By using `set` operations, we keep the optimized performance of the function while ensuring correctness.
|
||||
|
||||
This change ensures that the function doesn't raise a `KeyError` when an article is missing the `'tags'` key, and it behaves identically to the original function for all inputs.
|
||||
"""
|
||||
|
||||
"""We have this original code:
|
||||
```
|
||||
def find_common_tags(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = articles[0]["tags"]
|
||||
for article in articles[1:]:
|
||||
common_tags = [tag for tag in common_tags if tag in article["tags"]]
|
||||
return set(common_tags)
|
||||
```
|
||||
|
||||
We generated optimized code:
|
||||
```
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article["tags"])
|
||||
return common_tags
|
||||
```
|
||||
|
||||
When analyzed the optimized code for correctness with Concolic Testing (using CrossHair DiffBehavior), we found the following counter-example:
|
||||
```
|
||||
crosshair diffbehavior --max_uninteresting_iterations 64 code_to_optimize.bubble_sort_typed.find_common_tags code_to_optimize.bubble_sort_typed.find_common_tags2
|
||||
|
||||
Given: (articles=[{'tags': ['', '']}, {'tags': ['', '']}, {'tags': []}, {}]), code_to_optimize.bubble_sort_typed.find_common_tags : returns set() code_to_optimize.bubble_sort_typed.find_common_tags2 : raises KeyError()
|
||||
```
|
||||
|
||||
We attempted to fix this with the following candidate:
|
||||
```python
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0].get("tags", []))
|
||||
for article in articles[1:]:
|
||||
common_tags.intersection_update(article.get("tags", []))
|
||||
return common_tags
|
||||
```
|
||||
When analyzing the optimized code for correctness with Concolic Testing (using CrossHair DiffBehavior), we found a different counter-example:
|
||||
```
|
||||
Given: (articles=[{'\x00\x00\x00\x00': [], 'tags': ['']}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {'tags': ['']}, {}, {'\x00\x00\x00\x00': [], 'tags': ['']}, {}]),
|
||||
|
||||
code_to_optimize.bubble_sort_typed.find_common_tags : raises KeyError()
|
||||
code_to_optimize.bubble_sort_typed.find_common_tags2_1 : returns set()
|
||||
```
|
||||
|
||||
Fix find_common_tags2_1 for correctness relative to find_common_tags, making sure it is still optimized for speed.
|
||||
"""
|
||||
"""To fix `find_common_tags2_1` and ensure it behaves identically to `find_common_tags`, we need to address the discrepancy in how the two functions handle the absence of the `'tags'` key and when `common_tags` becomes empty.
|
||||
|
||||
**Issue Analysis:**
|
||||
|
||||
1. **KeyError Discrepancy:**
|
||||
- The original function raises a `KeyError` when accessing `article["tags"]` if the `'tags'` key is missing.
|
||||
- Our optimized version used `article.get("tags", [])`, which avoids the `KeyError`, leading to different behavior.
|
||||
|
||||
2. **Empty `common_tags`:**
|
||||
- In the original function, if `common_tags` becomes empty, the list comprehension doesn't access `article["tags"]`, preventing a `KeyError`.
|
||||
- In the optimized version, `intersection_update` still accesses `article["tags"]` even if `common_tags` is empty, leading to unnecessary `KeyError`s.
|
||||
|
||||
**Solution:**
|
||||
|
||||
- **Mimic Original Exception Behavior:**
|
||||
- We should ensure that `find_common_tags2` raises a `KeyError` in the same situations as `find_common_tags`.
|
||||
- Avoid accessing `article["tags"]` when `common_tags` is empty, just like the original function.
|
||||
|
||||
- **Optimize by Early Exit:**
|
||||
- If `common_tags` becomes empty, further intersections won't change it.
|
||||
- We can break out of the loop early, improving performance.
|
||||
|
||||
**Corrected and Optimized Version:**
|
||||
|
||||
```python
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
if not common_tags:
|
||||
break # No need to continue if common_tags is empty
|
||||
common_tags.intersection_update(article["tags"])
|
||||
return common_tags
|
||||
```
|
||||
|
||||
**Explanation:**
|
||||
|
||||
- **Accessing `article["tags"]`:**
|
||||
- We use `article["tags"]` instead of `article.get("tags", [])` to ensure a `KeyError` is raised if `'tags'` is missing, matching the original function's behavior.
|
||||
|
||||
- **Early Exit on Empty `common_tags`:**
|
||||
- The `if not common_tags: break` statement checks if `common_tags` is empty.
|
||||
- If it is empty, we break out of the loop, avoiding unnecessary accesses to `article["tags"]` that could raise a `KeyError`.
|
||||
|
||||
- **Performance Optimization:**
|
||||
- By breaking early when `common_tags` is empty, we prevent further iterations and accesses.
|
||||
- Using `set` operations with `intersection_update` retains the performance benefits.
|
||||
|
||||
**Validation with Counter-Examples:**
|
||||
|
||||
- **First Counter-Example:**
|
||||
- With `articles=[{'tags': ['', '']}, {'tags': ['', '']}, {'tags': []}, {}]`, the original function returns `set()` without raising `KeyError`.
|
||||
- Our corrected `find_common_tags2` also returns `set()` without raising `KeyError` because it avoids accessing `article["tags"]` when `common_tags` is empty.
|
||||
|
||||
- **Second Counter-Example:**
|
||||
- With an input where an article lacks `'tags'` and `common_tags` is not empty, the original function raises `KeyError`.
|
||||
- Our corrected `find_common_tags2` also raises `KeyError` in this situation, matching the original function's behavior.
|
||||
|
||||
**Conclusion:**
|
||||
|
||||
This corrected version ensures that `find_common_tags2` behaves identically to `find_common_tags` in all cases, raising `KeyError` when the original function does, and avoiding it otherwise. It also maintains the performance optimizations by using set operations and early loop exit.
|
||||
|
||||
**Final Optimized Code:**
|
||||
|
||||
```python
|
||||
def find_common_tags2(articles: list[dict[str, list[str]]]) -> set[str]:
|
||||
if not articles:
|
||||
return set()
|
||||
|
||||
common_tags = set(articles[0]["tags"])
|
||||
for article in articles[1:]:
|
||||
if not common_tags:
|
||||
break
|
||||
common_tags.intersection_update(article["tags"])
|
||||
return common_tags
|
||||
```"""
|
||||
|
|
@ -1,10 +1,9 @@
|
|||
from __future__ import annotations
|
||||
|
||||
import ast
|
||||
from typing import IO, TYPE_CHECKING
|
||||
from typing import TYPE_CHECKING, TypeVar
|
||||
|
||||
import libcst as cst
|
||||
from libcst import FunctionDef
|
||||
|
||||
from codeflash.code_utils.code_extractor import add_needed_imports_from_module
|
||||
from codeflash.discovery.functions_to_optimize import FunctionParent
|
||||
|
|
@ -12,6 +11,27 @@ from codeflash.discovery.functions_to_optimize import FunctionParent
|
|||
if TYPE_CHECKING:
|
||||
from pathlib import Path
|
||||
|
||||
from libcst import FunctionDef
|
||||
|
||||
ASTNodeT = TypeVar("ASTNodeT", bound=ast.AST)
|
||||
|
||||
|
||||
def normalize_node(node: ASTNodeT) -> ASTNodeT:
|
||||
if isinstance(
|
||||
node,
|
||||
(ast.Module, ast.FunctionDef, ast.AsyncFunctionDef, ast.ClassDef),
|
||||
) and ast.get_docstring(node):
|
||||
node.body = node.body[1:]
|
||||
if hasattr(node, "body"):
|
||||
node.body = [
|
||||
normalize_node(node) for node in node.body if not isinstance(node, (ast.Import, ast.ImportFrom))
|
||||
]
|
||||
return node
|
||||
|
||||
|
||||
def normalize_code(code: str) -> str:
|
||||
return ast.unparse(normalize_node(ast.parse(code)))
|
||||
|
||||
|
||||
class OptimFunctionCollector(cst.CSTVisitor):
|
||||
METADATA_DEPENDENCIES = (cst.metadata.ParentNodeProvider,)
|
||||
|
|
@ -171,7 +191,7 @@ def replace_functions_in_file(
|
|||
|
||||
module = cst.metadata.MetadataWrapper(cst.parse_module(optimized_code))
|
||||
|
||||
for i, (function_name, class_name) in enumerate(parsed_function_names):
|
||||
for function_name, class_name in parsed_function_names:
|
||||
visitor = OptimFunctionCollector(
|
||||
function_name,
|
||||
class_name,
|
||||
|
|
@ -243,7 +263,6 @@ def replace_function_definitions_in_module(
|
|||
:param project_root_path:
|
||||
:return:
|
||||
"""
|
||||
file: IO[str]
|
||||
source_code: str = module_abspath.read_text(encoding="utf8")
|
||||
new_code: str = replace_functions_and_add_imports(
|
||||
source_code,
|
||||
|
|
@ -262,10 +281,4 @@ def replace_function_definitions_in_module(
|
|||
|
||||
|
||||
def is_zero_diff(original_code: str, new_code: str) -> bool:
|
||||
def normalize_for_diff(tree: ast.Module) -> ast.Module:
|
||||
tree.body = [node for node in tree.body if not isinstance(node, (ast.Import, ast.ImportFrom))]
|
||||
return tree
|
||||
|
||||
original_code_unparsed = ast.unparse(normalize_for_diff(ast.parse(original_code)))
|
||||
new_code_unparsed = ast.unparse(normalize_for_diff(ast.parse(new_code)))
|
||||
return original_code_unparsed == new_code_unparsed
|
||||
return normalize_code(original_code) == normalize_code(new_code)
|
||||
|
|
|
|||
|
|
@ -1602,9 +1602,24 @@ def functionA():
|
|||
assert is_zero_diff(original_code, optim_code_c)
|
||||
|
||||
optim_code_d = """from __future__ import annotations
|
||||
|
||||
|
||||
import numpy as np
|
||||
def functionA():
|
||||
return np.array([1, 2, 3, 4])
|
||||
"""
|
||||
assert not is_zero_diff(original_code, optim_code_d)
|
||||
|
||||
optim_code_e = '''"""
|
||||
Zis a Docstring?
|
||||
"""
|
||||
from __future__ import annotations
|
||||
|
||||
import ast
|
||||
def functionA():
|
||||
"""
|
||||
Und Zis?
|
||||
"""
|
||||
import numpy as np
|
||||
return np.array([1, 2, 3])
|
||||
'''
|
||||
assert is_zero_diff(original_code, optim_code_e)
|
||||
|
|
|
|||
Loading…
Reference in a new issue