-
-
Notifications
You must be signed in to change notification settings - Fork 932
generative-ai: Add anti pattern example #1679
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 5 commits
55e0a9e
2a102b8
7a7bec2
f0e60d0
131f62f
f7d40f8
3b1b75c
fdc2c42
65f3bbe
cf1642e
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -24,3 +24,11 @@ Unacceptable uses | |
| Maintainers may close issues and PRs that are not useful or productive, including | ||
| those that are fully generated by AI. If a contributor repeatedly opens unproductive | ||
| issues or PRs, they may be blocked. | ||
|
|
||
| Anti-patterns | ||
|
||
| ============= | ||
| - While AI-assisted tools such as autocompletion can enhance productivity, they sometimes rewrite entire code blocks instead of making small, focused edits. | ||
|
||
| This can make it more difficult to review changes and to fully understand both the original intent of the code and the rationale behind the new modifications. | ||
| Maintaining consistency with the original code helps preserve clarity, traceability, and meaningful reviews and also helps us avoid unnecessary code churn. | ||
| - Sometimes AI assisted tools make failing unit tests pass by altering or bypassing the tests rather than addressing the underlying problem in the code. | ||
| Such changes do not represent a real fix and should be avoided. | ||
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do we have a second anti-pattern to include in the list? Other bad things: hallucinations, incorrectly changing tests instead of fixing code, ...
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
i think you covered the incorrectly changing tests instead of fixing one. Good call. I like having that in the list.
I don't personally see that so often anymore with Opus 4.1 or Sonnet 4.5, but older or lower end models tend to reward hack their way into that state more often - it's a thing we'll likely encounter coming from new contributors for a while as they can't be assumed to be using the latest and greatest.