Don't want to be too judgemental but does "Self Serve invite link" feature really needs 50 commits, an army of bots, countless nitpicking and 140+ messages? We are not launching Apollo to the Moon here.
I don't know this particular project but seeing threads like this kill any motivation to contribute.
"We are a fast moving start up (even at 3-4 years old), we believe in moving fast and breaking things ... That's why we don't do code reviews or unit tests.. we just edit live running code and restart the server"
Vs
"This one line change needs 4 -5 commits to add feature flags,unit tests, integration tests - all to be reviewed by different teams and wait 1-2 months to be deployed properly to production"
There's stories that bridge the gaps between the two models of explosions and fires and legally actionable fast moving breakages that bring new people into start-ups who put these processes into place and generally shift the culture towards cautiousness. Other industries need to start at point B, or otherwise learn the lessons that get you there _quickly_.
And now top nav breaks to second line on 63% on mobile devices and the submit button is now out of screen. I know these "how hard can it be?" type of changes
I worked for a big tech and may be I was lucky but code wise it was much more tame. You may need to get an army of people to sign off the feature but nobody was scrutinizing my code like that
> Haha, wait until you hear how long and how many people it takes to change the text for a single button at a company like Google
I feel this is a needlessly obtuse statement. I'll explain you why, as I've worked professionally with frontend development. From your comment it seems you don't have that type of context.
The text that is expected to feature in a UI element is a critical factor in cross cutting concerns such as product management and UX design, and it involves things like internationalization and accessibility support. This means that if you change a line of text, it needs to be translated to all supported languages, and the translation needs to meet usability and GUI requirements. This check needs to be done in each and every single locale supported.
I can give you a very concrete example. Once I was tasked with changing a single line of text in a button featured in a dialog. It turns out the french translation ended up being too long that forced line breaks. The UI framework didn't handled those line breaks well and reflowed the whole UI, causing a huge mess. This required new translation requests, but it turned out that the new translations were too vague and ambiguous. Product Managers got involved because the french translation resulted in poor user experience. Ultimately the whole dialog was redesigned.
But to you it's just a text on a single button, isn't it?
I think the problem is the detachment of a developer from what is being done. Many developers want to treat coding like a crative work, while corporations try to make it an automatic job that turns tickets into code. The fact that the person working on a dialog isn't aware that there are translations where things might force a different layout is a proof of a broken system.
I'm not saying that developer should also be UI specialist, that's nonsense. But a developer should know who to talk to regarding advice, and the contact should be easy. The model where a dev doesn't even know that French translation exists is wrong. The correct model is having a dev think "wait, this might affect translation, better send a Slack message to some UI guy and let him know".
Actually, what I said is a pipe dream. The reality is, most devs are average devs, and companies need to optimize processes for that. This results in very official communication that takes years to get anything done, because you can't trust average developer that they'll contact the UI guy. This leads to a lot of frustration from above-average developers, who simply need a different environment to shine (more freedom but also more responsibility)
> Antiwork emerged from Gumroad's mission to automate repetitive tasks. In 2025, we're taking a bold step by open-sourcing our entire suite of tools that helped run and scale Gumroad. We believe in making powerful automation accessible to everyone.
It looks like most of the coderabbit comments were genuinely addressed. Can't easily follow the Cursor ones - they definitely have some work to do on the right presentation / summary folding.
So yeah, it does look like they bring value for genuinely more reliable code.
I feel you, at our workplace someone added some AI Code review thing and it slops the whole PR with useless sequence diagrams and poems (why would I want to read an AI slop poem?) and pages and pages of instructions on how to interact with it.
Our Slack channel is also completely slopped to the point it's mostly bot conversations constantly spamming about review reminders, pull request statuses, and other useless info you can just look up yourself if you need it.
The signal to noise ratio is to the point where I just ignore everything.
My boss is obsessed with copilot PR summaries and they make me want to scream. A 12-line PR with a 40-line summary? And each description is "updated and enhanced <file>... Updated and enhanced <file>"
There's no utility in reading them, which means asking people more questions directly in slack instead, which genuinely makes the entire process slower for everyone. I know for a fact that no developer in my team would struggle to write 4-5 bullet points for any given PR in less than 2 minutes. My boss just genuinely thinks the slop is better (probably because he's not usually the one reading them)
Our Confluence is also rapidly turning into nothing but totally uninformative Claude-generated noise.
It's annoying to be repeatedly called a Luddite for questioning this stuff when the end results can be so obviously wasteful and unhelpful.
> It's annoying to be repeatedly called a Luddite for questioning this stuff
I've been employing the strategy of pushing on the customer and product delivery expectations. Ask some variants of the following in your next team meeting:
"Where are we at with Customer X?"
"How does Bob (the customer's CEO) feel about the project right now?"
"How will this sell us more copies or make the customer pay more money?"
These kinds of questions can rapidly reorient a team if a member with the power to terminate other employees is present. If the most senior management seems apathetic to these questions and you are interested in building an actual career, it might be time to move on. At some point, the person(s) paying for everything need to ask themselves if they intend to conduct a serious business or simply run a daycare for neurotic tech enthusiasts.
"Your PR is missing testing results. Now, with playright pro, enjoy agentic debugging analysis that coordinates with coding agents including claude code, cursor, copilot, and more! Only 19.99!"
This would be best implemented inline with the textbox of the comment form. If you want to give people feedback on their grammar then do it while they are actually writing - not after. Otherwise you just make an already hard to follow thread even more noisy.
Your last link seems to show alot of PRs being rejected because screenshots and test results werent included. I do get your point about the grammar rule being enforced badly, but the link isnt a great example it
This makes me wonder if AI usage will end up part of job position listings, similar to remote days, if it is not already. How much AI agent will you be able to use/be subjected to? Are people looking for this already when job searching?
Some companies are definitely already looking for it - I think as part of my job search, I've run into probably atleast 4 or 5 companies that, on top of having it as a qualification, ask specific questions on the job application about what AI tools you use and how they've made you more productive
I'm by no means a super keen or heavy user, I've barely dabbled really, but if we were hiring at the moment I think I'd ask about it - it's suddenly a huge part of understanding how a person works and what they might be like to interact with. Up there with considering cover letter/email/CV writing style, and commit messages from any take-home or their public repos.
And now when candidates submit take home tests for me to review, half of them are barely functional AI slopfests.
A goddamn staff engineer candidate with 20 years experience submitted an auto generated, broken app, and then tried to pass it off as though it didn't work because our API was down. No, the AI hard coded the wrong URL and he didn't notice (the correct URL was specified in the email containing the test instructions)
I run several job boards, and though there hasn't yet been any postings requiring the candidate's using AI (also not all that likely in the particular field), I am noticing a definite increase in the number of lists with emoji preceding each bullet point, and the use of em dashes. Personally, if I were a job seeker, I'd find that off-putting just as much as if I were presented with the requirement to use AI in my job.
Mathematicians use calculators, and so too do elementary school students, and grocery store clerks, and civil engineers. What each person needs from a calculator can be similar, but would you give a graphing calculator to the store clerk and expect them to be more "productive?"
Admittedly, my metaphor is leaky—and I also can't comment on the participants of the PR—but after reading the comments and the code itself, I'm getting a lot of "here’s a new calculator with a bunch of graphing functions, trigonometric menus, and poem generators—now go do the basic arithmetic you were already doing, but you work for the calculator now" vibes.
Said another way, it took me a lot more time and effort to understand what the bots were saying and if I agreed, than it did for me to formulate my own thoughts and questions.
Like the saying goes, "the best calculator is the one you have with you," and I'd much rather just use my own.
There's going to a wrapper around github PRs to summarize these issues, the mess they created in the first place. BTW this is the same guy which has famously stopped hiring engineers
In principle I don't see a problem with bots looking out for grammar mistakes and typos that could confuse readers (so long as it's less intrusive than this bot), but in this case the bot is just incorrect.
"Good to merge? Test suite passes locally?" is perfectly valid English. You need to make sure that the bot is configured to not insist on arbitrary prescriptivist style guides that nobody cares about.
The problem is these bots dont understand the concept of "goals" or "achieving something". So they cant judge for themselves when they are actually helping (like in correcting grammar where its so bad its hindering communications with other contributors) or being pedantic (like this) because they wont think to consider the difference unless wxplicitly told to do so
And NGI like a human would be given this task and consider the spirit of the law, the goal we want to acheive, and enforce it to reach that. The next token AI doesnt model that. It just predicts the next token, and understanding the spirit of the law does not seem to be in the emergent capabilities of that
I often see comments on Github issues where poor wording makes it difficult to understand what is actually meant. Things like "I reproduced the bug on Linux, then I tried Windows. I can't reproduce it now." Does that mean it's just not reproducible on Windows, or not reproducible at all anymore? Ambiguities like that are especially annoying when it's someone posting a solution to a problem. Sometimes it's because of grammatical errors, sometimes not.
I think LLMs are actually great for catching things like this, and you generally don't need some higher-level understanding about the goals involved to notice the ambiguity. My point wasn't that bots shouldn't be used like this, just that they need to be given the right instructions.
> just that they need to be given the right instructions.
this is the part im talking about. i also think LLMs are very capable at detecting different types and levels or grammar, but they cant decide which ones should be filtered out to meet a certain goal. they need detailed instructions for that, and thats somewhat inefficient and causes issues like this right here.
we have done this song and dance many times with AI. its the bitter lesson: you need a system that learns these things, you cant just give it a rule based engine to patch the things it cant learn. that works in the short term, but leads to a dead end. we need something that has the "common sense" to see when grammar is fine versus hindering communication, and this just isnt there yet. so it needs to be given detailed instructions to do so. which may or may not be sustainable
I agree with you but I think we're talking about different things haha.
It sounds like you mean that curating messages purely to conform to a particular style guide requires context-dependent information and can never be accomplished reliably with some unmodified generic prompt across many different projects.
I'm saying that while this is true, if you ignore grammar guidelines and just look for cases of ambiguous and confusing wording specifically, then this can actually be accomplished reliably with a generic prompt, if you get that prompt right. Not 100% accurately of course, but good enough that it would be beneficial overall in the long run.
Well the maintainers might. The both in (badly) enforcing guidelines they set out. So they might want to police grammar a little bit. And i could see their point if the grammar is so bad it hinders communication. Bit uncomfortably, it was also be a reasonable predictor of those spam PRs put in from people just trying to pad resumes.
But yeah the bot needs to loosen the boundary a bit there
In their defence, the originally-submitted title was edited so it didn't match a string in the PR thread. I've modified it so it does, but that also required removing words from the end.
Don't want to be too judgemental but does "Self Serve invite link" feature really needs 50 commits, an army of bots, countless nitpicking and 140+ messages? We are not launching Apollo to the Moon here.
I don't know this particular project but seeing threads like this kill any motivation to contribute.
Apollo won't ever make it to the Moon if the engineers were flooded with these bot replies XD
These resume's aren't going to write themselves.
Haha, wait until you hear how long and how many people it takes to change the text for a single button at a company like Google
Ngl .. the spectrum is really as wide as:
"We are a fast moving start up (even at 3-4 years old), we believe in moving fast and breaking things ... That's why we don't do code reviews or unit tests.. we just edit live running code and restart the server"
Vs
"This one line change needs 4 -5 commits to add feature flags,unit tests, integration tests - all to be reviewed by different teams and wait 1-2 months to be deployed properly to production"
There's stories that bridge the gaps between the two models of explosions and fires and legally actionable fast moving breakages that bring new people into start-ups who put these processes into place and generally shift the culture towards cautiousness. Other industries need to start at point B, or otherwise learn the lessons that get you there _quickly_.
Probably one person one minute, unless it needs translation in which case potentially unbounded time
And now top nav breaks to second line on 63% on mobile devices and the submit button is now out of screen. I know these "how hard can it be?" type of changes
I worked for a big tech and may be I was lucky but code wise it was much more tame. You may need to get an army of people to sign off the feature but nobody was scrutinizing my code like that
> Haha, wait until you hear how long and how many people it takes to change the text for a single button at a company like Google
I feel this is a needlessly obtuse statement. I'll explain you why, as I've worked professionally with frontend development. From your comment it seems you don't have that type of context.
The text that is expected to feature in a UI element is a critical factor in cross cutting concerns such as product management and UX design, and it involves things like internationalization and accessibility support. This means that if you change a line of text, it needs to be translated to all supported languages, and the translation needs to meet usability and GUI requirements. This check needs to be done in each and every single locale supported.
I can give you a very concrete example. Once I was tasked with changing a single line of text in a button featured in a dialog. It turns out the french translation ended up being too long that forced line breaks. The UI framework didn't handled those line breaks well and reflowed the whole UI, causing a huge mess. This required new translation requests, but it turned out that the new translations were too vague and ambiguous. Product Managers got involved because the french translation resulted in poor user experience. Ultimately the whole dialog was redesigned.
But to you it's just a text on a single button, isn't it?
I think the problem is the detachment of a developer from what is being done. Many developers want to treat coding like a crative work, while corporations try to make it an automatic job that turns tickets into code. The fact that the person working on a dialog isn't aware that there are translations where things might force a different layout is a proof of a broken system.
I'm not saying that developer should also be UI specialist, that's nonsense. But a developer should know who to talk to regarding advice, and the contact should be easy. The model where a dev doesn't even know that French translation exists is wrong. The correct model is having a dev think "wait, this might affect translation, better send a Slack message to some UI guy and let him know".
Actually, what I said is a pipe dream. The reality is, most devs are average devs, and companies need to optimize processes for that. This results in very official communication that takes years to get anything done, because you can't trust average developer that they'll contact the UI guy. This leads to a lot of frustration from above-average developers, who simply need a different environment to shine (more freedom but also more responsibility)
What an absolutely dystopian PR flow.
Shouldn't automation be somewhat useful? All these bot comments — do they really bring more value than they create distractions?
If you look the org, it might make more sense.
> Antiwork emerged from Gumroad's mission to automate repetitive tasks. In 2025, we're taking a bold step by open-sourcing our entire suite of tools that helped run and scale Gumroad. We believe in making powerful automation accessible to everyone.
It looks like most of the coderabbit comments were genuinely addressed. Can't easily follow the Cursor ones - they definitely have some work to do on the right presentation / summary folding.
So yeah, it does look like they bring value for genuinely more reliable code.
The PR thread looks like this to me:
https://www.reddit.com/r/memes/comments/7lkgdk/video_game_hu...
I feel you, at our workplace someone added some AI Code review thing and it slops the whole PR with useless sequence diagrams and poems (why would I want to read an AI slop poem?) and pages and pages of instructions on how to interact with it.
Our Slack channel is also completely slopped to the point it's mostly bot conversations constantly spamming about review reminders, pull request statuses, and other useless info you can just look up yourself if you need it.
The signal to noise ratio is to the point where I just ignore everything.
My boss is obsessed with copilot PR summaries and they make me want to scream. A 12-line PR with a 40-line summary? And each description is "updated and enhanced <file>... Updated and enhanced <file>"
There's no utility in reading them, which means asking people more questions directly in slack instead, which genuinely makes the entire process slower for everyone. I know for a fact that no developer in my team would struggle to write 4-5 bullet points for any given PR in less than 2 minutes. My boss just genuinely thinks the slop is better (probably because he's not usually the one reading them)
Our Confluence is also rapidly turning into nothing but totally uninformative Claude-generated noise.
It's annoying to be repeatedly called a Luddite for questioning this stuff when the end results can be so obviously wasteful and unhelpful.
> It's annoying to be repeatedly called a Luddite for questioning this stuff
I've been employing the strategy of pushing on the customer and product delivery expectations. Ask some variants of the following in your next team meeting:
These kinds of questions can rapidly reorient a team if a member with the power to terminate other employees is present. If the most senior management seems apathetic to these questions and you are interested in building an actual career, it might be time to move on. At some point, the person(s) paying for everything need to ask themselves if they intend to conduct a serious business or simply run a daycare for neurotic tech enthusiasts.You should add a bot to summarize all the messages in those channels!
this picture is missing some ads, it's the logical next step
"Your PR is missing testing results. Now, with playright pro, enjoy agentic debugging analysis that coordinates with coding agents including claude code, cursor, copilot, and more! Only 19.99!"
Wow what an obnoxious and extremely verbose PR flow. Bot overflow
This would be best implemented inline with the textbox of the comment form. If you want to give people feedback on their grammar then do it while they are actually writing - not after. Otherwise you just make an already hard to follow thread even more noisy.
Is this the future of collaborative coding?
When
https://github.com/antiwork/flexile/pull/427#issuecomment-30...
results in
https://github.com/antiwork/flexile/pull/427#issuecomment-30...
More incredible examples where a LLM flags contributors' pull requests because their comments contain minor grammar errors:
https://github.com/antiwork/flexile/pulls?q="our+contributin...
Your last link seems to show alot of PRs being rejected because screenshots and test results werent included. I do get your point about the grammar rule being enforced badly, but the link isnt a great example it
Such future is instrumental in helping me accept and embrace my own mortality.
This makes me wonder if AI usage will end up part of job position listings, similar to remote days, if it is not already. How much AI agent will you be able to use/be subjected to? Are people looking for this already when job searching?
Some companies are definitely already looking for it - I think as part of my job search, I've run into probably atleast 4 or 5 companies that, on top of having it as a qualification, ask specific questions on the job application about what AI tools you use and how they've made you more productive
I'm by no means a super keen or heavy user, I've barely dabbled really, but if we were hiring at the moment I think I'd ask about it - it's suddenly a huge part of understanding how a person works and what they might be like to interact with. Up there with considering cover letter/email/CV writing style, and commit messages from any take-home or their public repos.
And now when candidates submit take home tests for me to review, half of them are barely functional AI slopfests.
A goddamn staff engineer candidate with 20 years experience submitted an auto generated, broken app, and then tried to pass it off as though it didn't work because our API was down. No, the AI hard coded the wrong URL and he didn't notice (the correct URL was specified in the email containing the test instructions)
I want off Mr Altman's wild ride lol
I run several job boards, and though there hasn't yet been any postings requiring the candidate's using AI (also not all that likely in the particular field), I am noticing a definite increase in the number of lists with emoji preceding each bullet point, and the use of em dashes. Personally, if I were a job seeker, I'd find that off-putting just as much as if I were presented with the requirement to use AI in my job.
Oh my god, the fucking emoji bullets
I basically just stop reading on sight
I often use the metaphor of LLMs as calculators.
Mathematicians use calculators, and so too do elementary school students, and grocery store clerks, and civil engineers. What each person needs from a calculator can be similar, but would you give a graphing calculator to the store clerk and expect them to be more "productive?"
Admittedly, my metaphor is leaky—and I also can't comment on the participants of the PR—but after reading the comments and the code itself, I'm getting a lot of "here’s a new calculator with a bunch of graphing functions, trigonometric menus, and poem generators—now go do the basic arithmetic you were already doing, but you work for the calculator now" vibes.
Said another way, it took me a lot more time and effort to understand what the bots were saying and if I agreed, than it did for me to formulate my own thoughts and questions.
Like the saying goes, "the best calculator is the one you have with you," and I'd much rather just use my own.
Unfortinately, LLMs arnt deterministic. So they cannot be consistent. So calculators arnt comparable.
I know there's something to be said for retaining git history when merging, but merging 50 commits is a great way to pollute your commit history.
Some of the commit descriptions: "fix", "fixes", "clean up".
Github will squash all commits into a single one that actually gets merged: https://github.com/antiwork/flexile/commit/123404b3caa52870b...
> # Self-Serve Invite Link Feature — Fixes #348 This feature enables admins to generate an invite link for contractors
How is that self-service?
Shhh don't tell the bots!
How are these bots still a thing? Who actually wants then???
This profession has gone from rewarding and enjoyable to a confusing, miserable hellscape faster than I possibly could have imagined.
Dear bot,
Fuck off.
Dear humans who advocated for installing the bot, let me use anodyne, US corporate bullshit language so you'll understand:
Your bot does not add value. Get rid of it, before it drives out all voluntary contributors.
There's going to a wrapper around github PRs to summarize these issues, the mess they created in the first place. BTW this is the same guy which has famously stopped hiring engineers
Like we needed more ways to stifle contribution to FOSS projects. Where is the GaaS (gatekeeping as a service) unicorn?
Imagine a contributor forgetting to rinse his message through an AI verbalizer. The horror!
In principle I don't see a problem with bots looking out for grammar mistakes and typos that could confuse readers (so long as it's less intrusive than this bot), but in this case the bot is just incorrect.
"Good to merge? Test suite passes locally?" is perfectly valid English. You need to make sure that the bot is configured to not insist on arbitrary prescriptivist style guides that nobody cares about.
The problem is these bots dont understand the concept of "goals" or "achieving something". So they cant judge for themselves when they are actually helping (like in correcting grammar where its so bad its hindering communications with other contributors) or being pedantic (like this) because they wont think to consider the difference unless wxplicitly told to do so
And NGI like a human would be given this task and consider the spirit of the law, the goal we want to acheive, and enforce it to reach that. The next token AI doesnt model that. It just predicts the next token, and understanding the spirit of the law does not seem to be in the emergent capabilities of that
I often see comments on Github issues where poor wording makes it difficult to understand what is actually meant. Things like "I reproduced the bug on Linux, then I tried Windows. I can't reproduce it now." Does that mean it's just not reproducible on Windows, or not reproducible at all anymore? Ambiguities like that are especially annoying when it's someone posting a solution to a problem. Sometimes it's because of grammatical errors, sometimes not.
I think LLMs are actually great for catching things like this, and you generally don't need some higher-level understanding about the goals involved to notice the ambiguity. My point wasn't that bots shouldn't be used like this, just that they need to be given the right instructions.
> just that they need to be given the right instructions.
this is the part im talking about. i also think LLMs are very capable at detecting different types and levels or grammar, but they cant decide which ones should be filtered out to meet a certain goal. they need detailed instructions for that, and thats somewhat inefficient and causes issues like this right here.
we have done this song and dance many times with AI. its the bitter lesson: you need a system that learns these things, you cant just give it a rule based engine to patch the things it cant learn. that works in the short term, but leads to a dead end. we need something that has the "common sense" to see when grammar is fine versus hindering communication, and this just isnt there yet. so it needs to be given detailed instructions to do so. which may or may not be sustainable
I agree with you but I think we're talking about different things haha.
It sounds like you mean that curating messages purely to conform to a particular style guide requires context-dependent information and can never be accomplished reliably with some unmodified generic prompt across many different projects.
I'm saying that while this is true, if you ignore grammar guidelines and just look for cases of ambiguous and confusing wording specifically, then this can actually be accomplished reliably with a generic prompt, if you get that prompt right. Not 100% accurately of course, but good enough that it would be beneficial overall in the long run.
Yeah, no-one wants Grammarly on pull requests.
Well the maintainers might. The both in (badly) enforcing guidelines they set out. So they might want to police grammar a little bit. And i could see their point if the grammar is so bad it hinders communication. Bit uncomfortably, it was also be a reasonable predictor of those spam PRs put in from people just trying to pad resumes.
But yeah the bot needs to loosen the boundary a bit there
> thank you for the clarification! I appreciate you sharing that domain knowledge about the document-signature relationship
> …
> Your expertise about the system's constraints helps provide important context that static analysis tools can't capture.
So much fawning bullshit bloating the message and the token count. I think this might be the thing with LLMs I dislike most.
Suggestion for prompt writers: “Don’t waste tokens. Keep messages succinct and direct.”
commit number n: add tests
Imagine this in perfect French.
Grammar MechaHitler.
The comments of coderabbitai are ... well.
Of arguable quality I would say. The length size limit is arbitrary and the > 0 is ridiculous.where exactly does the bot say that?
I kept reading and reading and the "violation of our guidelines" phrase wasn't appearing, so I got bored
https://github.com/antiwork/flexile/pull/427#issuecomment-30...
It's in the second sentence of linked comment...
You can type Ctrl-F to search for a phrase within a page.
In their defence, the originally-submitted title was edited so it didn't match a string in the PR thread. I've modified it so it does, but that also required removing words from the end.