I have no unique perspective to add other than an obvious question: If the PR is low quality, why not just close/reject it? Does it matter if it's AI assisted or not?
If AI writes a for loop the same way you would... Does it automatically mean the code is bad because you—or someone you approve of—didn't write it? What is the actual argument being made here? All code has trade offs, does AI make a bad cost/benefit analysis? Hell yeah it does. Do humans make the same mistake? I can tell you for certain they do, because at least half of my career was spent fixing those mistakes... Before there ever was an LLM in sight. So again... What's the argument here? AI can produce more code, so like more possibility for fuck up? Well, don't vibe code with "approve everything" like what are we even talking about? It's not the tool it's the users, and as with any tool theres going to be misuse, especially new and emerging ones lol
> A 19k lines-of-code Pull Request was opened in January, 2026.
Such a PR should be rejected simply because of the shear size of it, regardless of AI use. Seriously, who submits a 19k line PR? Just make many small ones.
The PR touched a lot of internals, including module code and mirrors the fs APIs. So, yes it was big, but the commit history was largely clean and followed a development story, and it was tested. The code quality was decent too. I didn't review all of it because I don't have a personal stake in this though.
I suggest EVERYONE in this thread go read the the GitHub PR in question. There's some good arguments for and against AI, and what it means for FOSS... But good lord you will have to sift through the virtue signalling bullshit and have patience for the constant moving of goalposts
How would you go about breaking up this particular set of functionality into smaller PRs, exactly? It's meant to introduce a virtualized file system... the size is dictated by the feature itself.
Also, no mention at all regarding the test coverage, or impact if any on existing code paths specifically.
Well, survivorship bias means that Elixir is loudly populated by AI maximalists now. Just go look at the last several years worth of US/EU Elixirconf talks schedules, it's maybe a third of each cohort and included in keynote slots.
This is how I would deal with the problem if I maintained node: "Please, use your tokens and experimental energies to port to Rust and pass the following test suite. Let us know when you've got something that works."
This is a silly reactionary response. Where is the line? Can I use AI to look up APIs? Write documentation? What if I write a function and ask AI to test it? What if I manually implemented an idea that I thought about after chatting with AI a few weeks ago?
Stop treating this like it's going to go away. We need actual solutions for the FOSS community that make reviewing AI assisted work tractable.
On the other hand, I haven’t and I believe many of us, have never paid node any money so it feels weird to dictate their approach.
Think carefully before responding.
Such a PR should be rejected simply because of the shear size of it, regardless of AI use. Seriously, who submits a 19k line PR? Just make many small ones.
I suggest EVERYONE in this thread go read the the GitHub PR in question. There's some good arguments for and against AI, and what it means for FOSS... But good lord you will have to sift through the virtue signalling bullshit and have patience for the constant moving of goalposts
Also, no mention at all regarding the test coverage, or impact if any on existing code paths specifically.
@indutny explains their views in that thread.
Stop treating this like it's going to go away. We need actual solutions for the FOSS community that make reviewing AI assisted work tractable.