Your position implies that we need to prove that we're not smoking our own supply. I would argue that you are the one who should prove that we're not working (conservatively) 5-8x faster.
The most telling part is when you said "most of the time reading the prompt". That strongly implies that you're attempting to one-shot whatever it is that you're working on.
There is no "the prompt" in my current application. It's a 275k LoC ESP-IDF app spread across ~30 components that interact via FreeRTOS mechanisms as well as an app-wide event bus. It manages non-blocking UI, IO over multiple protocols, drives an OLED using a customized version of lvgl. It is, by any estimation, a serious and non-trivial application, and it was almost entirely crafted by LLM coding models being closely driven by yours truly across several hundred distinct Cursor conversations.
It's probably taken me 10% of the time it would have taken me to do by hand, and that's precisely because I lean on it so heavily for initial buildout, thoughtful troubleshooting (it is never tired, never not available, and also knows more than I do about electronics as a bonus) and the occasional large cross-component refactor.
I don't suspect that you're wrong. I know that you're wrong.
Sorry, that line about "reading the prompt" was unclear, I was referencing people who use it to write emails, project updates, etc. Boilerplate natural language that most people skim over because most of it is fluff. I'm well aware that large scale AI coding requires much more guidance than a single prompt.
Regardless, I would be more charitable if people like you didn't make statements such as, paraphrased, "the burden of proof is on the person suggesting the null hypothesis." I'm open to being convinced, but until now the only real study I've seen for coding shows no gain from AI. The anecdotes I have are mostly things like a coworker's 50kloc project that hasn't seen the light of day because it's massive, unreviewable, and takes as long to test as it would to have for a human to write something equivalent in <<50kloc.
Your project is the first one I've heard of that seems to be successful, and I'm curious to learn more. Just to confirm, your project is not majority boilerplate, functional enough that you'd feel comfortable releasing it to others, and you're able to get bugs fixed without too much trouble? That's pretty rare from what I've seen, and is definitely going at the top of my list of anecdotes in support of AI coding!
We're going to have to agree to disagree; the fact that you're having trouble proving a negative is not our fault or our problem.
Drop your email or some other way to contact you. I will put you on the project mailing list and you'll get a notification when the GitHub project goes public; this will happen when the first beta units go out to the presales buyers.
The most telling part is when you said "most of the time reading the prompt". That strongly implies that you're attempting to one-shot whatever it is that you're working on.
There is no "the prompt" in my current application. It's a 275k LoC ESP-IDF app spread across ~30 components that interact via FreeRTOS mechanisms as well as an app-wide event bus. It manages non-blocking UI, IO over multiple protocols, drives an OLED using a customized version of lvgl. It is, by any estimation, a serious and non-trivial application, and it was almost entirely crafted by LLM coding models being closely driven by yours truly across several hundred distinct Cursor conversations.
It's probably taken me 10% of the time it would have taken me to do by hand, and that's precisely because I lean on it so heavily for initial buildout, thoughtful troubleshooting (it is never tired, never not available, and also knows more than I do about electronics as a bonus) and the occasional large cross-component refactor.
I don't suspect that you're wrong. I know that you're wrong.