> If you haven't seen coding agents produce working code you've not been paying attention for the past 3-12 months.
If you believe coding agents produce working code, why was the decision below made?
Amazon orders 90-day reset after code mishaps cause
millions of lost orders[0]
0 - https://www.businessinsider.com/amazon-tightens-code-control...
Good journalism would include : https://www.aboutamazon.com/news/company-news/amazon-outage-...
I find it somewhat overblown.
Also, I think there's a difference between working code and exceptionally bug-free code. Humans produce bugs all the time. I know I do at least.
> Good journalism would include ...
The link you provided begins with the declaration:
I am not a journalist and even I would question the "good journalism would include" assertion given the source provided.> I find it somewhat overblown.
As I quoted in a peer comment:
If the above is "overblown", then the SVP has done so. I have no evidence to believe this is the case however.Do you?
> I am not a journalist and even I would question the "good journalism would include" assertion given the source provided.
You've misunderstood. I was saying good journalism would include both sides, and hopefully primary sources alongside the reporting, so readers can evaluate both.
> If the above is "overblown", then the SVP has done so. I have no evidence to believe this is the case however.
It says "at least one of those disruptions were tied to Amazon's AI coding assistant Q, while others exposed deeper issues." You initially cited this article as evidence that coding agents don't produce working code. But the SVP is describing a broader trend of deployment and control plane failures, most of which are classic infrastructure problems that predate AI tooling entirely. You're attributing a systemic operational failure to AI code generation when even your own source doesn't support that.
More fundamentally, your original argument was that the premise "software can write working code" is flawed. One company having incidents, where some of those incidents involved AI tooling doesn't prove that. Humans cause production incidents every single day. By your logic, the existence of any bug would prove humans can't write working code either.
You appear to be confusing "produce working code" with "exclusively produce working code".
> You appear to be confusing "produce working code" with "exclusively produce working code".
The confusion is not mine own. From the article cited:
It appears to me that "Amazon's SVP of e-commerce services" desires producing working code and has identified the ramifications of not producing same.That's why I'm writing a guide about how to use this stuff to produce good code.
> That's why I'm writing a guide about how to use this stuff to produce good code.
Consider the halting problem[0]:
Essentially, it identifies that mathematics cannot prove an arbitrary program will or will not terminate based on the input given to it. So if math cannot express a solution to this conundrum, how can any mathematical algorithm generate solutions to arbitrary problems which can be trusted to complete (a.k.a. "halt")?Put another way, we all know "1 + 2 = 3" since elementary school. Basic math assumed everyone knows.
Imagine an environment where "1 + 2" 99% of the time results in "3", but may throw a `DivisionByZeroException`, return NaN[1], or rewrite the equation to be "PI x r x r".
Why would anyone trust that environment to reliably do what they instructed it to do?
0 - https://en.wikipedia.org/wiki/Halting_problem
1 - https://en.wikipedia.org/wiki/NaN
I find the challenge of using LLMs to usefully write software despite their non-deterministic nature to be interesting and deserving of study.
I get the appeal and respect the study you are engaging.
A meta-question I posit is; at what point does the investment in trying to get "LLMs to usefully write software despite their non-deterministic nature" become more than solving the problems at hand without using those tools?
For the purpose of the aforementioned, please assume commercial use as opposed to academic research.