Hacker Newsnew | past | comments | ask | show | jobs | submit | codexon's commentslogin

You can replace it with a much lower paid employee though.


A lower paid and less qualified employee won’t be able to spot when the AI screws up.

Having a higher-paid, qualified employee supervising multiple AIs as the human only needs to spot for mistakes - maybe.


I'm not sure that's entirely true. For most things, checking if a solution is correct is much easier than implementing it (page looks wrong, can't login etc...)


You’re looking at the end result, I’m looking at implementation. Engineering management, not QA.


You definitely cannot. Code org, architecture, and system design are senior level roles and responsibilities.


AI is already aware of the best practices. It does not just blindly do what you ask of it in the simplest way.


Best practices are always situation dependent.


Claude code will prompt you and explain to you what practice fits a situation. It might not do it perfectly, but the foundations are there.


That's not growth. Growth is having the existing employee do more.


I'm not arguing about growth. I was addressing this statement which seems to presume that AI has no effect if the job can't be removed.

> If AI can't do 100% of a job then you can't remove the job.


I think some of it might be genuine. For people that don't code (like management), going from 0 to being able to create a landing page that looks like it came from a big corporation is a miracle.

They are not able to comprehend that for anything more complicated than that, the code might compile, but the logical errors and failure to implement the specs start piling up.


This paper creates a new benchmark comprised of real remote work tasks sourced from the remote working website Upwork. The best commercial LLMs like Opus, GPT, Gemini, and Grok were tested.

Models released a few days ago, Opus 4.6 and GPT 5.3, haven't been tested yet, but given the performance on other micro-benchmarks, they will probably not be much different on this benchmark.


They didn't test Opus at all, only Sonnet.

One of the tasks was "Build an interactive dashboard for exploring data from the World Happiness Report." -- I can't imagine how Opus4.5 could've failed that.


Check the link to the study. It has been updated for Opus 4.5.


Even for coding, it seems to still make A LOT of mistakes.

https://youtu.be/8brENzmq1pE?t=1544

I feel like everyone is counting chickens before they hatch here with all the doomsday predictions and extrapolating LLM capability into infinity.

People that seem to overhype this seem to either be non-technical or are just making landing pages.


Waiting until the moment they get good enough is not a smart thing to do either. If you are a farmer and know it is going to snow, at some point in the next 5 months, you make plans NOW, you don't wait until the temperatures drop and you see the snow falling. Right now, people are waiting for the snowfall before moving their proverbial chickens indoors


Top AI researchers like Yann LeCunn have said that LLMs are a dead end.

It seems to me that LLM performance is plateuing and not improving exponentially anymore. This recent hubbub about rewriting a worse GCC for $20,000 is another example of overhype and regurgitating training data.

You don't know for sure if it is going to "snow" (AI reaches general intelligence) Snow happens frequently, AI reaching general intelligence has never happened. If it ever happens, 99% of jobs are gone and there is really nothing you can do to prepare for this other than maybe buy guns and ammo, and even that might not do anything to robotic soldiers.

People were worried about AI taking their jobs 60 years ago when perceptrons came out, and anyone who avoided a tech career because of that back then would have lost out majorly.


There is no reason why an AI model capable of pushing a significant chunk of devs into lower paid and highly competitive dev jobs as a result of automation needs to be a general artificial intelligence. There is a lack of nuance that comes with thinking that either AI is dumb or it has human level general intelligence. As much as devs hate to admit it, you don't need that much of what we understand as general intelligence to write software. Only a portion of your intelligence is needed and arguably not all of it at the same time.

While general purpose models might be plateauing soon (arguably they have for a while). Highly specialised models (especially for programming) haven't necessarily plateaud yet. And anyway, existing functionality seem like a good foundation to build upon systems that remove the need of hiring as many devs. It's not the "being out of a job" that should worry you. Open up your binary thinking and consider that facing a 08 job market for the rest of your career is not the same permanent unemployment but it is not a market you would like to have.

That is the real concern.


You don't need to be a genius or rocket scientist to write code, but llm don't even reach the bar for anything but the most simple things. Take a look at the video I posted earlier for an example.

And specialised models for programming HAVE plateaued.

https://livebench.ai/#/?sort=Agentic+Coding+Average

From Claude 4.1 to 4.5 was only an 18% gain, and from 4.5 to 4.6 it even DECLINED. Codex 5.1 to 5.2 also shows a decline.


https://arxiv.org/abs/2510.26787

Testing the top llms on wework, the highest performing one only succeeded with a rate of 2.5%

Can you imagine not being fired when you can only do 2.5% of all tasks?

This study is dated October 30th, very recent.


> Can you imagine not being fired when you can only do 2.5% of all tasks?

You are not competing against LLMs though. You are competing against people (who in a pre-LLM world wouldn't be in tech) using LLMs tools to beat you in terms of value. In the new world, you either are a top 1% dev or you beat everyone in race to the bottom pricewise. The middle will become vanishingly small. Think of manufacturing in developed countries.


Didn't this paper demonstrate that you only need 1.58 bits to be equivalent to 16 bits in performance?

https://arxiv.org/abs/2402.17764


This technique showed that there are ways during training to optimize weights to neatly quantize while remaining performant. This isn't a post training quantization like int4.


For Kimi quantization is part of the training also. Specifically they say they use QAT, quantization aware training.

That doesn't mean training with all integer math, but certain tricks are used to specifically plan for the end weight size. I.e. fake quantization nodes are inserted to simulate int4.


Iirc the paper was solid, but it still hasn’t been adopted/proven out at large scale. Harder to adapt hardware and code kernels to something like this compared to int4.


just call it one trit


Has anyone noticed a lot of polymarket posts on their X (formerly known as twitter) feed claiming to be making a fortune? It makes me feel like its some kind of coordinated guerilla marketing scheme.


I find it interesting that there's so much pushback against ai generated art and music while there seems to be very little for ai generated code.


Perhaps that's because there's an enormous difference between fine art and computer programs.

Also, there's quite a lot of pushback against AI-generated code, but also because unlike music, normal people have no interest in and aren't aware of the code.


They are obviously different things, but aren't the people who spent thousands of hours honing their coding and releasing their code spending just as much time and effort if not more than the people who made non-ai images and music?


There certainly are. If you by chance imply equivalence based on time and effort spent, neither is the thing that differentiates arts, crafts and other activities.


I won't merge anything AI generated in any of my FOSS projects, unless I'm successfully deceived.

In the first place, I do not regard a copyright notice and license on AI generated code to be valid in my eyes, so on those grounds alone, I cannot use it any more than I could merge a piece of proprietary, leaked source code.


The copyright office agreed with you about the non-copyrightability of AI generated media so in that sense you can safely ignore copyright claims on anything AI-generated.


You can safely ignore claims made by the operator of the AI who had the material generated, true enough.

You cannot ignore a credible/plausible claim from a third party.

It's not simply the case that the output has a public-domain-like status.


Music is art, code is engineering. "Hackers and painters"[1] was always wishful fluff, unfortunately.

When it comes to code, I don't think anyone cares how the sausage is made, and only very rarely do people care by whom. The only question is "does it work well?"

Art is totally different. Provenance is much more important - sometimes essential. David is a beautiful work, but you could 3d print or cast a replica of "David". No one would pretend that the copy is the same as the original though - even if they're indistinguishable to the untrained eye - because one was painstakingly hand sculpted and the others were cheaply produced. This sense of provenance is the property that NFTs were (unsuccessfully) trying to capture.

[1] https://www.paulgraham.com/hp.html


If someone painstakingly hand sculpted an exact replica of "David", does it make it art, or a forgery? Is hand written code to produce generative art not art?

It's difficult to pin down the line. Ultimately it's up to the individual to define them. "The relationship to art, and this kind of painting, to their work, varied with the person entirely."[1]

[1]: https://news.berkeley.edu/2025/03/31/berkeley-voices-transfo...


> If someone painstakingly hand sculpted an exact replica of "David", does it make it art, or a forgery? Is hand written code to produce generative art not art?

No and no.

If you raise and teach a child and they generate a painting, are you the artist?


Musicians and artists are under pressure to make money, but they cant rush it

while programmers have to rush it these days or they lose their jobs. Programmers don't have much of say in their companies.


Devs are quite used to using others peoples work for free via packages, frameworks and entire operating systems and IDE’s. It’s just part of the culture.

Music has its history in IP, royalties, and most things need to be paid for in the creation of music or art itself.

It’s going to be much easier for devs to accept AI when remixing code is such a huge part of the culture already. The expectation in the arts is entirely different.


This doesn't make sense to me. I mean, the term "remix" literally comes from the music scene.

Artists are constantly getting inspiration from one another, referencing one another, performing together or having their works exhibited together...

While there are some big name artists who are famously protective of the concept of IP, those artists have made headlines exactly because when they litigate they seem so unreasonable compared to the bedroom musicians and pub bands and church choirs and school teachers and wedding DJs and millions of other artists and performers whose way of participating in "the culture" is much less tied to ownership.


Most code people interact with are creations shat out by soulless corporations, why would they care? Being honest here, the vast majority of people have their code experience dictated by less than a handful of companies; at their jobs they are told to use these tools or get file for welfare. The animosity has been baked into the industry for quite a while, it's only very very recently that the masses have been able to interact with open source code and even that is getting torn down by big tech.

Compare this to music where you are free to choose and listen to whatever you want, or stare at art that moves you. IF you don

At work most people are force to deal with code like SalesForce or MSFT garbage, not the same experience at all.

Why would people care about code coming from an industry that has been bleeding them dry and making their society worse for nearly 20+ years?


What???

Every thread on HN that touches on the topic has countless people talking about how LLM generated code is always bad, buggy and people that utilize them are inexperienced juniors that don't understand anything.

And they're not completely wrong. If you don't know what you're doing, you'll absolutely create dumster fires instead of software


Sure, I am one of the people who will say that. But where are the people calling for it to be banned? Where are the stores and websites that are banning AI generated software?


I feel like part of the difference is how art vs code is viewed. You could make the argument code is art, though most don't have that stance. Visual art and music tend to be made by a few people, there is ego involved, you care who the artist is. Code tends to be made by shops and consumers don't know who the coders are. Programmers are already faceless.

I think it's also about money. Places code and code samples are stored tend to be large companies that are in tech and on the AI hype wagon. Bandcamp is not one of those places.


There's one popular platform that requires disclosing whether and how AI was used (Steam), and if you search anything about it, all you can find is like a sea of articles opposing it.



Yes and it was probably only done because of people complaining about AI art, not AI code.


Really? You've not seen the numerous open source projects banning AI-generated PRs with extreme prejudice?


That's not really the same as stores outright banning AI code.

An apt analogy would be like a shared drawing taking merge requests and having to spend 30 minutes looking at every single merge request zoomed in to see if there was a microscopic phallus embedded somewhere.

It is completely fair for an open source project to have their own standards, and you are also free to fork it so you can accept as many AI PRs as you want.

None of these options are available for someone that wants to sell AI generated music. There are really only 2 marketplaces to sell your own music and if both of them banned AI, then you are effectively locked out of the entire market.


Positing that AI generated code is always bad and buggy is delusional.

I have dozens of little programs and websites that are AI generated and do their job perfectly.


I think a key factor there is that programmers (in the actual sense, rather than so-called “vibe coders”) are more likely on average than (current) artists and musicians to have intimate knowledge of how AI works and what AI can and can't do well — and consequently, the quality of their output is high enough that it's harder to notice the use of AI.

Eventually that'll change, as artists and musicians continue to experiment with AI and come up with novel uses for it, just as digital artists did with tablets and digital painting software, and just as musicians did with keyboards and DAWs.


AI music from suno sounds indistinguishable to non-ai generated music to me.

In terms of how well it works, the quality of AI music is far better than art or code. In art there are noticeble glitches like multiple fingers. For code, it can call non existent functions, not do what it is supposed to do, or have security issues or memory leaks. From what I can tell, there is no such deal breaker for AI music.


The tells in music are there. The most common being: vocals have a subtle constant hiss to them, voices and instruments sound different in the second half than they did in the first, the hiss filter gets more prominent and affects all instruments towards the end of the song, auditory artifacts like volume jumps or random notes/noises near transitions.

More subjective tells: drums are hissy and weak, lyrics are generic or weird like "Went to the grocery store to buy coffee beans for my sadness", weirdly uniform loudness and density from start to finish, drops/climaxes are underwhelming, and (if you've listened to enough of them) a general uncanny feel to them.

I've generated about 70 hours of AI music and have listened to all of the songs at least once, so it's become intuitive for me to pick them out.

Some examples for listening for the hiss filter:

https://suno.com/s/qvUKLxVV6HDifknq (Easiest to hear at 0:00 with the inhale)

https://suno.com/s/QZx1t0aii0HVZYGx (Really strong at 0:09)

Some examples for more hiss and other (subjective) tells like weak drums:

https://suno.com/s/tTYygsVFo88SX6OV

https://suno.com/s/CzFgC6dxSQLWyGSn


> For code, it can call non existent functions, not do what it is supposed to do, or have security issues or memory leaks.

I guess what I'm getting at is that, since programmers are typically more inclined than the average person to understand how AI works, programmers are therefore ahead of the curve when it comes to understanding those pitfalls and structuring their workflows to minimize them — to play to the strengths and weaknesses of LLMs. A “fancy” autocomplete v. a “fancy” linter v. something pretending to be a junior programmer are all going to have very different rates of success.

The issue hindering art and music is that most people using generative AI for art and music are doing so analogously to the “something pretending to be a junior programmer” role instead of the “fancy autocomplete” or “fancy linter” roles. That is: they're typically using AI to generate works end-to-end, whereas (non-vibe-coder) programmers are typically using AI in far narrower scopes, with more direct control over the final output. I think the quality of AI-based art and music will improve as more narrowly-scoped AI-driven workflows catch on among actually-skilled artists and musicians — and the result will be works that are very different from existing works, rather than works that only cheaply imitate some statistical average of existing works.


Companies sell products built on code, not the code itself. Code is a means to an end.


The pushback is motivated by the interests of the petty-bourgeois class, and those are a larger proportion of the former.


Imagine being a Marxist and not respecting the craft and labor required for art production. Couldn't be me.


Actually many networks will prefer routing over a cheap AS path no matter how long it is.


> > and is (one of many attributes) used to judge the quality of route


Lower cost usually means lower quality and is an example of how a long path being leaked can result in traffic flowing away from high quality path to the leaked path.

Not saying that this is the case with Venezuela, just explaining the reality of BGP where path prepends are often ignored.


His claim-- as best as I can read it-- is that B leaking a long-length route changes where traffic is routed, but not to B.

It's possible he's saying something else, but I can't figure out, and he hasn't clarified.


I've actually tried hetzner on and off with 1 server for the past 2 years and keep running into downtime every few months.

First I used an ex101 with an i9-13900. Within a week it just froze. It could not be reset remotely. Nothing in kern.log. Support offered no solution but a hard reboot. No mention of what might be wrong other than user error.

A few months later, one of the drives just disconnects from raid by itself. It took support 1 hour to respond and they said they found no issue so it must be my fault.

Then I changed to a ryzen based server and it also mysteriously had problems like this. Again the support blamed the user.

It was only after I cancelled the server and several months later that I see this so I know it isn't just me.

https://docs.hetzner.com/robot/dedicated-server/general-info...


Why wouldn't microsoft advertise this though? If they had the ability to take the attack and others might not, then it'll result in more customers for them.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: