

deleted by creator
deleted by creator
deleted by creator
May the stress take its toll on his heart.
Whatever he has to tell himself is fine. The important part is he’s learning to realize he fucked up and correcting it. Never thought we’d see that.
I don’t think they necessarily need logical reasoning. Solid enough test cases, automated test plans, and the ability to use trial & error rapidly means that they can throw a bunch of stuff at the wall and release whatever sticks.
I’ve already seen some crazy stuff setup just with a customized model connected to a bunch of ADO pipelines that can shit out reasonably functional code, test, and release it autonomously. It’s front-ended by a chatbot, where the devs can provide a requested tweak in plan English and have their webapp updated in a few minutes. Right now, there’s a manual review/approval process in place, but this is using commodity shit in 2025. Imagine describing that scenario to someone in 2015 and tell me we can accurately predict the limitations there will be in 2035, '45, etc.
I don’t think the industry’s disappearing anytime soon, but I do think we’ll see AI eating up some of the offshore/junior/mid-level work before I get to retire.
The only thing AI will replace is small, standalone scripts/programs.
For now. Eventually, I’d expect LLMs to be better at ingesting the massive existing codebase and taking it into account and either planning an approach or spitting out the first iteration of code. Holding large amounts of a language in memory and adding to it is their whole thing.
Societies and people mean less than nothing to them.