I remember Travis Kalanik spouting the talking points about self-driving in 2017, that after Waymo, Tesla had the advantage because they had the best data, that they were going to crack self-driving soon. Then I remember Dara scuttling Uber’s entire self driving division in 2019.
Self-driving is possible but it requires a massive sustained investment in custom hardware on the car, in real and simulation testing, in painstaking software developlment covering tens of thousands of scenarios, realtime remote control failsafes, fleet management capabilities in every city. Waymo is the only company that comes close to the right approach. All these other Elons, GM, Uber CEOs are just jangling shiny objects in front of investors. A moonshot on the financial model for what are otherwise mature stagnant businesses.
I recall a lot of the original funding (Billions of dollars) was spent on 3rd party consultants to run various multi-year long review processes (environmental, legal compliance, eminent domain, community agreements etc.)
at this level they just decide and spin up a swat team to execute it in a couple weeks without politicking. the bureaucratic ways, reviews are just for the low levels, to keep them busy with feature scraps while they mostly do operations
yup, I think there are few public articles on aws mantle so you can look it up, but internally this is pretty common knowledge. The entire inference engine of bedrock is built and maintained by a handful of ec2 engineers (all principals and above). Judging by the commit history of the project they are able to just build independent of any of the traditional bureaucracy.
the way in which Mantle was built is highlighted internally by PEs as some sort of triumph but really its a fairly tone deaf indicted of AWS’s engineering culture ... “To achieve a meaningful result in a reasonable amount of time we had to break nearly ever constraint that we force all other engineers to work under. good luck to you plebs of L6 and below”
I'm going to play devil's advocate here: _what if_ most of lower level engineers are actually not able to self-organize themselves like this dedicated, I bet hand-picked group of PE can? I'm pretty sure AWS ruthless culture would gladly use less middle-managers and be swifter in time to market if it were so easy, no? What works for a single, highly-focused project (or a handful of similar situations) doesn't scale when you have to take care of bazillion customers, do boring/smaller tasks and keep the machine ticking.
Having seen how the sausage is made, Amazon's internal engineering culture is a textbook example of the principal/agent problem. You only get a raise from a promo, and you only get a promo from scope. It's way easier for management to demonstrate scope and so those folks build out empires of crap. In the best case they're just consuming oxygen, but I have definitely seen orgs created that were a negative influence on productivity.
Never worked for BigTech, so I'mm really not in my territory, but I wonder what is the purpose of all those interviews to get "the best of the best" and then let these things happen. What are the incentives for the management layer?
probably the main thing is just having low-level engineers competent enough to run operations on the dozens of poorly maintained 50k LoC codebases per team, hundreds of bespoke internal dependencies, processes, and custom tooling. build ecosystems of 5+ languages.
secondarily its an industrialization of software development. the hiring process is where they try define the labor as a replaceable component. grab the best cogs you can annually for the lowest price, run them in the machine for 2-4 years and swap most of them out before they get too expensive, or specialized or uppity.
Lol, spinning up swat teams because someone high up decides "drop everything this is my pet priority now" is politicking. It looks good for the leaders, meanwhile its the engineers pulling the all nighters and dealing with having to maintain systems that are operationally compromised from day 0 because there's no proper planning/scoping involved other than "Big Man says this needs to be done in 2 weeks"
I think the subtext of the last few weeks is the Anthropic was becoming severely capacity constrained (or approaching that). They seem to have had to sign two somewhat adverse contracts with Amazon and Google in short succession. suddenly model quality is back up again.
What is all this AI doing? People are spending 10’s to 100’s of billions and no service or technology seems better or cheaper. Everything is more expensive and worse.
- Development velocity is very noticeably much higher across the board. Quality is not obviously worse, but it's LLM assisted, not vibe coding (except for experiments and internal tools).
- Things that would have been tactically built with TypeScript are now Rust apps.
- Things that would have been small Python scripts are full web apps and dashboards.
- Vibe coding (with Claude Desktop, nobody is using Replit or any of the others) is the new Excel for non tech people.
- Every time someone has any idea it's accompanied by a multi page "Clauded" memo explaining why it's a great idea and what exactly should be done (about 20% of which is useful).
- 80% of what were web searches now go to Claude instead (for at least a significant minority of people, could easily be over 50%).
- Nobody talks about ChatGPT any more. It's Claude or (sometimes) Gemini.
- My main job isn't writing code but I try to keep Claude Code (both my personal and corpo accounts) and OpenCode (also almost always Claude, via Copilot) busy and churning away on something as close to 100% of the time as I can without getting in the way of my other priorities.
We (~20 people) are probably using 2 orders of magnitude more inference than we were at the start of the year and it's consolidated away from cursor, ChatGPT and Claude to just be almost all Claude (plus a little Gemini as that's part of our Google Whateverspace plan and some people like it, mostly for non-engineering tasks).
No idea if any of this will make things better, exactly, but I think we'd be at a severe competitive disadvantage if we dropped it all and went back how things were.
I am hobbyist playing around. Recently dropped CC (which gave me a sense of awe 2 months ago), but they realized GPUs need CapEx and I want to screw around with pi.dev on a budget. Then on to GH Copilot but couldn't understand their cost structure, ran out of quota half month in, now on Codex. I don't really see any difference for little stuff. I also have Antigravity through a personal Gmail account with access to Opus et al and I don't understand if I am paying for it or not. They don't have my CC so that's a breather.
It's all romantic, but a bunch of devs are getting canned left and right, a slice of the population whose disposable income the economy depends on.
It's too late to be a contrarian pundit, but what's been done besides uncovering some 0-days? The correction will be brutal, worse than the Industrial Revolution. Just the recent news about Meta cuts, SalesForce, Snap, Block, the list is long.
Have you shipped anything commercially viable because of AI or are you/we just keeping up?
> The correction will be brutal, worse than the Industrial Revolution.
Has it occurred to you that there might not be a correction, and that the outcome would still be brutal, at least on par with the industrial revolution.
It's physically impossible to build out the datacenters required for the "AI is actually good and we have mass layoffs" scenario. This Anthropic investment is spurred on because they've already hit a brick wall with capacity.
$40B goes a long way, but not for datacenters where nearly every single component and service is now backordered. Even if you could build the DC, the power connection won't be there.
The current oil crisis just makes all of that even worse.
Doesn't that just draw out the AI revolution by a few years? I don't see why it would stop anything though.
Imagine a scenario where someone claimed that it was physically impossible to replace all the buggies with automobiles because everything was backordered and there were labor shortages. Surely the replacement still happens eventually though?
A drawn out long change simply doesn't have the major societal upset that imminent mass-unemployment has.
With how much scale AI datacenters want and how the Trump administration has made supply problems significantly worse, we'd be talking decades, plural.
I don't think lowering the rate a bit is going to be sufficient to avoid major upsets. If (arbitrary example) every software developer were forced to switch jobs over a 10 year period that would still be an extremely disruptive sequence of events. And I don't think there's any scenario in which software developers are widely impacted but other industries somehow aren't.
Digitization was already fairly disruptive and that involved much smaller changes than what we appear to be facing while also taking place over something like 30 years or more.
We must have a very different view of the world because in my neck of the woods companies are desperate for senior talent. And it's become even harder to find seniors now that everyone has access to a machine that can create the appearance of experience.
I mean as in living through the industrial revolution would have been wild. So whether we have an AI revolution or an AI bubble it's bound to be a roller coaster.
And that's without accounting for the various wars (and resultant economic impacts) that are already in progress. A large part of what drove the meat grinder of WWI was (very approximately) the various actors repeatedly misjudging the overall situation and being overly enthusiastic to try out their shiny new weapons systems. If one or more superpowers decide to have a showdown the only thing that might minimize loss of life this time around is (ironically enough) the rise of autonomous weapons systems. Even in that case as we know from WWII the logical outcome is a decimated economy and manufacturing sector regardless of anything else that might happen.
What strategic merit is there in targeting civilians or life critical infrastructure in a fully automated battlebot scenario? Perhaps it's naive but I would expect stockpiles, datacenters, and any key infrastructure on which the local semiconductor fabrication depends to be the primary targets.
Look au Ukraine for answers and how russians target almost purely civilian infrastructure and civilians in terror campaigns every single day and night, same as nazis did to Britain in WWII. With exactly same results but they just double down and send more drones next day.
russia is really and empire of the dumb and subjugated serfs at this point (again, history repeats), but they are far from only such place.
Dont expect more, most people are not that nice when SHTF.
The aim of war is to make political change and gain control of opponent. it is much more valuable to capture datacenters, infra and semiconductor fabricaton than to destroy and rebuild it.
Bubbles like the AI bubble are a game theoretic outcome of a revolution. Many players invest heavily to avoid losing, but as a whole the market over invests. This leads to a bubble.
We're not talking about the LLMs of today but whatever shows up 2 years from now and then again 2 years after that. Don't look at the present state of things but instead project the trend line.
There has always been a gap between the experience of solo/small shop developers, vs. developers who work in teams in a large corporate environment. But thanks to open source, we have for the past twenty years at least mostly all been using the same tools.
But right now, the difference in developer experience between a dev on a team at a business which has corporate copilot or Claude licenses and bosses encouraging them to maximize token usage, vs a solo dev experimenting once every few months with a consumer grade chat model is vast.
Meta seemingly has a constant stream of product managers. If llm’s really augment the productivity of engineers, why isn’t meta launching lots more stuff? I mean there’s no harm in at least launching one new thing.
What are all those people doing with the so called productivity enhancements?
What I’m calling into question is how much does generating more code matter if the bottle neck is creativity/imagination for projects?
The only thing I’ve seen is a really crummy meta AI thing implemented within WhatsApp.
It’s allowed a sludge of internal tools to spin up, and more bloat. The ability to sand bag and over build these tools has gotten 2-10x worse.
Only solution I can think of is to drastically cut headcount so productivity is back to prior levels, and profitability is raised. Big Tech is mostly market constrained with not much room to grow beyond the market itself growing.
As for startups, seems like AI tools have drastically reduced their time to market and accelerated their growth curves.
Im convinced the most scarce skill on the planet is the ability to a) envision something that needs to exist in the world b) explain how the thing creates value from a financial perspective.
Most people tend to think they know what they are talking about (e.g. surface level understanding of how to think economically) and end up making basket-case decisions - only realising it months later. By that point they will fail to admit defeat and keep going on.
"As for startups, seems like AI tools have drastically reduced their time to market and accelerated their growth curves."
What I see in my backyard: coding now takes significantly less time, but its just coding. Before one gets to building there are squabbles between business and product people. Testing takes just as much as it used to. Since nice to haves are easy to add and product people begin to take it for granted, the product cycles don't get shorter.
Give it time. Right now its just coding, but procedural AI will come after product development, architecture, and then whatever is left of management.
The best people can not only envision products but also possess great judgement without needing data. For AI to even come close it would need an insane amount of data that is nuanced and subtle - by the the time the AI has obtained all the necessary data and made sense of it the human is long gone working on something else.
But these people will age out and juniors do not get hired. “Good judgment comes from experience, and experience comes from bad judgment.” and all that.
Is LLM going to invent its own languages that no average programmer will understand? As in "I don't need your C++ human, I will rewrite your fart app in ClaudASM and you will like it". These are naive questions, but I can't visualize how all of this will unfold.
A neutral hobbyist on a $20 budget will build something and immediately bump into quotas. Its not going to be an enjoyable experience.
A negatively predisposed pro who only dabbles in AI gets to the first disappointment, smiles, and thinks "yeah, about what i expected" and quits.
To learn those new tools one needs to not be stingy. Invest as much as needed into tokens, subscriptions, and maybe most importantly invest the time. Spend time building various things. Try out various models not just for coding, but as part of apps being built. For bonus points, meaningfully experiment with local models. I try to avoid discussions with sceptics who have not put at least a few months of effort into learning those tools. It's like discussing driving with my mother in law, who spent maybe 20hrs behind the wheel through her whole life (and is very, very opinionated!).
In my opinion it's a complete waste of time and money to learn something that is gated by a company that might disappear tomorrow.
It's akin to company courses to learn something that is specific to that company. Of course you do them on the job, there is no point in doing them if you don't work there.
Similarly what's the point of trying 300 different models if any job will decide for you which one they approve the use of, and you are liable to get fired and asked for damages if you let anything else access company intellectual property?
The difference is (if you'll forgive me recruiting a couple of straw men for the purpose of illustrating the spectrum we are talking about here):
Hobbyist solo dev, counting tokens, hitting quotas, trying things on little projects, giving up and not seeing what the fuss is about.
vs
Corporate developer, increasingly held accountable by their boss for hitting metrics for token usage; being handed every new model as soon as it comes out; working with the tools every day on code changes that impact other developers on other teams all of whom have access to those same tools.
Okay, so just to be clear you're not commenting on productivity? Or what does "changes that impact" mean?
I might be missing a lot of self-evident assumptions here but I feel like I'm still missing so much context and have no idea what this difference is actually describing.
If you have some objective measure of productivity in mind, feel free to share it, but no that's not what I'm commenting on.
I'm talking more about why threads like this seem to be full of people saying 'this has completely changed how corporate development works' and other people saying 'I tried it a few times and I don't get the hype'
Developers being let go is about the economy. Every time we see a slowdown people are let go and we always blame the fad but it's the economy not whatever.
> Every time someone has any idea it's accompanied by a multi page "Clauded" memo explaining why it's a great idea and what exactly should be done (about 20% of which is useful).
we're in the same boat, and currently trying to fix that 20% problem because it's the biggest hindrance to shipping things quickly
there is a ton of learned ceremony that we have to undue gracefully because it's extremely tempting to vibe code a problem spec as opposed to just... talking to users directly and understanding what the actual problem is
As a CTO I can say that this is not my experience.
My experience these days is fighting corporate bureaucracy and inertia to make sure we reap the benefits of faster coding. Feeding agents with work is not a problem. Building teams that use those tools effectively is the problem. (Say, shall we merge product and engineering teams? Do we start getting rid of people who refuse to use AI? What do we do with pentests? How do we strengthen the tools that do code analysis and weed out lazy devs who can now more easily pretend to be invested in their work? Stuff like this keeps me busy.
As a CTO this has been my experience as well. I would add in every non-technical C-suite member aiming to use AI as some magic lever to avoid prioritizing projects or engaging in real critical thinking. Too many people are offloading their cognitive decisionmaking to some magic box, thinking it has all the answers, because its output appears magical and complete.
After 25 years in programming I think I’ll finally start that farm ;)
> Do we start getting rid of people who refuse to use AI?
I don't even think the bigger companies are going to waste time on figuring out how to retrain, they're just going to do industrial scale layoffs and then rebuild from the ground up with people who won't get past interviews without demonstrating hard skills in this area.
There is a shocking gap growing right now, it's a Wile E. Coyote not realizing he already walked off the cliff type of situation for a lot of people.
Ultimately the shareholders want to see the money. They dont give a crap about what you think or what the poster above thinks - you're both accountable to the shareholders who do not employ you for fun. They employ you for the sole purpose of making them wealthier. All this incremental spend on tokens shows up in the financials positively or it doesn't.
> Ultimately the shareholders want to see the money.
Seems like we're saying the same thing?
> All this incremental spend on tokens shows up in the financials positively or it doesn't.
Right, and we're talking about the staff failing to spend the incremental tokens at all, thus failing to discover whether or not they'll show up positively. I'm just saying, investors are probably going to decide to roll the dice on a complete staffing rebuild rather than try to wait for the existing corporate culture to adapt because they're going to get fomo. Arguably it's already happening.
> My impression has always been it's more important the build the correct thing (what the customer needs/wants) rather than more stuff faster.
The process of learning what the customer needs/wants is a heavily iterative one, often involving throwing prototypes at them or betting at a solution, then course-correcting based on their reaction. Similarly, the process of building the correct thing is almost always an iterative approximation - correctness is something you discover and arrive at after research and prototypes and trying and getting it wrong.
All of that benefits from any of its steps being done faster - but it's up to the org/team whether they translate this speedup to quality or velocity. For example, if AI lets you knock out prototypes and hypothesis-testing scripts much faster, you can choose whether to finish earlier (and start work on next thing sooner), or do more thorough research, test more hypothesis, and finish as normally, but with better result.
(Well, at least theoretically. If you're under competitive pressure, the usual market dynamics will take the choice away, but that's another topic.)
You have a specific idea of customer in mind. Likely different than the gp’s. Many types of customers are quite happy to have prototypes thrown at them. Sometimes it’s even contractually required in agency work.
Is it just me or is this whole mania exposing those people who thought they were great ‘thinkers’? The takes I see are so utterly flawed it’s ironic - people refer to llm’s as hallucinating when the real halluncinations are from people cosplaying the role of management/investors when they have never done said role professionally in their life.
For sure, but these days product management mistakes can be more easily rectified. Before that, if we invested 4 months in building something that did not land, we'd be quite reluctant to jettison this and start fresh. Egos, career considerations, sunk cost, etc. I think I will soon be able to say "not any more", since doing a U turn can be cheaper than pretending the bad choice is the best choice. "Oops, lets redo this" vs. 6 months of executive squabbling about whose fault it is that we wasted $3M in development costs on something that clearly does not perform.
Also, give it time. Real adoption in boring companies started Q1. Q2 is, I think, this settling in and people learning how to do their work and manage their responsibilities. Q3/Q4 will be the time when I expect to start seeing higher velocities across all IT-adjacent products I use.
Iteration speed is far more important than the volume of features delivered, though we are tackling aspirational features that would previously have been considered too complex or niche, too.
This with the ability to research, iterate on prototypes, in my opinion allows to determine the right thing quicker as well. Of course right now the value is largely intuition based, there may be some immediate revenue/profit, but revenue/profit will take time to follow, so in a way it is a speculative intuition based bet. Financial gains will take time to follow, so for a period of time it will be "trust me bro" for at least some cases, but I suppose future will show, since the intuition seems so strong about it. You can't have good data about an emerging tech like that.
Isn't that trivially true? Scenario 1) Spend $10,000 to make one prototype. You get one shot, so you prepare and do as much pre-work as humanly possible, but because you only get one shot, you forgot the ask the question that in Hobbs sight was obvious. Scenario 2) prototypes cost $1,000 so you get multiple shots. So you don't do as much pre-work, throw a half dozen things at the wall. One of them sticks! It really resonates with customers. You iterate a few more times, and when it's finally on the market, you have a successful business.
The difference is all that pre-work. The problem with that is some things are only obvious after you've built one and it doesn't fit just right for some reason. That reason is impossibly harder to just reason about and figure out vs iterating where possible. For software things that's easier. For hardware, we have stories like the palm pilot engineer having a wooden block with them for a week before deciding on the form factor for it. Such pre-work is valuable, but if the cost of prototypes is way down, you can afford to iterate instead of trying to psychically predict everything up front. Of course that doesn't work for eg trips to the Moon, but most busineeses aren't doing that.
The problem is in validating the prototype. Whether the users are consumers or enterprises or internal stakeholders, they aren’t going to try 10 different prototypes. They will try one or two.
Most business software isn’t complicated to implement (i.e. it doesn’t require multiple prototypes to determine which technical approach is best). Usually for most apps you approximately know the technical implementation. What requires taste, experience, or whatever you want to call it, is the user experience and if your software actually solves a real problem. You can’t really just churn on prototypes to solve that. You will lose the patience of your user base.
Even so-called UX and product experts get stuff wrong all the time. Going from idea to prototype to feedback in hours or days rather than days or weeks feels like a superpower, at least in the very customer facing parts of what we do.
Did CAD make engineers better? certain products are only possible because of CAD but the pen and paper guys weren’t obviously less efficient, and I personally think they were very efficient.
When prototypes are harder to build you focus on answering the biggest questions. I feel like you spend more time iterating on details in CAD, even when the larger idea is invalid.
It’s no more exhausting than the alternative. It feels good being able to build more and experiment more.
The biggest downside is the feeling that people sometimes turn their brain off and aren’t even doing basic checks on some of the slop their LLMs produce.
It sounds very similar to my shop. I have QA people and Product Managers using Claude to develop better integration and reporting tools in Python. Business users are vibe coding all kinds of tools shared as Claude Artifacts, the more ambitious ones are building single page app prototypes. We ported one prototype to Next.js and hosted on Vercel in a couple of days and then handed it back to them with a Devcontainer and Claude Code so they can iterate on it themselves; and we also developed all the security infrastructure, scaffolding, agent instructions & policy required to do this for low stakes apps in a responsible way.
It hardly seems worth it to try to iterate on design when they can just build a completely functional prototype themselves in a few hours. We're building APIs for internal users in preference to UIs, because they can build the UIs themselves and get exactly what they need for their specific use cases and then share it with whoever wants it.
We replaced an expensive, proprietary vendor product in a couple of weeks.
I have no delusions about the scale or complexity limits of these projects. They can help with large, complex systems but mostly at the margins: help with impact analysis, production support, test cases, code review. We generate a lot of code too but we're not vibe coding a new system of record and review standards have actually increased because refactoring is so much cheaper.
The fact is that ordinary businesses have a LOT of unmet demand for low stakes custom software. The ones that lean into this will not develop superpowers but I do think they will out-compete slow adopters and those companies will be forced to catch up in the next few years.
I develop presentations now by dumping a bunch of context in a folder with a template and telling Claude Cowork what I want (it does much better than web version because of its python and shell tools and it can iterate, render, review, repeat until its excellent). The copy is quite good, I rewrite less than a third of it and the style and graphics are so much better than I could do myself in many hours.
No one likes reading a bunch of vibe coded slop and cultural norms about this are still evolving; but on balance its worth it by far.
Personally at my place, there hasn't been a noticable velocity change since the adoption of Claude Code. I'd say it's even slightly worse as now you have junior frontend engineers making nonsense PRs in the backend.
Mainn blockers are still product, legal, management ... which Claude code didn't help with.
Is your team measuring how much of your code is being written with claude and comparing amongst the team, like what works best in your codebase? How are you learning from each other?
I’m making a team version of my buildermark.dev open source project and trying to learn about how teams would like to use it.
Different teams are using it in very different ways so it can be tough to compare meaningfully.
Backends handling tens to hundreds of thousands of messages per second with extremely high correctness and resilience requirements are necessarily taking a different approach to less critical services that power various ancillary sites/pages or to front end web apps.
That said there's a lot of very open discussion around tooling, "skills", MCP, etc., harnesses, and approaches and plenty of sharing and cross-pollination of techniques.
It would be great to find ways to better quantify the actual value add from LLMs and from the various ways of using them, but our experience so far is that the landscape in terms of both model capability and tooling is shifting so fast that that's quite hard to do.
Thanks for the feedback. I agree that it’s changing very fast, which is why my thesis is that this tooling will be needed to help everyone on the team keep up.
I am an early Gemini daily driver type engineer, feels like Node, Firefox, React and Tailwind all over again, Claude Sonnet is 10x more expensive, quick thought experiment do you think 10 Gemini prompts is needed to match the quality of one Claude Code prompt? The harness around Gemini is an issue but I built my own (in Rust)
I'm not sure. I have a buddy that's one of the better engineers I know personally, and he struggled to maintain an "AI Lent" for even a month. He found he just wasn't productive enough without it.
The kicker is he wasn't able to compete without the agents.
IMO this is the natural end state of LLM fueled capitalism: products skating along the razor edge between "has value under capitalism" and "is a heap of garbage" until we suddenly realize there's nothing under our feet.
This sounds like my office, but we're a bit more tilted toward Codex. I personally use Claude Cowork for drudge-admin work, GPT 5.5-Pro for several big research tasks daily, and the LLMs munge on each other's slop all day as I try my best to wrap my head around what has been produced and get it into our document repository -- all the while being conscious that the enormous volume of stuff I'm producing is a bit overwhelming for everyone.
We are definitely reaching the point where you need an LLM to deal with the onslaught of LLM-generated content, even if the humans are being judicious about editing everything. We're all just cranking on an inhumanly massive amount of output and it's frankly scary.
Everything from complex backend logic and processing to new user facing festivals, ops and infra tools, analytical apps, etc.
A lot of engineers now describe the problem, discuss the outline of the solution with the LLM, and then get it to write and test most of it ahead of their review. They tell me it usually takes the same approach they would anyway and even when it doesn’t, it’s often faster to explain what’s wrong and give the LLM another try.
Very little (and even then, only simple internal tools) gets written without a human owning the code and reviewing it thoroughly, but even with that overhead the productivity boost is impressive.
I kept asking this question last year, especially after that initial METR report showing people believed themselves to be faster when they were slower. Then I decided to dive in feet-first for a few weeks so that nobody could say I hadn't tried all I could.
At work, what I see happening is that tickets that would have lingered in a backlog "forever" are getting done. Ideas that would have come up in conversation but never been turned into scoped work is getting done, too. Some things are no faster at all, and some things are slower, mostly because the clankers can't be trusted and human understanding can't be sped up, or because input is needed from product team, etc. But the sorts of things that don't make it into release notes, and are never announced to customers, those are happening faster, and more of them are happening.
We review server logs, create tickets for every error message we see, and chase them down, either fixing the cause or mitigating and downgrading the error message, or however is appropriate to the issue. This was already a practice, but it used to feel like we were falling farther behind every week, as the backlog of such tickets grew longer. Most low-priority stuff, since obviously we prioritized errors based on user impact, but now remediation is so fast that we've eliminated almost the entire backlog. It's the sort of things that if we were a mobile app, would be described as "improvement and bug fixes" generically. It's a lot of quality-of-life issues for use as backend devs.
At home, I'm creating projects I don't intend for anyone outside my family to see. So far things I could theoretically have done myself, even related to things I've done myself before, but at a scale I wouldn't bother. Like a price-checker that tracks a watchlist of grocery items at nine local stores and notifies me in discord of sales on items and in categories I care about. It's a little agent posting to a discord channel that I can check before heading out for groceries.
Or several projects related to my hobbies, automating the parts I don't enjoy so much to give me more time for the parts I do. My collection of a half-dozen python scripts and three cron jobs related to those hobbies has grown to just over 20 such scripts and 14 cron jobs. Plus some that are used by an agent as part of a skill, although still scripts I can call manually, because I'll go back to cron jobs for everything if the price of tokens rises a bit more.
I was super-skeptical, and now I'm not. I think companies laying off employees are delusional or using LLMs as an excuse, but there is zero question in my mind that these things can be a huge boon to productivity for some categories of coding.
I went through a similar journey. That and having read all the other experienced engineers’ anecdotes I think the current consensus is that it can boost productivity and does so for a lot of people but that vibe coding still remains unviable in a lot of situations.
Development velocity is faster, but the code quality hits take a while to manifest.
Some places are more diligent, but most are not. We HATE reading other people's code, and we only have so much focus capacity per day to review all the shit these clunkers spew out.
Over time, the errors induced by Looks Good To Me code reviews compound.
I'm burning an insane number of tokens 8-12 hours a day for the dramatic improvement of some internal tooling at a big tech company. Using it heavily for an unannounced future project as well.
We suddenly have a proliferation of new internal tools and resources, nearly all of which are barely functional and largely useless with no discernible impact on the overall business trajectory but sure do seem to help come promo time.
Barely an hour goes by without a new 4-page document about something that that everyone is apparently ment to read, digest and respond to, despite its 'author' having done none of those steps, it's starting to feel actively adversarial.
Without good management AI is just a new way to make terrible work in unprecedented quantities.
With good management you will get great work faster.
The distinguishing feature between organisations competing in the AI era is process. AI can automate a lot of the work but the human side owns process. If it’s no good everything collapses. Functional companies become hyper functional while dysfunctional companies will collapse.
Bad ideas used to be warded off by workers who in some shape or form of malicious compliance just would slow down and redirect the work while advocating for better solutions.
That can’t happen as much anymore as your manager or CEO can vibe code stuff and throw it down the pipeline for the workers to fix.
If you have bad processes your company will die, or shrivel or stagnate at best. Companies with good process will beat you.
I personally noticed this. The speed at which development was happening at one gig I had was impossible to keep up with without agentic development, and serious review wasn't really possibile because there wasn't really even time to learn the codebase. Had a huge stack of rules and MCPs to leverage that kinda kept things on the rails and apps were coming out but like, for why? It was like we were all just abandoning the idea of good code and caring about the user and just trying to close tickets and keep management/the client happy, I'm not sure if anyone anywhere on the line was measuring real world outcomes. Apparently the client was thrilled.
It felt like... You know that story where two economists pass each other fifty bucks back and forth and in doing so skyrocket the local GDP? Felt like that.
My main use of vibecoding is creating dozens of internal tools that have sped up tasks, or made tasks possible that were previously not. These tools would have taken weeks of time to build manually and would have been hard to justify, rather than just struggling with manual processes every now and again. AI has been life-changing in creating these kinda janky tools with janky UI that do everything they're supposed to perfectly, but are ugly as hell.
Are you able to describe any of those internal tools in more detail? How important are they on average? (For example, at a prior job I spent a bit of time creating a slackbot command "/wtf acronym" which would query our company's giant glossary of acronyms and return the definition. It wasn't very popular (read: not very useful/important) but it saved myself some time at least looking things up (saving more time than it took to create I'm sure). I'd expect modern LLMs to be able to recreate it within a few minutes as a one-shot task.)
If it's useless that's a you problem. I've been building CRUDs that would have taken me a month to get perfectly right in the span of 4-5 days which save an enormous number of human tech support hours.
Sorry man but the software world is littered with CRUD apps, they are called CRUD apps for a reason. They're basically the mass assembled stamped L-bracket of the software world. CRUD apps have also had template generators for like 30 years now too.
Still useless in the sense that if you died tomorrow and your app was forgotten in a week the world will still carry on. As it should. Utterly useless in pushing humanity forward but completely competent at creating busy work that does not matter (much like 99% of CRUD apps and dashboards).
But sure yeah, the dashboard for your SMB is amazing.
The software industry's value proposition for the vast majority of businesses running the world lies in CRUD apps that properly capture business requirements. That's infinitely more relevant in insurance, pharma, banking and logistics than any technological breakthrough of the past 25 years.
Your rant just shows you don't understand why people pay for software.
I have one that serves a few functions- Tracks certificates and licenses (you can export certs in any of the majorly requested formats), a dashboard that tells you when licenses and certs are close to expiring, a user count, a notification system for alerts (otherwise it's a mostly buried Teams channel most people miss), a Downtime Tracker that doesn't require people to input easily calculatable fields, a way for teams to reset their service account password and manage permissions, as well as add, remove, switch which project is sponsoring which person, edit points of contact, verify project statuses, and a lot more. It even has some quick charts that pull from our Jira helpdesk queue- charts that people used to run once a week for a meeting are just live now in one place. It also has application statuses and links, and a lot more.
I'd been fighting to make this for two years and kept getting told no. I got claude to make a PoC in a day, then got management support to continue for a couple weeks. It's super beneficial, and targets so many of our pain points that really bog us down.
A lot of businesses can get by just fine with making it one person's responsibility to maintain a spreadsheet for this. It can be fragile though as the company grows and/or the number of items increases, and you have to make sure it's all still centralized and teams aren't randomly purchasing licenses or subscriptions without telling anyone, it needs to be properly handed off if the person leaves/dies/takes a vacation, backed up if not using a cloud spreadsheet... I've probably seen at least a dozen startups come and go over the years purporting to solve this kind of problem, other businesses integrate it into an existing Salesforce/other deployment... it seems like a fine choice for an internal tool, so long as the tool is running on infrastructure that is no less stable than a spreadsheet on someone's machine.
In the startup world something like "every emailed spreadsheet is a business" used to be a motivating phrase, it must be more rough out there when LLMs can business-ify so many spreadsheet processes (whether it's necessary for the business yet or not). And of course with this sort of tool in particular, more eyes seeing "we're paying $x/mo for this service?" naturally leads to "can't we just use our $y/mo LLM to make our own version?". Not sure I'd want to be in small-time b2b right now.
Why are you ignoring the fact that grabbing data from heterogeneous sources, combining it and presenting it is generally never a trivial task? This is exactly what LLMs are good for.
If you are using an LLM to actually fetch that data, combine it, and present it to you in an ad hoc way (like you run the same prompt every month or something), I wouldn't trust that at all. It still hallucinates, invents things and takes short cuts too often.
If you are using an LLM to create an application to grab data from heterogeneous sources, combine it and present it, that is much better, but could also basically be the excel spreadsheet they are describing.
Your knowledge of LLMs is outdated by at least a year. For the past three months at least my team has been one-shotting complex SQL queries that are as semantically correct as your ability to describe them.
And why do you diminish the skill of good data wrangling as if it weren’t the most valuable skill in the vast majority of computer programming jobs? Your cynicism doesn’t correspond with the current ground truth in LLM usage.
Well, that is still having the LLM write code which is more like my second scenario. I use SOTA LLMs for coding literally every day. I don't think my knowledge is "outdated by at least a year".
The ones I can mention.. one that watches a specific web site until an offer that is listed expires and then clicks renew (happens about once a day, but there is no automated way in the system to do it and having the app do it saves it being unlisted for hours and saves someone logging in to do it). Several that download specific combinations of documents from several different portals, where the user would just suck it up previously and right-click on each one to save it (this has a bunch of heuristics because it really required a human before to determine which links to click and in what order, but Claude was able to determine a solid algo for it). Another one that opens PDFs and pulls the titles and dates from the first page of the documents, which again was just done manually before, but now sends the docs via Gemma4 free API on Google to extract the data (the docs are a mess of thousands of different layouts).
My team has also adopted this - it's much easier to add another layer than to refine or simplify what exists. We have AI skills to help us debug microservices that call microservices that have circular dependencies.
This was possible before but someone would maybe notice the insane spaghetti. Now it's just "we'll fix it with another layer of noodles".
That's so interesting because where I work, the push was to "add one more API" to existing services, turning them into near monoliths for the sake of deployment and access. Still a mess of util and helper functions recursively calling each other, but at least it's one binary in one container.
Unfortunately I saw this pre-AI with microservices, where while empowering developers with their beloved microservices, we create intense complexity and deployment headaches. AI will fix the slop with an obscuring layer of complexity on top.
Oh the "micro services" are all coupled. To test anything you have to deploy a constellation of interdependent services with redundant DBs, each generating new IDs for the same underlying resource.
We’re seeing the exact same where I work. Our main Slack channels have become inundated with “new tool announcements!”, multiple per day, often solving duplicate problems or problems that don’t exist. We’ve had to stop using those channels for any real conversation because most people are muting them due to the slop noise.
And what’s worse is that when someone does build a decent tool, you can’t help but be skeptical because of all the absolute slop that has come out. And everyone thinks their slop doesn’t stink, so you can’t take them at their word when they say it doesn’t. Even in this thread, how are you to know who is talking about building something useful vs something they think is useful?
A lot of people that have always wanted to be developers but didn’t have the skills are now empowered to go and build… things. But AI hasn’t equipped them with the skill of understanding if it actually makes sense to build a thing, or how to maintain it, or how to evolve it, or how to integrate it with other tools. And then they get upset when you tell them their tool isn’t the best thing since sliced bread. It’s exhausting, and I think we’ve yet to see the true consequences of the slop firehose.
>Barely an hour goes by without a new 4-page document about something that that everyone is apparently ment to read, digest and respond to, despite its 'author' having done none of those steps, it's starting to feel actively adversarial.
well, isn't that what AI can be used effectively for - to generate [auto]response to the AI generated content.
ideally we'd delegate all mindless/routine stuff to AI, and we'd dedicate ourselves to the higher creative and scientific pursuits. Somehow though i think that ideal will come with some adjustments/distortions/bugs :)
I'll throw this out as something where it has saved literally weeks of work: debugging pathological behaviour in third-party code. Prompt example: "Today, when I did U, V, and W. I ended up with X happening. I fixed it by doing Y. The second time I tried, Z happened instead (which was the expected behaviour). Can you work out a plausible explanation for why X happened the first time and why Y fixed it? Please keep track of the specific lines of code where the behaviour difference shows up."
This is in a real-time stateful system, not a system where I'd necessarily expect the exact same thing to happen every time. I just wanted to understand why it behaved differently because there wasn't any obvious reason, to me, why it would.
The explanation it came back with was pretty wild. It essentially boiled down to a module not being adequately initialized before it was used the first time and then it maintained its state from then on out. The narrative touched a lot of code, and the source references it provided did an excellent job of walking me through the narrative. I independently validated the explanation using some telemetry data that the LLM didn't have access to. It was correct. This would have taken me a very long time to work out by hand.
Edit: I have done this multiple times and have been blown away each time.
This seems to be a common denominator for what LLMs actually do well: Finding bugs and explaining code. Anything about producing code is still a success to be seen.
> Prompt example: "Today, when I did U, V, and W. I ended up with X happening. I fixed it by doing Y. The second time I tried, Z happened instead (which was the expected behaviour). Can you work out a plausible explanation for why X happened the first time and why Y fixed it? Please keep track of the specific lines of code where the behaviour difference shows up."
> The explanation it came back with was pretty wild. It essentially boiled down to a module not being adequately initialized before it was used the first time and then it maintained its state from then on out.
Even without knowing any of the variable values, that explanation doesn't sound wild at all to me. It sounds in fact entirely plausible, and very much like what I'd expect the right answer to sound like.
The wild part, for me at the time, was how many steps there were from cause and effect and how perfectly they'd been reasoned through. The first time I had that experience was my first real "this LLM stuff might have some legs". My second similar experience several days later was "hmmm that wasn't a fluke..."
I'm still at a stage where I'm not completely sure that I like the code that Codex or Claude wants to write. Sometimes it's good, sometimes it takes 5 or 6 iterations to get it somewhere I'm happy with. But wow, on the front end of the work, they are great design/review/iterate partners; sometimes I let the tools write the first draft and then I find the gaps, sometimes I write the first draft and let the tools find the gaps. Either way has worked really well for making solid debt-free progress.
I answered this in a different comment below, but a lot of the friction is around the amount of time it takes to test/review/submit etc, and a lot of this is centered around tooling that no one has had the time to improve, perf problems in clunky processes that have been around longer than anyone individual, and other things of this nature. Addressing these issues is now approachable and doable in one's "spare time".
The point of that friction is to keep the human in the loop wrt code quality, it's not meant to be meaningless busywork. It's difficult to believe that you sustain the benefit of those systems. Anthropic and Microsoft publicly failed to keep up code quality. They would probably be in a better spot currently if they used neither, no friction, no AI. But that friction exists for a reason and AI doesn't have the "context length" to benefit from it.
This the the difference between intentional and incidental friction, if your CI/CD pipeline is bad it should be improved not sidestepped. The first step in large projects is paving over the lower layer so that all that incidental friction, the kind AI can help with, is removed. If you are constantly going outside that paved area, sure AI will help, but not with the success of the project which is more contingent on the fact that you've failed to lay the groundwork correctly.
For me/my team, I use it to fix DevProd pain points that I would otherwise never get the investment to go solve. Just removed Webpack for Rspack, for example. Could easily do it myself, which is why I can prompt it correctly and review the output properly, but I can let it run while I’m in meetings over more important product or architectural decisions
it's crazy that the experiences are still so wildly varying that we get people that use this strategy as a 'valid' gotcha.
AI works for the vast majority of nowhere-near-the-edge CS work -- you know, all the stuff the majority of people have to do every day.
I don't touch any kind of SQL manually anymore. I don't touch iptables or UFW. I don't touch polkit, dbus, or any other human-hostile IPC anymore. I don't write cron jobs, or system unit files. I query for documentation rather than slogging through a stupid web wiki or equivalent. a decent LLM model does it all with fairly easy 5-10 word prompts.
ever do real work with a mic and speech-to-text? It's 50x'd by LLM support. Gone are the days of saying "H T T P COLON FORWARD SLASH FORWARD SLASH W W W".
this isn't some untested frontier land anymore. People that embrace it find it really empowering except on the edges, and even those state-of-the-art edge people are using it to do the crap work.
This whole "Yeah, well let me see the proof!" ostrich-head-in-the-sand thing works about as long as it takes for everyone to make you eat their dust.
People ask for examples because they want to know what other people are doing. Everything you mention here is VERY reasonable. It's exactly the kind of stuff no one is going to be surprised that you are getting good results with the current AI. But none of that is particularly groundbreaking.
I'm not trying to marginalize your or anyone else's usage of AI. The reason people are saying "such as" is to gauge where the value lies. The US GDP is around 30T. Right now there's is something like ~12T reasonably involved in the current AI economy. That's massive company valuations, data center and infrastructure build out a lot of it is underpinning and heavily influencing traditional sectors of the economy that have a real risk of being going down the wrong path.
So the question isn't what can AI do, it can do a lot, even very cheap models can handle most of what you have listed. The real question is what can the cutting edge state of the art models do so much better that is productively value added to justify such a massive economic presence.
That's all well and good, but what happens when the price to run these AIs goes up 10x or even 100x.
It's the same model as Uber, and I can't afford Uber most of the time anymore. It's become cost prohibitive just to take a short ride, but it used to cost like $7.
It's all fun and games until someone has to pay the bill, and these companies are losing many billions of dollars with no end in sight for the losses.
I doubt the tech and costs for the tech will improve fast enough to stop the flood of money going out, and I doubt people are going to want to pay what it really costs. That $200/month plan might not look so good when it's $2000/month, or more.
Why not try it yourself? Inference providers like BaseTen and AWS Bedrock have perfectly capable open source models as well as some licensed closed source models they host.
You can use "API-style" pricing on these providers which is more transparent to costs. It's very likely to end up more than 200 a month, but the question is, are you going to see more than that in value?
Bedrock and other third party open weight hosted model costs are not subsidized. What could possibly be the investment strategy for being one of twelve fly-by-night openrouter operators hosting the latest Qwen?
It's an important concern for those footing the bill, but I expect companies really in the face of being impacted by it to be able to do a cost-benefit calculation and use a mix of models. For the sorts of things GP described (iptables whatever, recalling how to scan open ports on the network, the sorts of things you usually could answer for yourself with 10-600 seconds in a manpage / help text / google search / stack overflow thread), local/open-weight models are already good enough and fast enough on a lot of commodity hardware to suffice. Whereas now companies might say just offload such queries to the frontier $200/mo plan because why not, tokens are plentiful and it's already being paid for, if in the future it goes to $2000/mo with more limited tokens, you might save them for the actual important or latency-sensitive work and use lower-cost local models for simpler stuff. That lower-cost might involve a $2000 GPU to be really usable, but it pays for itself shortly by comparison. To use your Uber analogy, people might have used it to get to downtown and the airport, but now it's way more expensive, so they'll take a bus or walk or drive downtown instead -- but the airport trip, even though it's more expensive than it used to be, is still attractive in the face of competing alternatives like taxis/long term parking.
None of that is concrete though; it's all alleged speed-ups with no discernable (though a lot of claimed) impact.
> This whole "Yeah, well let me see the proof!" ostrich-head-in-the-sand thing works about as long as it takes for everyone to make you eat their dust.
People will stop asking for the proof when the dust-eating commences.
Im convinced none of these people have any training in corporate finance. For if they did they'd realise they were wasting money.
I guess you gotta look busy. But the stick will come when the shareholders look at the income statement and ask... So I see an increase in operating expenses. Let me go calculate the ROIC. Hm its lower, what to do? Oh I know, lets fire the people who caused this (it wont be the C-Suite or management who takes the fall) lmao.
Do you really think companies have started spending millions on tokens and no one from finance has been involved?
You could argue that all the spending is wasted (doubtless some is), but insisting that the decision is being made in complete ignorance of financial concerns reeks of that “everyone’s dumb but me” energy.
There is a difference to just noticing and attributing it to and recognizing negative financial outcomes. Right now for most companies they are still adjusting to declining inflation. Their bottom lines are doing quite well because consumer price inflation is much stickier than supply inflation. We are coming off of one of the quickest and largest supply lead inflationary cycles. It may not be immediately apparent for many companies that new expenditures are a drag on profitability.
The real thing to look at is whether or not the future outlook for company AI spend is heading up or down?
> Do you really think companies have started spending millions on tokens and no one from finance has been involved?
Oh, they were involved all right. They ran their analyses and realized that the increase in Acme Corp's share price from becoming "AI-enabled" will pay for the tokens several times over. For today. They plan to be retired before tomorrow.
That magic trick only works for publicly traded stocks.
Most firms are not a google or a Microsoft - a firms cash balance can become a strategic weapon in the right environment. So wasting money is not a great idea. Lest we forget dividends.
Moreover if you have a budget set re. Spend on tokens - you have rationing. Therefore the firm should be trying to get the most out of token spend. If you are wasting tokens on stuff that doesn’t create a benefit financially for the firm then indeed it is not inline with proper corporate financial theory.
No, it works for any VC-backed companies. Something like 60% of VC funding last year went to AI companies. VCs aren't going to give you a money unless you're building an agentic AI-native agent platform for agents.
No Employees of publicly traded firms benefit from short-term gains in the stock price, assuming the stock price jump holds throughout the period of grant/vesting.
People who work at VC-backed firms do not get to enjoy the same degree of liquidity, not even close. There can be some outliers but that is 0.1% of all.
Can't believe simple stuff like this has to be said.
CFOs or VPs absolutely benefit by hyping their company up to private investors by allowing tokenmaxxing to go on unchecked. Tender offers, acquisitions, and aquihires all exist. Or just good old fashioned resume padding by saying you "enabled AI transformation" or whatever helps you land a big payday at some other company.
More that there is a poor incentive structure. Just like how PE can make money by leveraged buyouts and running businesses into the ground. Many of the financial instruments that make both that and the current AI bubble possible were legal then made illegal within the lifetimes of the last 16 presidents.
Round-tripping used to be regulated. SPVs used to be regulated. If you need a loan you used to have to go to something called a bank, now it comes from ???? who knows drug cartels, child traffickers, blackstone, russians & chinese oligarchs. Even assuming it doesn't collapse tommorow why should they make double digit returns on AI datacenters built on the backs of Americans?
AI is truly perfect for internal tooling. Security is less or no concern, bugs are more acceptable, performance / scalability rarely a concern. Quickest way to get things done, and speed up production development, MVP development etc.
"We are writing down X billions over 4 years, and have cancel several ambitious programs related to our AI experiments. We were following standard practice in the industry, so [shareholders] can't blame us for these chickens coming to roost. If everyone is guilty, is anyone really guilty?"
If security was the prime concern, there would be no chickens and no coop and no farm - people would still be living in caves. After all, outside is dangerous, and Grug Chief said, smart ass grugs with their smart ass ideas like fire or agriculture just invite complexity and create security vulnerabilities.
After all (Grug Chief reminds us), the only truly secure computing system is an inert rock.
No problems at all except, unauthorized access to a model they were claiming was a weapon and couldn't be released to the public and having their cli code leaked in the last two weeks. Everything's just fine
This is what happens when entire industries go all in on "Move fast and break things." Imagine what they said about software applying to everything else in the world. That's what's coming.
> Security is less or no concern, bugs are more acceptable, performance / scalability rarely a concern. Quickest way to get things done
> This is what happens when entire industries go all in on "Move fast and break things." Imagine what they said about software applying to everything else in the world. That's what's coming.
This is literally how rest of the world works already, and always had. We'd still be living in caves otherwise. Fortunately most people (at least outside software) seem to understand that security is a trade-off against usefulness, and not an end goal in itself.
I am, oddly, able to get really quite a lot of mileage out of $20/mo of OpenAI plan, and I have never encountered a usage limit. I have gotten warnings that I was close a couple times.
I wonder what I’m doing differently.
I did spend quite a bit of time, mostly manually, improving development processes such that the agent could effectively check its work. This made a difference between the agent mostly not working and mostly working. Maybe if I had instead spent gobs of money it would have worked output tooling improvements?
I wonder if you're like me? I tried out the MCPs and sub agents and rules and bells and whistles and always just came back to a plain Codex / Claude Code / Cursor Agent terminal window, where I say what I want, @ a few files, let it rip, check the diff, ask for some adjustments, then commit and start the process over after clearing context.
Haven't found a process that beats this yet and I burn very few tokens this way.
I don’t really write code with it at all, and that’s why I burn so many tokens.
I like writing code, I’m good at writing code. What I hate doing is dredging through logs, filtering out test scenarios and putting together disparate information from knowledge silos - so I have the AI doing that. It’s my research assistant.
Effectively I’m using it like an automated search engine that indexes anything I want and refines the results by using the statistical near neighbors of how other people explained their searches.
I’m not them but we have vastly improved our internal pipeline monitoring/triage/root cause/etc by having a new system that basically its whole purpose is to hook into all of our other systems and consolidate it under a single view with an emphasis on shortening the amount of time it takes to triage and refine issues.
This will have previously been too ambitious to ever scope but we’ve been able to build essentially all of it in just two months. Since it sits on top of our other systems and acts as more of a window/pass through control pane, the fact that it’s vibe coded poses little risk since we still have all the existing infrastructure under it if something goes awry.
I guess that's one way to tout a technology as revolutionary without actually needing to provide any proof of it. Just say you're using it for "internal tooling" and "unannounced projects", that way nobody can look at them and notice they're indistinguishable from the slop that clogs up Show HN nowadays.
It's better than the "here's my code, it a giant pile of spaghetti but only luddites care about code quality and maintainability anyway" method, at least.
I'm using it to write frontend code literally 5 times faster. What would have been a shell script is now a GUI backed by an API layer that doesn't require looking up internal documentation to know that it exists.
I've been using it to write tools that drastically facilitate spinning up local k8s cluster with an entire suite of development services that used to take two days to set up in Docker.
I have some coworker who says something similar, he vibe coded tons of cryptic code, which indeed solves some problem though could be way more compact and well structured. Now it is hitting complexity limitation, since llm now cant comprehend it, and human cant comprehend it by large a margin.
it will comprehend it well enough to complicate it further into a rats-nest that only Opus 4.9 can comprehend, and so on. Good luck if you run into a bug before the N+1 version launches.
Its a bit of workspace politics, I would need to call that guy out to tell that he is not hyper-performer, but just pushed lots of low quality code which will produce lots of negative impact in a long term.
Also, I am not sure if it is trivial to implement. The code is injected into many scenarios and workflows, so replacement will be painful and risky if new solution break some edge case.
It sounds like you might have some larger process problems if someone can just inject a bunch of vibe-coded slop into critical workflows while more discerning eyes are dubious of the quality/reliability etc.
In some sense, sure. There’s a lot of processes that weren’t previously needed, because sloppy people who couldn’t or wouldn’t think things through were mostly incapable of producing PRs that passed all the existing tests.
its partially/largely management problem. One of tier1 productivity metric in the group is # of LoC created by engineers, so it creates dynamics of people exchanging favors of pushing AI slop to codebase, or be labeled as low performers.
Exactly. Software quality has become worse, online media has become even more trash than before, and life is otherwise basically the same, lack of jobs notwithstanding. The legitimately useful things regular people can use AI for would be mostly solved by locally run quantized models. This AI "revolution" may be setting several billion on fire without even 1% of that being real value added to the world.
Coding velocity doesn't matter if it the net result is software that sucks massive schlong. The real world doesn't care if programmers can write code faster.
I'm wondering whether the layoffs are partly targeting people who haven't adapted to using AI tools, particularly those who are openly dismissive of AI-assisted work.
Because the job itself has now changed, and they haven't. Their output speed might have been eclipsed by that of the engineers who efficiently adopted the new tooling.
Where I work, the power dynamics have shifted wildly. There are a number of senior engineers who refuse to touch the stuff, and as a result, they can barely keep up with their peers. Some of our juniors are now running laps around them.
When a stranger to your craft can now teach themselves what you know, how to do your job, and even how to automate your tasks in the span of the same workday as you, all while reliably being able to gauge the innacuracy of the output they're reading, how much longer do you really hold relevance?
Because the job changed out from under them - it's now to use AI as much as possible and generate so much and so convoluted content that humans have no chance of keeping up the "velocity" without being entirely dependent on it.
I'm spending a ton of tokens because it insists on manually correcting code that fails the linter, despite the instructions in the AGENTS.md to run the linter with autocorrect.
And also because the Plan agent generates a huge plan, asks me a couple yes/no questions with an obvious answer, and then regenerates the entire plan again. Then the Build agent gets confused anyway and does something else, and I have to round-trip about 5 times with that full context each time.
It's not just code generation, either - more and more people in my own org are using Claude Code for infrastructure automation, devops, etc. Obviously some amount of code in there, but an absolute ton of tokens being consumed just dealing with Kubernetes work at scale.
I work in a large infrastructure design consultancy. There are massive unmet needs for narrow-purpose automation in project delivery that are being developed by the users that need them for a few pennies at a time. Previously, getting any project of any size done by the in-house developers required 6-24 months of alignment-building with multiple levels of management in multiple business units. Getting anything done comes with an administrative price tag in the $10,000s. Those people spent all their time on a few outward-facing products and building dashboards for EVPs and C-level people, and the people doing real work our customers actually care about are banging rocks together and fingerpainting walls of their caves with excel and email based workflow.
Now there are pockets of people who are extremely productive, and maybe 80-90% of the rest who will never adapt. When I mean extremely, I mean people producing weeks of effort of marketing teams and hundreds of unbillable hours of senior-level professionals with only a few minutes of human involvement. Paid software extensions for (awful) design software we are required by clients to use can all be duplicated in house. Technical leadership is now aware of this, and our spend on software licenses is going to drop fast. I think every project in my own portfolio has some kind of custom automation supporting it, which was unthinkable 5 years ago.
It's going to take years for practical knowledge of how to use these systems to spread and even more for market discipline to expel those who cannot or will not learn. The Industrial Revolution took nearly a century, depending on how you're counting. LLMs have only been producing coherent output in the last 5 years. They've only been as good or better than people at some things you would have done on a computer for about a year or so. Be patient. These are massive changes.
>What is all this AI doing? People are spending 10’s to 100’s of billions and no service or technology seems better or cheaper. Everything is more expensive and worse.
That "more expensive" is someone's revenue. May be AI is the kind of technology that allows to make more and more revenue by making things more expensive and worse than by making them better and cheaper.
This is my take away too. I see some interesting toys here and there, but not much of substance. Meanwhile all the GitHub issues I follow for open source projects have slowed to a halt, the products I use have no significant updates. Even AI products are slow to improve their interfaces.
It's a great tool, and at 1/10 or 1/100th the cost of actual developers. In the context of yc I guess watch out getting re-disrupted by a smaller team faster than before. But that's really the trend the past 40 years so nothing is new. Well maybe the velocity combined with us loosing it's footing at the same time.
But yea it's not gonna make facebook 20% better tomorrow just that you need 5 people instead of 40 to build the next facebook.
You seem to be under the impression that making services better or cheaper _for the consumer_ is the goal of any corporation. The goal is to make their own operations better and cheaper for them. They are laying off employees and adding features of questionable value as a pretext to raise prices. The playbook has not changed, it has only accelerated.
Real-world counter-anecdote from mid-market finance (sub-$500M revenue, 2-3k employees). On a customer's JD Edwards data, a playbook run via MCP returns AP aging by company plus the three vendors driving the variance plus the prior-month delta plus citations to the JDE F0411 rows in roughly 30 seconds. The same in Excel takes ~45 minutes. That's not 100x, it's not transformative, and it's not autonomous. It's one analyst hour reclaimed per query. Multiply by ~80 queries a month and a $50/hour fully loaded cost and the math is mundane and positive. Value is real and unsexy. The marketing has compressed all the unsexy ROI into the same sentence as the trillion-dollar promises. (Disclosure: I co-founded eyko, the platform doing this.)
For myself, its a massive boost when solo developing. Perhaps this is a different use case than most. It can work across multiple programming languages and frameworks that I had zero experience in. I use my existing knowledge of programming to ensure the new code written is correct. Also it really excels at translating from one language/framework to another. I can spend time getting it working well in a platform I know then just ask it to convert to another platform. It gets it 90% right in the first prompt, then its just a matter of fine-tuning, reviewing etc. This last 10% is where I supercharge my learning on those languages/framework. To lean all the new languages and frameworks would have taken me months before I would be productive. Now with a single prompt, we get 90% of the way there. That is incredible value for us.
And yet.. building shit is no longer the sole domain of the software engineer.
That's the sea change.
I've literally had finance and GTM stand things up for themselves in the last few weeks. A few tweaks (obviously around security and access), and they are good to go.
They've gone from wrangling spreadsheets to smooth automated workflows that allow them to work at a higher level in a matter of months.
That's what all this AI is doing. The shit we could never get the time to get around to doing.
The only thing that matters is the impact on the financials. The shareholders (the people who employ you) dont care about any of this if it does not enhance value.
I can say in one role in my job, I'm getting a lot of use and I know my colleagues are at least trying a lot of things. One use is a first-pass review of animal care and use protocols. The Claude project was given all of the relevant policies and guidelines as well as a fairly long prompt that explains the things we look for in protocol review. It's checking some things that the software we use makes very tedious to check and raising inconsistencies between sections. Some places have a full time "protocol reader" who does this kind of first check, but we've never had that, so it's helpful.
Another project I'm seeing in the same realm is taking an approved protocol and some study results and checking that the records of what was done match what they said they could do in the approved protocol. It can also make sure that surgical records have all the things they should have. This can help meet one of the requirements from the national accreditation organization to do "post approval monitoring".
Another way I've used it is to have it collate and compare a particular kind of policy across many institutions who transparently put their policies online. Seeing the commonality between the policies and where some excel helped me rewrite our policy.
This is work that just wasn't happening before or, more accurately, it was being spread over lots of people, and any improvement in efficiency or consistency is hard to measure.
They're pointing out that run-rate revenue is based on essentially sampling revenue over some limited time interval, then extrapolating from there assuming revenue always occurs at the same rate (or greater) over all similar intervals in the future. More specifically, they're pointing out that estimates of ARR derived from this kind of sampling are fundamentally prone to error and can be arbitrarily inflated based on how the time interval is sampled.
Of course, but the fact of the matter is that the same technique was used for the quarter prior to that, and there’s a 3x increase quarter over quarter.
As far as I understand run rate revenue is just a fancy way of saying that "the last month we had sales, and if that continues for a year we will have a AAR of 30B. meaning it's not 30B yet, but the sales numbers indicates that we get there by continue selling at the current speed. But to have revenue of $100 and get $30B in ARR I guess the period looked at needs to be seconds....
(Run Rate = Revenue in Period / # of Days in Period x 365)
Not even that. It's not based on actual sales in, for example, the past month. It's based on an expected continuous growth based on the growth of the past month (or whatever period you pick).
I don't follow Anthropic closely enough to know what claims its CEO has made, but it is factual that Altman is a pathological liar. You can observe this for yourself by reading and listening to the things he says and then comparing them to reality. We have years of evidence to look back on. The chasm between Altman's reality and everyone else's is so large and so well-known that it was one of the chief factors cited by the board when he was fired.
(I would then argue that he was re-hired specifically because others involved with OpenAI understood that it is literally his job to lie and that OpenAI would not be where it is today as a corporate behemoth rather than a research non-profit without a world-class liar marketing it, but that is merely conjecture.)
I mean.. kinda everything about Mythos for example? Anthropic has a good product, but they also pretty consistently say some stupid ass shit if you're being generous, and blatant lies if you aren't
I agree about the core motivation behind these deals, however I'm skeptical as to how "suddenly" we'll see substantial improvements. Despite their size, I'd be surprised if Google or Amazon had uncommitted chunks of Anthropic-scale, top-tier AI compute sitting around waiting to be activated.
They're already over-subscribed and waiting for new data centers (and power plants) to come online. I suspect Anthropic will get a modest amount of new capacity right away with more added over coming quarters. These two deals don't change the total amount of AI compute available on planet Earth over the next 18 months. Anthropic parting with high-value equity has now made them the new highest bidder for an already over-bid resource. I suspect the net impact will be Amazon & Google pushing prices even higher on everyone else as they reallocate compute to their new top whale.
> Despite their size, I'd be surprised if Google or Amazon had uncommitted chunks of Anthropic-scale, top-tier AI compute sitting around waiting to be activated.
I doubt it was idle capacity. But for a chunk of equity in Anthropic I imagine they are willing to deprioritize other, possibly internal, uses. Certainly anything that's not contractually obligated could be on the chopping block.
When people are moving from other models (e.g. GPT, Gemini, etc) the compute that is previously powering that inference now becomes available. Of course, I'm certainly doubtful that Google would break commits and give OpenAI GPUs to Anthropic, but the underlying effect is present and probably sorted out somehow. It's not completely net new compute for the world.
It seems like they shifted heavily to prioritizing enterprise users. Starting in the last day or two I started getting much faster responses on an enterprise plan.
Perhaps the adversity of the contracts cancels out with their sudden success and increase in valuation and it ends up a wash compared to the counterfactual scenario where they would have speculated on high growth early on.
another llm user exhibiting the behavior of a gambling addict "That table is cold this week" "that machine is't hot, you wont win on it" Every month llm users are coming up with reasons why they think the model is under preforming from some occulted reasoning only they perceive. Maybe you're just frying your brain by using llms the way you do.
They should probably look at moving away from general purpose hardware for their actual products, and reserve GP hardware for RnD. You don't need frontier nodes to run circles around GPGPUs, an ASIC made with 28nm is more than enough to embarrass an H100 (and much cheaper)
AI is in such desperate need to adopt software-hardware co-development practices, it's infuriating watching the industry drag its feet about it. We are wasting so much electricity and absolutely wrecking the "free" market just because these companies are incentivized to work at an unsustainable breakneck speed in getting shit to market.
Well to a certain extent it also blunts competition, Gemini is less of a threat if their main investor is also backing Anthropic. The issue is when the pyramid scheme collapses...
Both Amazon and Google provide the Claude models via their Kiro and Antigravity IDEs respectively. It could also be investing in their attempt to own the IDE space.
also if youre in a large organization, this is a great way to sabotage other peoples projects while elevating your stature. Require that they go evaluate alternatives and prior art, and write a slew of analysis and decision documentation
its not “normal” when companies have 10s of Billion in net profit per quarter
Axing low/negative ROI product lines, sure. But recently these cuts have been across-the-board and in product lines that are net profitable and have strong technical product roadmaps. Moreover they are firing longer tenured (expensive) engineers
I understand they’re managing a transition to a capital intensive strategy but the whole era reeks of stock price focused financial engineering and these large companies flexing oligopoly power in the face of their customers and the labor that builds their technology.
true but the government will inevitably demand their own stanza of (blocking) system prompts in the major AI services. then they will ban local LLM and foreign ones.
Self-driving is possible but it requires a massive sustained investment in custom hardware on the car, in real and simulation testing, in painstaking software developlment covering tens of thousands of scenarios, realtime remote control failsafes, fleet management capabilities in every city. Waymo is the only company that comes close to the right approach. All these other Elons, GM, Uber CEOs are just jangling shiny objects in front of investors. A moonshot on the financial model for what are otherwise mature stagnant businesses.
reply