Spending Too Much Money on a Coding Agent

allenpike.com

151 points by GavinAnderegg 4 days ago


iamleppert - a day ago

"Now we don't need to hire a founding engineer! Yippee!" I wonder all these people who are building companies that are built on prompts (not even a person) from other companies. The minute there is a rug pull (and there WILL be one), what are you going to do? You'll be in even worse shape because in this case there won't be someone who can help you figure out your next move, there won't be an old team, there will just be NO team. Is this the future?

jbentley1 - a day ago

My Claude Code usage would have been $24k last month if I didn't have a max plan, at least according to Claude-Monitor.

I've been using a tool I developed (https://github.com/stravu/crystal) to run several sessions in parallel. Sometimes I will run the same prompt multiple times and pick the winner, or sometimes I'll be working on multiple features at once, reviewing and testing one while waiting on the others.

Basically, with the right tooling you can burn tokens incredibly fast while still receiving a ton of value from them.

rogerkirkness - a day ago

Early stage founder here. You have no idea how worth it $200/month is as a multiple on what compensation is required to fund good engineers. Absolutely the highest ROI thing I have done in the life of the company so far.

alwillis - a day ago

After reading many of the comments in this thread, I suspect many (not all) issues come from lack of planning and poor prompting.

For anything moderately complex, use Claude's plan mode; you get to approve the plan before turning it loose. The planning phase is where you want to use a more sophisticated model or use extended thinking mode.

Once you have a great plan, you can use a less sophisticated model to execute it.

Even if you're a great programmer, you may suck at prompting. There's an art and a science to prompting; perhaps learn about it? [1]

Don't forget; in addition to telling Claude or any other model what to do, you can also tell them what not to do in the CLAUDE.md or equivalent file.

[1]: https://docs.anthropic.com/en/docs/build-with-claude/prompt-...

lvl155 - a day ago

Is $200/month a lot of money when you can multiply your productivity? It depends but the most valuable currency in life is time. For some, spending thousands a month would be worth it.

chis - a day ago

Has anyone else done this and felt the same? Every now and then I try to reevaluate all the models. So far it still feels like Claude is in the lead just because it will predictably do what I want when given a mid-sized problem. Meanwhile o3 will sometimes one-shot a masterpiece, sometimes go down the complete wrong path.

This might also just be a feature of the change in problem size - perhaps the larger problems that necessitate o3 are also too open-ended and would require much more planning up front. But at that point it's actually more natural to just iterate with sonnet and stay in the driver's seat a bit. Plus sonnet runs 5x faster.

feintruled - a day ago

Interesting. Though it seems they are themselves building Agentic AI tooling. It's vibe coding all the way down - when's something real going to pop out the bottom?

jasonthorsness - a day ago

I really hope we can avoid metered stuff for the long-term. One of the best aspects of software development is the low capital barrier to entry, and the cost of the AI tools right now is threatening that.

I'm fortunate in that my own use of the AI tools I'm personally paying for is squished into my off-time on nights and weekends, so I get buy with a $20/month Claude subscription :).

bicepjai - a day ago

I can see how pricing at 100 to 200$ per month per employee could make sense for companies, it’s a clear value proposition at that scale. But for personal projects and open source work, it feels out of reach. I’d really like to see more accessible pricing tiers for individuals and hobbyists. Pay per token models don’t work for me either; earlier this year, I racked up almost $1,000 in a month just experimenting with personal projects, and that experience has made me wary of using these tools since.

Sources

quonn - a day ago

Charging $200/month is economically only possible if there is not a true market for LLMs or some sort of monopoly power. Currently there is no evidence that this will be the case. There are already multiple competitors and the barrier to entry is relatively low (compared to e.g. the car industry or other manufacturing industries), there are no network effects (like for social networks) and no need to get the product 100% right (like compatibility to Photoshop or Office) and the prices for training will drop further. Furthermore $200 is not free (like Google).

Can anyone name one single widely-used digital product that does _not_ have to be precisely correct/compatible/identical to The Original and that everyone _does_ pay $200/month for?

Therefore, should prices that users pay get anywhere even close to that number, there will naturally be opportunities for competitors to bring prices down to a reasonable level.

pshirshov - a day ago

> Use boring technology: LLMs do much better with well-documented and well-understood dependencies than obscure, novel, or magical ones. Now is not the time to let Steve load in a Haskell-to-WebAssembly pipeline.

If we all go that way, there might be no new haskells and webassemblies in the future.

tomjuggler - 13 hours ago

I believe this story, but being from a third world country it's not feasible to spend anywhere near that much. Also, as others have mentioned I am wary of "rug-pulls" when it comes to proprietary models and services. That is why I am all-in on Deepseek currently, with Aider (and Roo for MCP integration). When the main api is lagging I just switch to the same model with a different provider on OpenRouter. Theoretically I could host my own if I got that busy.

Solo developer doing small jobs but I code every day and $10 per month would be a busy month for me. I still read every line of code though..

georgeecollins - a day ago

I am blown away that you can get a founding engineer for $10k / month. I guess that is not counting stock options, in which case it makes sense. But I think if you include options the opportunity cost is much higher. IMO great engineers are worth a lot, no shade.

hoistbypetard - a day ago

> literally changing failing tests into skipped tests to resolve “the tests are failing.”

Wow. It really is like a ridiculous, over-confident, *very* junior developer.

mathiaspoint - a day ago

I can't imagine using something like this and not self hosting. Moving around in your editor costs money? That would completely crush my velocity.

nico - a day ago

The article reads almost like an ad for o3 and spending a lot of money on LLM APIs

In my experience, o4-mini-high is good enough, even just through the chat interface

Cursor et al can be more comfy because they have access to the files directly. But when working on a sufficiently large/old/complex code base, the main limitation is the human in the loop and managing context, so things end up evening out. Not only that, but a lot of times it’s just easier/better to manually feed things to ChatGPT/Claude - that way you get to more carefully curate and understand the tasks and the changes

I still haven’t seen any convincing real life scenario with larger in-production code bases in which agents are able to autonomously write most of the code

If anyone has a video/demo, would love to see it

tabs_or_spaces - a day ago

Since this is a business problem.

* It's not clear on how much revenue or new customers is generated by using a coding agent

* It's not clear on how things are going on production. There's only talks about development in the article

I feel ai coding agents will give you the edge. Just this article doesn't talk about revenue or PnL side of things, just perceived costs saved from not employing an engineer.

scrubs - 16 hours ago

The one thing AI hates and doesn't want you to know about AI:

Having typing skills >= 120 wpm will triple your efficacy.

cmrdporcupine - a day ago

I get a lot of value out of Claude Max at $100 USD/month. I use it almost exclusively for my personal open source projects. For work, I'm more cautious.

I worry, with an article like this floating around, and with this as the competition, and with the economics of all this stuff generally... major price increases are on the horizon.

Businesses (some) can afford this, after all it's still just a portion of the costs of a SWE salary (tho $1000/m is getting up there). But open source developers cannot.

I worry about this trend, and when the other shoe will drop on Anthropic's products, at least.

suralind - a day ago

How does GitHub Copilot stack against API access directly from OpenAI, etc.? Is it faster to use API keys than Copilot?

v5v3 - a day ago

No need to use the most expensive models for every query? Use it for the ones the cheaper models don't do well.

butlike - a day ago

I love how paying for prompts stuck. Like, if someone's going to do your homework for you, they should get compensated.

suninsight - a day ago

So what we do at NonBioS.ai is to use a cheaper model to do routine tasks, but switch to a higher thinking model seamlessly if the agent get stuck. Its most cost efficient, and we take that switching cost away from the engineer.

But broadly agree to the argument of the post - just spending more might still be worth it.

andrewstuart - a day ago

I must be holding OpenAI wrong.

Everyone time I try it I find it to be useless compared to Claude or Gemini.

nickjj - a day ago

Serious question, how do you justify paying for any of this without feeling like it's a waste?

I occasionally use ChatGPT (free version without logging in) and the amount of times it's really wrong is very high. Often times it takes a lot of prompting and feeding it information from third party sources for it to realize it has incorrect information and then it corrects itself.

All of these prompts would be using money on a paid plan right?

I also used Cursor (free trial on their paid plan) for a bit and I didn't find much of a difference. I would say whatever back-end it was using was possibly worse. The code it wrote was busted and over engineered.

I want to like AI and in some cases it helps gain insight on something but I feel like literally 90% of my time is it prodiving me information that straight up doesn't work and eventually it might work but to get there is a lot of time and effort.

throwaway984393 - a day ago

[dead]

deadbabe - a day ago

I find it kind of boggling that employers spend $200/month to make employees lives easier, for no real gain.

That’s right. Productivity does go up, but most of these employees aren’t really contributing directly to revenue. There is no code to dollar pipeline. Finishing work faster means some roadmap items move quicker, but they just move quicker toward true bottlenecks that can’t really be resolved quickly with AI. So the engineers sit around doing nothing for longer periods of time waiting to be unblocked. Deadlines aren’t being estimated tighter, they are still as long as ever.

Enjoy this time while it lasts. Someday employers might realize they need to hire less and just cram more work into individual engineers schedules, because AI should supposedly make work much easier.

delduca - a day ago

I just pay $20/month on ChatGPT and spend the entire day coding with its help, no need to pay for tokens, no need to integrate it on your IDE.