IMO other than the Microsoft IP issue, I think the biggest thing that has shifted since this acquisition was first in the works is Claude Code has absolutely exploded. Forking an IDE and all the expense that comes with that feels like a waste of effort, considering the number of free/open source CLI agentic tools that are out there.
Let's review the current state of things:
- Terminal CLI agents are several orders of magnitude less $$$ to develop than forking an entire IDE.
- CC is dead simple to onboard (use whatever IDE you're using now, with a simple extension for some UX improvements).
- Anthropic is free to aggressively undercut their own API margins (and middlemen like Cursor) in exchange for more predictable subscription revenue + training data access.
What does Cursor/Windsurf offer over VS Code + CC?
- Tab completion model (Cursor's remaining moat)
- Some UI niceties like "add selection to chat", and etc.
Personally I think this is a harbinger of where things are going. Cursor was fastest to $900M ARR and IMO will be fastest back down again.
Agreed on everything. Just to add, not only anthropic is offering CC at like a 500% loss, they restricted sonnet/opus 4 access to windsurf, and jacked up their enterprise deal to Cursor. The increase in price was so big that it forced cursor to make that disastrous downgrade to their plans.
I think only way Cursor and other UX wrappers still win is if on device models or at least open source models catch up in the next 2 years. Then i can see a big push for UX if models are truly a commodity. But as long as claude is much better then yes they hold all the cards. (And don't have a bigger company to have a civil war with like openai)
The way I am doing the math with my Max subscription and assuming DeepSeek API prices, it is still x5 times cheaper. So either DeepSeek is losing money (unlikely) or Anthropic is losing lots of money (more likely). Grok kinda confirms my suspicions. Assuming DeepSeek prices, I've probably spent north of $100 of Grok compute. I didn't pay Grok or Twitter a single cent. $100 is a lot of loss for a single user.
I'm also curious about this. Claude Code feels very expensive to me, but at the same time I don't have much perspective (nothing to compare it to, really, other than Codex or other agent editors I guess. And CC is way better so likely worth the extra money anyway)
Pretty easy to hit $100 an hour using Opus on API credits. The model providers are heavily subsidized, the datacenters appear to be too. If you look at the Coreweave stuff and the private datacenters it starts looking like the telecom bubble. Even Meta is looking to finance datacenter expansion - https://www.reuters.com/business/meta-seeks-29-billion-priva...
The reason they are talking about building new nuclear power plants in the US isn't just for a few training runs, its for inference. At scale the AI tools are going to be extremely expensive.
Also note China produces twice as much electricity as the United States. Software development and agent demand is going to be competitive across industries. You may think, oh I can just use a few hours of this a day and I got a week of work done (happens to me some days), but you are going to end up needing to match what your competitors are doing - not what you got comfortable with. This is the recurring trap of new technology (no capitalism required.)
There is a danger to independent developers becoming reliant on models. $100-$200 is a customer acquisition cost giveaway. The state of the art models probably will end up costing hourly what a human developer costs. There is also the speed and batching part. How willing is the developer to, for example, get 50% off but maybe wait twice as long for the output. Hopefully the good dev models end up only costing $1000-$2000 a month in a year. At least that will be more accessible.
Somewhere in the future these good models will run on device and just cost the price of your hardware. Will it be the AGI models? We will find out.
I wonder how this comment will age, will look back at it in 5 or 10 years.
The SOTA models will always run in data centers, because they have 5x or more VRAM and 10-100x the compute allowance. Plus, they can make good use of scaling w/ batch inference which is a huge power savings, and which a single developer machine doesn’t make full use of.
Not really, it's possible with any market economy, even a hypothetical socialist one (that is, one where all market actors are worker-owned co-ops).
And, since there is no global super-state, the world economy is a market economy, so even if every state were a state-owned planned economy, North Korea style, still there would exist this type of competition between states.
Consider also that VC funds often have pension funds as their limited partners. Workers have a claim to their pension, and thus a claim to the startup returns that the VC invests in.
So yeah it basically comes down to your definition of "worker-owned". What fraction of worker ownership is necessary? Do C-level execs count as workers? Can it be "worker-owned" if the "workers" are people working elsewhere?
Beyond the "worker-owned" terminology, why is this distinction supposed to matter exactly? Supposing there was an SV startup that was relatively generous with equity compensation, so over 50% of equity is owned by non-C-level employees. What would you expect to change, if anything, if that threshold was passed?
Can you give me an idea of how much interaction would be $50-$100 per day? Like are you pretty constantly in a back and forth with CC? And if you wouldn’t mind, any chance you can give me an idea of productivity gains pre/post LLM?
Yes, a lot of usage, I’d guess top 10% among my peers. I do 6-10hrs of constant iterating across mid-size codebases of 750k tokens. CC is set to use Opus by default, which further drives up costs.
Estimating productivity gains is a flame war I don’t want to start, but as a signal: if the CC Max plan goes up 10x in price, I’m still keeping my subscription.
I maintain top-tier subscription to every frontier service (~$1k/mo) and throughout the week spend multiple hours with each of Cursor, Amp, Augment, Windsurf, Codex CLI, Gemini CLI, but keep on defaulting to Claude Code.
I’m a founder/CTO of an enterprise SaaS, and I code everything from data modeling, to algos, backend integrations, frontend architecture, UI widgets, etc. All in TypeScript, which is perfectly suited to LLMs because we can fit the types and repo map into context without loading all code.
As to “why”: I’ve been coding for 25 years, and LLMs is the first technology that has a non-linear impact on my output. It’s simultaneously moronic and jaw-dropping. I’m good at what I do (eg, merged fixes into Node) and Claude/o3 regularly finds material edge cases in my code that I was confident in. Then they add a test case (as per our style), write a fix, and update docs/examples within two minutes.
I love coding and the art&craft of software development. I’ve written millions of lines of revenue generating code, and made millions doing it. If someone forced me to stop using LLMs in my production process, I’d quit on the spot.
Why not self host: open source models are a generation behind SOTA. R1 is just not in the same league as the pro commercial models.
> data modeling, to algos, backend integrations, frontend architecture, UI widgets, etc. All in TypeScript, which is perfectly suited to LLMs because we can fit the types and repo map into context without loading all code.
Which frameworks & libraries have you found work well in this (agentic) context? I feel much of the js lib. landscape does not do enough to enforce an easily-understood project structure that would "constrain" the architecture and force modularity. (I might have this bias from my many years of work with Rails that is highly opinionated in this regard).
> If someone forced me to stop using LLMs in my production process, I’d quit on the spot.
Yup 100% agree. I’d rather try to convince them of the benefits than go back to what feels like an unnecessarily inefficient process of writing all code by hand again.
And I’ve got 25+ years of solid coding experience. Never going back.
When you say generation behind, can you give a sense of what that means in functionality per your current use? Slower/lower quality, it would take more iterations to get what you want?
I am curious what kind of development you’re doing and where your projects fall on the fast iteration<->correctness curve (no judgment). I’ve used CC Pro for a few weeks now and I will keep it, it’s fantastically useful for some things, but it has wasted more of my time than it saved when I’ve experimented with giving it harder tasks.
It's interesting to work with a number of people using various models and interaction modes in slightly different capacities. I can see where the huge productivity gains are and can feel them, but the same is true for the opposite. I'm pretty sure I lost a full day or more trying to track down a build error because it was relatively trivial fpr someone to ask CC or something to refactor a ton of files, which it seems to have done a bit too eagerly. On the other hand, that refactor would have been super tedious, so maybe worth it?
The project is just a web backend. I give Claude Code grunt work tasks. Things like "make X operation also return Y data" or "create Z new model + CRUD operations". Also asking it to implement well-known patterns like denouncing or caching for an existing operation works well.
My app builds and runs fine on Termux, so my CLAUDE.md says to always run unit tests after making changes. So I punch in a request, close my phone for a bit, then check back later and review the diff. Usually takes one or two follow-up asks to get right, but since it always builds and passes tests, I never get complete garbage back.
There are some tasks that I never give it. Most of that is just intuition. Anything I need to understand deeply or care about the implementation of I do myself. And the app was originally hand-built by me, which I think is important - I would not trust CC to design the entire thing from scratch. It's much easier to review changes when you understand the overall architecture deeply.
Unless you're getting paid for your commute, you're just giving your employer free productivity. I would recommend doing literally anything else with that time. Read a book, maybe.
you can easily reach 50$ per day.
by force switching model to opus
/model opus
it will continue to use opus eventhough there is a warning about approaching limit.
i found opus is significantly more capable in coding than sonnet, especcially for the task that is poorly defined, thinking mode can fulfill alot of missing detail and you just need to edit a little before let it code.
Claude Code with a Claude subscription is the cheap version for current SOTA.
"Agentic" workflows burn through tokens like there's no tomorrow, and the new Opus model is so expensive per-token that the Max plan pays itself back in one or two days of moderate usage. When people reports their Claude Code sessions costing $100+ per day, I read that as the API price equivalent - it makes no sense to actually "pay as you go" with Claude right now.
This is arguably the cheapest option available on the market right now in terms of results per dollar, but only if you can afford the subscription itself. There's also time/value component here: on Max x5, it's quite easy to hit the usage limits of Opus (fortunately the limit is per 5 hours or so); Max x20 is only twice the price of Max x5 but gives you 4x more Opus; better model = less time spent fighting with and cleaning up after the AI. It's expensive to be poor, unfortunately.
Claude Code pro is ~$20USD/ month and is nearly enough for someone like me who can’t use it at work and is just playing around with it after work. I’m loving it.
Zed is fantastic. Just dipping my toes in agentic AI, but I was able to fix a failing test I spent maybe 15 minutes trying to untangle in a couple minutes with Zed. (It did proceed to break other tests in that file though, but I quickly reverted that.)
It is also BYOA or you can buy a subscription from Zed themselves and help them out. I currently use it with my free Copilot+ subscription (GitHub hands it out to pretty much any free/open source dev).
You can tell Claude Code to use opus using /model and then it doesn't fall back to Sonnet btw. I am on the $100 plan and I hit rate-limits every now and then, but not enough to warrant using Sonnet instead of Opus.
This is what I don’t get about the cost being reported by Claude code. At work I use it against our AWS Bedrock instance, and most sessions will say 15/20 dollars and I’ll have multiple agents running. So I can easily spend 60 bucks a day in reported cost. Our AWS Bedrock bill is only a small fraction of that? Why would you over charge on direct usage of your API?
Seems like the survival strategy for cursor would be to develop their own frontier coding model. Maybe they can leverage the data from their still somewhat significant lead in the space to make a solid effort.
I don’t think that’s a viable strategy. It is very very hard and not many people can do it. Just look at how much Meta is paying to poach the few people in the world capable of training a next gen frontier model.
The basic concept plus a lot of money spent on compute and training data gets you pretraining. After that to get a really good model there’s a lot more fine-tuning / RL steps that companies are pretty secretive about. That is where the “smart decisions” and knowledge gained by training previous generations of sota models comes in.
We’d probably see more companies training their own models if it was cheaper, for sure. Maybe some of them would do very well. But even having a lot of money to throw at this doesn’t guarantee success, e.g. Meta’s Llama 4 was a big disappointment.
That said, it’s not impossible to catch up to close to state-of-the-art, as Deepseek showed.
I'd recommend reading some of the papers on what it takes to actually train a proper foundation model, such as the Llama 3 Herd of Models paper. It is a deeply sophisticated process.
Coding startups also try to fine-tune OSS models to their own ends. But this is also very difficult, and usually just done as a cost optimization, not as a way to get better functionality.
I also just prefer CC's UX. I've tried to make myself use Copilot and Roo and I just couldn't. The extra mental overhead and UI context-switching took me out of the flow. And tab completion has never felt valuable to me.
But the chat UX is so simple it doesn't take up any extra brain-cycles. It's easier to alt-tab to and from; it feels like slacking a coworker. I can have one or more terminal windows open with agents I'm managing, and still monitor/intervene in my editor as they work. Fits much nicer with my brain, and accelerates my flow instead of disrupting it
There's something starkly different for me about not having to think about exactly what context to feed to the tool, which text to highlight or tabs to open, which predefined agent to select, which IDE button to press
Just formulate my concepts and intent and then express those in words. If I need to be more precise in my words then I will be, but I stay in a concepts + words headspace. That's very important for conserving my own mental context window
Their base is $20/mth. That would equal 3.75M people paying a sub to Cursor.
If literally everyone is on their $200/mth plan, then that would be 375K paid users.
There’s 50M VS Code + VS users (May 2025). [1] 7% of all VS Code users having switched to Cursor does not match my personal circle of developers. 0.7% . . . Maybe? But, that would be if everyone using Cursor were paying $200/month.
Seems impossibly high, especially given the number of other AI subscription options as well.
The $20/month cursor sub is heavily limited though, for basic casual usage that's fine but you VERY soon run into its limits when working at any speed.
- Forking VSCode is very easy; you can do it in 1 hour.
- Anthropic doesn't use the inputs for training.
- Cursor doesn't have $900M ARR. That was the raise. Their ARR is ~$500m [1].
- Claude Code already support the niceties, including "add selection to chat", accessing IDE's realtime warnings and errors (built-in tool 'ideDiagnostics'), and using IDE's native diff viewer for reviewing the edits.
The cost of vscode fork is that microsoft has restricted extension marketplace for forks. You have to maintain separate one, that is the real dealbreaker
Claude Code is just proving that coding agents can be successful. The interface isn’t magic, it just fits the model and integrates with a system in all the right ways. The Anthropic team for that product is very small comparatively (their most prolific contributor is Claude), and I think it’s more of a technology proof than a core competency - it’s a great API $ business lever, but there’s no reason for them to try and win the “agentic coding UI” market. Unless Generative AI flops everywhere else, these markets will continue to emerge and need focus. The Windsurf kerfuffle is further proof that OpenAI doesn’t see the market as must-win for a frontier model shop.
And so I’d say this isn’t a harbinger of the death of Cursor, instead proof that there’s a future in the market they were just recently winning.
I was being hyperbolic saying their ARR will go to zero. That's obviously not the case, but the point is that CC has revealed their real product was not "agentic coding UI", it was "insanely cheap tokens". I have no doubt they will continue to see success, but their future right now looks closer to being a competitor to free/open tools like cline/roo code, as well as the CLI entrants, not a standalone $500M ARR juggarnaut. They have no horse in the race in the token market, they're a middleman.
They either need to create their own model and compete on cost, or hope that token costs come down dramatically so as to be too cheap to meter.
I think CC is just far more useful; I use it for literally everything and without MCP (except puppeteer sometimes) as it just writes python/bash scripts to do that far better than all those hacked together MCP garbage bins. It controls my computer & writes code. It made me better as well as now I actually write code, including GUI/web apps, that's are always fully scriptable. It helps me, but it definitely helps CC; it can just interrogate/test everything I make without puppeteer (or other web browser control, which is always brittle as hell).
Cursor see it coming - it's why they're moving to the web and mobile[0]
The bigger issue is the advantage Anthropic, Google and OpenAI have in developing and deploying their own models. It wasn't that long ago that Cursor was reading 50 lines of code at a time to save on token costs. Anthropic just came out and yolo'd the context window because they could afford to, and it blew everything else away.
Cursor could release a cli tomorrow but it wouldn't help them compete when Anthropic and Google can always be multiples cheaper
> Anthropic just came out and yolo'd the context window because they could afford to
I don’t think this is true at all. The reason CC is so good is that they’re very deliberate about what goes in the context. CC often spends ages reading 5 LOC snippets, but afterwards it only has relevant stuff in context.
Heard a lot of this context bs parroted all over HN, don't buy it. If simply increasing context size can solve problem, Gemini would be the best model for everything.
I think this is an interesting and cool direction for Cursor to be going in and I don't doubt something like this is the future. But I have my doubts whether it will save them in the short/medium term:
- AI is not good enough yet to abandon the traditional IDE experierence if you're doing anything non-trivial. Hard finding use cases for this right now.
- There's no moat here. There are already a dozen "Claude Code UI" OSS projects with similar basic functionality.
Strictly speaking about large, complex, sprawling codebases, I don't think you can beat the experience that an IDE + coding agent brings with a terminal-based coding agent.
Auto-regressive nature of these things mean that errors accumulate, and IDEs are well placed to give that observability to the human, than a coding agent. I can course correct more easily in an IDE with clear diffs, coding navigation, than following a terminal timeline.
You can view and navigate the diffs made by the terminal agent in your IDE in realtime, just like Cursor, as well as commit, revert, etc. That’s really all the “integration” you need.
Some excellent points. On “add selection to chat”, I just want to add that the Claude Code VS code extension automatically passes the current selection to the model. :)
I am genuinely curious if any Cursor or Windsurf users who have also tried Claude Code could speak to why they prefer the IDE-fork tools? I’ve only ever used Claude Code myself - what am I missing?
Cursor's tab completion model is legitimately fantastic and for many people is worth the entire $20 subscription. Lint fixes or syntax-level refactors are guessed and executed instantly with TAB with close to 100% accuracy. This is their final moat IMO, if Copilot manages to bring their tab completion up to near parity, very little reason to use Cursor.
Idk. When you're doing something it really gets it's super nice, but it's also off a lot of times and it's IMO super distracting when it constantly pop up. No way to explicitly request it instead - other than toggling, which seems to also turn off context/edit tracking, because after toggling on it does not suggest anything until you make some edits.
While Zed's model is not as good the UI is so much better IMO.
Just to offer a different perspective, I use Cursor at work and, coming from emacs (which I still use) with copilot completions only when I request them with a shortcut, Cursor’s behavior drives me crazy.
Which Emacs Package do you use for CoPilot, i tried using Copilot.el a long while ago, but had problems with it. Is there something new or does copilot.el fulfill your needs?
I haven't used Cursor or Claude much, how different is it from Copilot? I bounce between desktop ChatGPT (which can update VS Code) and copilot. Is there an impression that those have fallen behind?
It's quite interesting how little the Cursor power users use tab. Majority of the posts are some insane number of agent edits and close to (or exactly) 0 tabs.
At my company we have an enterprise subscription and we're also all allowed to see the analytics for the entire company. Last I checked, I was literally the number one user of Tab and middle of the pack for agent.
It's interesting when I see videos or reddit posts about cursor and people getting rate limited and being super angry. In my experience tab is the number one feature, and I feel like most people using agent are probably overusing it tasks that would honestly take less time to do myself or using models way smarter than they need to be for the task at hand.
I'd like to ask the opposite question: why do people prefer command line tools? I tried both and I prefer working in IDE. The main reason is that I don't trust the LLMs too much and I like to see and potentially quickly edit the changes they make. With an IDE, I can iterate much faster than with the command line tool.
I haven't tried Claude Code VS Code extension. Did anyone replaced Cursor with this setup?
I replaced. My opinion: Cursor sucks as an IDE.
Cursor may have a average to above average quality in IDE assistance - but the IDE seems to get in the way. It's entire performance is based on the real-time performance and latency from their servers and sometimes it is way too slow. The TAB autocomplete that was working for you in the last 30 minutes suddenly doesn't work randomly, or just experiences severe delays that it stops making sense.
Besides that, the IDE seems poorly designed - some navigation options are confusing and it makes way too many intrusive changes (ex: automatically finishing strings).
I've since gone back to VS Code - with Cline (with OpenRouter and super cheap Qwen Coder models, Windsurf FREE, Claude Code with $20 per month) and I get great mileage from all of them.
You're looking at (coloured) diffs in your shell is all when it comes to coding. It's pretty easy to setup MCP and have claude be the director. Like I have zen MCP running with an OpenRouter API key, and will ask claude to consult with pro (gemini) or o3, or both to come up with an architecture review / plan.
I honestly don't know how great that is, because it just reiterates what I was planning anyways, and I can't tell if it's just glazing, or it's just drawing the same general conclusions. Seriously though, it does a decent job, and you can discuss / ruminate over approaches.
I assume you can do all the same things in an editor. I'm just comfortable with a shell is all, and as a hardcore Vi user, I don't really want to use Visual Studio.
I also use vim heavily and I've found that I'm really enjoying Cursor + VS Code Vim extension. The cursor tab completion works very nicely in conjunction with vim navigate mode.
I can roll back to different checkpoints with Cursor easily. Maybe CC has it but the fact that I haven’t found it after using it daily is an example of Cursor having a better UX for me.
Does anyone have a comparison between this and OpenAI Codex? I find OpenAI's thing really good actually (vastly better workflow that Windsurf). Maybe I am missing out however.
Codex CLI is very bad, it often struggles to even find the file and goes on a rampage inside the home directory trying to find the file and commenting on random folders. Using o3/o4-mini in Aider is decent though.
Using cline for a bit made me realize cursor was doomed. Everything is just a gpt/anthropic wrapper of fancy prompts.
I can do most of what I want with cline, and I've gone back from large changes to just small changes and been moving much quicker. Large refactors/changes start to deviate from what you actually want to accomplish unless you have written a dissertation, and even then they fail.
I use Augment extensively and find it superior to cursor in every way - and operates as an extension. It has a really handy task planning interface and meta prompt refinement feature and the costs are remarkably low. The quality of output implantation is higher IMO and I don’t have to do a lot of model selection and don’t get Max model bill explosions. If there’s something Cursor provided that Augment doesn’t via extension it was not functionally useful enough to notice.
I think Augment has been flying under the radar for many people, and really reserve better marketing.
I've been using Augment for over a year with IntelliJ, and never understood why my colleagues were all raving about Cursor and Windsurf. I gave Cursor a real try, but it wasn't any better, and the value proposition of having to adopt a dedicated IDE wasn't attractive to me.
A plugin to leverage your existing tools makes a lot more sense than an IDE. Or at least until/if AI agents get so smart that you don't need most of the IDE's functionality, which might change what kinds of tooling are needed when you're in the passenger seat rather than the driver's seat.
Not if you want custom UI. There are a lot of things you can do in extension land (continue, cline, roocode, kilocode, etc. are good examples) but there are some things you can't.
One thing I would have thought would be really cool to try is to integrate it at the LSP level, and use all that good stuff, but apparently people trying (I think there was a company from .il trying) either went closed or didn't release anything note worthy...
IIRC problem is that VS Code does not allow extensions to create custom UI in the panels areas except for WebViews(?). It makes for not a great experience. Plus Cursor does a lot with background indexing to make their tab completion model really good - more than would be possible with the extensions APIs available.
When the Copilot extension needs a new VS Code feature it gets added, but it isn't available to third party extensions until months later... Err, years later... well, whenever Microsoft feels like it.
So an extension will never be able to compete with Copilot.
> with a simple extension for some UX improvements
What are the UX improvements?
I was using the Pycharm plugin and didn’t notice any actual integration.
I had problems with pycharm’s terminal—not least of which was default 5k line scroll back which while easy to change was worst part of CC for me at first.
I finally jumped to using iterm and then using pycharm separately to do code review, visual git workflows, some run config etc.
But the actual value of Pycharm—-and I’ve been a real booster of that IDE has shrank due to CC and moving out of the built in terminal is a threat to usage of the product for me.
If the plugin offered some big value I might stick with it but I’m not sure what they could even do.
#1 improvement for VS Code users is giving the agent MCP tools to get diagnostics from the editor LSPs. Saves a tremendous amount of time having the agent run and rerun linting commands.
I never got the valuation. I (and many others) have built open source agent plugins that are pretty much just as good, in our free time (check out magenta nvim btw, I think it turned out neat!)
I use Windsurf so I remain in the driver's seat. Using AI coding tools too much feels like brain rot where I can't think sharply anymore. Having auto complete guess my next edit as I'm typing is great because I still retain all the control over the code base. There's never any blocks of code that I can't be bothered to look at, because I wrote everything still.
CC would explode even further if they had official Team/Enterprise plan (likely in the work, Claude Code Waffle flag), and worked on Windows without WSL (supposedly pretty easy to fix, they just didn't bother). Cursor learnt the % of Windows user was really high when they started looking, even before they really supported it.
They're likely artificially holding it back either because its a loss leader they want to use a very specific way, or because they're planning the next big boom/launch (maybe with a new model to build hype?).
> What does Cursor/Windsurf offer over VS Code + CC?
Cursor's @Docs is still unparalleled and no MCP server for documentation fetching even comes close. That is the only reason why I still use Cursor, sometimes I have esoteric packages that must be used in my code and other IDEs will simply hallucinate due to not having such a robust docs feature, if any, which is useless to me, and I believe Claude Code also falls into that bucket.
> Cursor's @Docs is still unparalleled and no MCP server for documentation
I strongly disagree. It will put the wrong doc snippets into context 99% of the time. If the docs are slightly long then forget it, it’ll be even worse.
What packages do you use it for? I honestly never had that issue, it's very good in my use cases to find some specific function to call or to figure out some specific syntax.
The forked IDE thing I don't understand either, but...
During the evaluation at a previous job, we found that windsurf is waaaay better than anything else. They were expensive (to train on our source code directly) but the solution the offered was outperforming others.
Windsurf big claim to fame was that you could run their model in airgap and they said they did not train on GPL code. This was an option available for Enterprise customers until they took it away recently to prevent self hosting
just curious because I'm inexperienced with all the latest tools here
> - Tab completion model (Cursor's remaining moat)
What is that? I have Gemini Code Assist installed in VSCode and I'm getting tab completion. (yes, LLM based tab completion)
Which, as an aside I find useful when it works but also often extremely confusing to read. Like say in C++ I type
int myVar = 123
The editor might show
int myVar = 123;
And it's nearly impossible to tell that I didn't enter that `;` so I move on to the next line instead of pressing tab only to find the `;` wasn't really there. That's also probably an easy example. Literally it feels like 1 of 6 lines I type I can't tell what is actually in the file and what is being suggested. Any tips? Maybe I just need to set some special background color for text being suggested.
and PS: that tiny example is not an example of a great tab completion. A better one is when I start editing 1 of 10 similar lines, I edit the first one, it sees the pattern and auto does the other 9. Can also do the "type a comment and it fills in the code" thing. Just trying to be clear I'm getting LLM tab completion and not using Cursor
It gets even worse when all three of IntelliSense, AI completion, and the human are all vying for control of the input. This can be very frustrating at times.
Claude Code is totally different paradigm. You don't edit your files directly so there is no tab autocomplete. It's a chat session.
There are IDE integrations where you can run it in a terminal session while perusing the files through your IDE, but it's not powering any autocomplete there AFAIK.
Yes or running claude code in the cursor/vscode terminal and watching the files change and then reviewing in IDE. I often like to be able to see an entire file when reviewing a diff, rather than just the lines that changed. Plus it's nice to have go-to-definition when reviewing.
Depending on what I'm doing with it I have 3 modes:
Trivial/easy stuff - let it make a PR at the end and review in GitHub. It rarely gets this stuff wrong IME or does anything stupid.
Moderately complex stuff - let it code away, review/test it in my IDE and make any changes myself and tell claude what I've changed (and get it to do a quick review of my code)
Complex stuff - watch it like a hawk as it is thinking and interrupt it constantly asking questions/telling it what to do, then review in my IDE.
Yes, it shows you the file diff. But generally, the workflow is that you git commit a checkpoint, then let it make all the changes it wants freely, then in your IDE, review what has changed since previous commit, iterate the prompts/make your own adjustments to the code, and when you like it, git commit.
Apparently they are, which is crazy to me. Zed agent mode shows modified hunks and you can accept/reject them individually. I can't imagine doing it all through the CLI, it seems extremely primitive.
As far as I can tell, terminal agents are inferior to hosted agents in sandboxed/imaged environments when it comes to concurrent execution and far inferior to assisted ide in terms of UX so what exactly is the point?. The "UI niceties" is the whole point of using cursor and somehow, everyone else sucks at it.
Done. Now you have a SOTA agentic AI with pretty forgiving usage limits up and running immediately. This is why it's capturing developer mindshare. The simplicity of getting up and going with it is a selling point.
Plus it’s straightforward to make Claude Code run agents in parallel/background just like Codex and Cursor, in local sandboxes: https://github.com/dagger/container-use
You’re missing the point tho. The point of the cli agent is that it’s a building block to put this thing everywhere. Look at CCs GitHub plugin, it’s great
CC on github just looks like Codex. I see your point, but it seems like all the big players basically have a CLI agent and most of them think that its just an implementation detail so they dont expose it.
It's another Character.ai situation [0]. Unfortunate for any employees who aren't founders or researchers, as they don't get any payout or a nice new job from this exit structure. In fact they lose their whole time invested at the company.
What a harsh time to work for an AI startup as a rank and file employee! I wonder how the founders justify going along with it inside their mind.
Edit: Thank you @jonny_eh for the clarification. I can't imagine it feels awesome being a leftover but at least you vested out. "Take the money and leave" is still a bit raw when the founders and researchers are now getting the initial payout + generous Google RSU's.
2.5x in 10 months. With returns like that - if I were the full time chef of the investors spare private jet I would be updating my CV and looking for a new gig.
Honestly depends on when they got in. Seed investors? They're probably fine with their preferences. Series B and beyond? That's where it gets messy. What round you thinking?
Character.ai reached out to me for an opportunity, but they've already been carved up.
I think it's great that the rank and file got some of their equity cash-out (based on the other comment), but I imagine it isn't an attractive prospect as a start-up to join at this point.
I just ignored the recruiter. I can't imagine their would be a second liquidity event.
> Unfortunate for any employees who aren't founders or researchers, as they don't get any payout or a nice new job from this exit structure. In fact they lose their whole time invested at the company.
Windsurf’s value didn’t go to $0 overnight. The company will continue and their equity is likely still worth a decent amount wherever the company ends up.
Obviously a disappointing outcome for the people who thought life changing money was right around the corner, but they didn’t lose everything.
Just like with Character I'm assuming the employees get something. Whatever nonsense "licensing" fee Google is paying to not cause an antitrust investigation should be paid out straight to employees
The general character of capital markets is to pay as little as possible. Otherwise you lose out to those who are more ruthless. It is plausible that Windsurf employees really are getting very little value for their work. We need to see details of the deal.
Isn't there not some contractual agreement between the VCs and the founders? (I understand that a non-compete might not apply [in CA], but taking VC money is a little different that simply getting hired).
Were I a Windsurf investor, I'd be pissed right now and calling my lawyer.
the founder is on a vesting schedule set with the vc. walking away forfeits his ownership in the company (not sure of the specifics of this weird deal, but this is true in 99% of situations) which returns his ownership to the VCs either directly or functionally.
the only reason he'd walk away is because he thinks other opportunities are higher EV. if he believes this, a) the investors investment is likely worth virtually 0 anyway and b) if it's not, removing a leader who doesn't want to be there probably increases P(success) for the company and further increases the value of the investment.
founder departure isn't good for the narrative, but it's a symptom of an investment going bad, not often a cause.
I was so surprised (or shocked) to hear that Windsurf was getting acquired for 3 billion dollars, I made an HN post asking about the truth of that news - https://news.ycombinator.com/item?id=43933825. HN's system didn't like my tone I guess and removed it, lol.
But in any case, I just can't see how AI code editors like Windsurf or Cursor, without any proprietary model, can be valued at billions. What's the underlying IP that justifies these valuations?
Similar thing to what we’ve witnessed with crypto coins.
It’s AI season and those with money invest in it, pump it and will exit post IPO.
Difference here is, that besides value that those products “hold”, it’s possible also to provide AI as a service, making Google / Microsoft etc interested.
Cursor (and Garry Tan’s X post) has shown us that the VC money is propping up these companies astounding growth, the only way for them to become profitable is to increase the cost per a request, which means they need to innovate like crazy.
The moat is paper thin.
GitHub has open sourced copilot.
The open source community is working hard on their own projects.
No doubt Cursor is moving fast to create amazing innovations, but if the competition only focuses on thin wrappers they are not worth the billion dollar valuations.
I love watching this space as it is moving extremely fast.
There is no moat. If you’re a true believer that strong agents are around the corner, then all of these add on companies will be obsolete in a few years. The first company to strong agents can trivially rebuild Cursor or Windsurf.
That would be true if the product was the goal. In my experience,
marketing > market > product
Even with AGI in hand, there will still be competition between offerings based on externalities, inertia, or battle-testedness, or authority. Maybe super-intelligence would change the calculus of that, but you'd still probably find opportunities beyond just letting your pool of agents vibe code it.
If you believe AGI is round the corner (I don't), then you face the dilemma of investing in a company that will be the one profitable survivor in a disintegrating world.
This would only be true if it was cheap to run and would return results quickly. If AGI only has compute to serve 1 customer per hour then their is an upper bound of market share it can take from other products.
Cursor just committed mass consumer fraud at worst, and at best pissed off all their best customers. I feel really sorry for those who invested at a 9bb valuation.
flip-flopping on pricing has led users to feel nickel-and-dimed
i like cursor fine, but check out the forum/subreddit to see people talking like addicts, pissed their fix is getting more expensive
i think this aggressive reaction is more pronounced for non-programmers who are making things for the first time. they tasted a new power and they don't want it taken away.
I agree with your take, but I still don't excuse anti consumer practices like that. It annoys me because this is a repeat problem in this space, where these companies don't take into account the market dynamics, or costs of their service. From the start I've been looking at these $20.00 subscriptions, and then my own personal api per token costs and been wondering how they aren't all bankrupt.
Look no further than founders in the sports betting space, like the fanduel founders. Borrow a bunch of money at huge valuations because of hype and ignore the fact, that despite it being exciting and popular, the margins are like <5%. Fanduel founders sold for 400 something million, walked away with nothing. Its now a multibillion dollar company when the new owners realized the product was marketing, not the vig. These AI companies are shifting towards their "marketing" eras.
What surprises me is just how much they've missed the mark.
I'm not an extreme user of Cursor. It has become an essential part of my workflow, but I also probably on the lower/medium section of users. I know that a lot of my friends were spending $XXX amounts/month on extra usage with them, while I've never gone beyond 50% included premium credits usage.
After their changes I'm getting hit with throttling multiple times a day, which likely means that the same thing happens to almost every Cursor user. So that means one or more of:
- They are jacking up the prices, to squeeze out more profit, so it looks good in the VC game
- They had to jack up the prices, so that they aren't running at a loss anymore (that would be a bad indicator regarding profitability for the whole field)
- They are really incompetent about simulating/estimating the impact of their pricing decisions, which also isn't a good future indicator for their customers
My guess is that it's your second scenario there (avoid running at a loss). In the start-up game scale/growth is the most important thing and profits really aren't that important. you want to show to later stage VCs that your idea has traction and there's a large addressable market.
Whilst profits aren't important you also can't burn all your current capital, so if the burn rate gets too high you have to put up prices, which seems likely to be what Cursor is doing.
Until something else comes around for similar price and is much better.
Good thing for consumers who use AI coding tools is that there is no lock-in like in Photoshop or similar software where you hone your skills for years to use particular tool. Switching from Cursor to any other platform would literally take 10 minutes.
I think the recent Grok release and considering xAI was relatively late to the game shows that the only moat to training giant models is how many GPUs you can buy. ChatGPT was earth-shattering and it took less than two years for multiple credible competitors to match or exceed them. Making these models profitable is proving extremely difficult in the face of so much competition and such unsustainable expectations being set. Google seems to be most likely to sustain themselves through this melee. Them and the Chinese companies.
His reaction seems entirely appropriate. He could ignore you, but then you might keep replying to his posts and potentially spread incorrect but damaging information. He is losing close to zero by blocking you, but preventing a potential big loss. Why did you make that remark, if not to damage YC's reputation? This does not seem like the correct approach, if you wanted to improve their selection process.
> Why did you make that remark, if not to damage YC's reputation?
Seems harsh and cultish to assume malice. He didnt say you parents have false credentials
I would say calling out people and institutions like that is important so as to keep them honest, and if they arent honest and are trying to grift/defraud people then they deserve the reputation loss
> He is losing close to zero by blocking you, but preventing a potential big loss.
Thats great for gary, but the rest of the world isnt there waiting to be optimized for his benefit. If people trust YC to incubate good talent, but feel its becoming a hub for grifters, then some accountability is in order. Institutions are beholden to their public stakeholders, even private institutions, because they still have people who are using and supporting them
This whole situation feels shockingly close to the Meta/Scale situation, where founders and specific employees were plucked out, and effectively gutted any future prospects for the company.
At least in the Scale case there seemed to be some form of payout to employees and equity holders, but this takes it a whole lot further by just throwing out all other employees.
There is supposed to be the concept that “all common stock is the same”. These fake-acquisitions completely undermine that.
I never knew anyone who used Windsurf. These AI acquisitions have been unbelievable(in a bad way). WIX acquired some garbage Lovable.dev clone for 80 million. I think many of us are waiting for this bubble to pop(economy will likely pop too)
It was barely better than Cursor and they got shafted by Anthropic because of the takeover announcement so nobody really used it anymore because let's face it - Claude Sonnet is just the best coding model. Design-wise the chat panel and autocomplete integration was a bit nicer than in Cursor but not by much. Subscription for Windsurf was/is also 5$ cheaper.
i don't think it was better than or comparable to cursor at all. except for the month prior to the OpenAI Acquisition news where some minor influencers on X were calling it better.
Everyone has a niche, Windsurf is the only large provider if you are a Jetbrains shop.
There are some alternatives like continue.dev or Jetbrains own AI offering but no Cursor or Claude Code ( Sonnet 3.7/4) you can get through Jetbrains plugin or others, but Anthropic does not provide support same with cursor.
Check out sweep. Completely unaffiliated, their only offering is the jetbrains plugin so it gets a lot more focus than windsurf. Only downside is that Claude code is still a better agent, but at least its tab complete is some of the best
Base44 is absolutely not garbage. I’ve tried it and can say it’s as good or better of a vibe-builder than Lovable or Bolt. Have you benchmarked it against the competition or can you otherwise substantiate the “garbage” claim? FWIW I do know one amazing engineer using Windsurf
Agree in principle, but when evaluated against the competition and likely acquisition targets of Wix, it's certainly not garbage. I've seen it vibe code an entire app that was -- admittedly mostly working -- and deploy it with a prompt of 5 words, in about 2 minutes.
So Google, Meta, and Microsoft will just hollow out the best AI startups of their talent instead of buying them - out of fear of monopoly lawsuits I'm assuming?
Nice plan I guess. Kind of obvious to spot though.
I actually don't know if there's much that can be done unless there's some non-competes in those employees' contracts which are usually not very enforceable outside of finance iirc.
I’m honestly just surprised that the CEO and co-founder decided to walk away from the company and leave behind all these employees he was leading. Especially considering many of them probably joined for lower pay, hoping for a big upside.
Yes, startups are always a bit of a gamble, but this feels like a captain abandoning ship while it’s still full of sailors (many of whom have families depending on them).
This really is a whole new level of getting screwed.
This is why advice is always to treat options for a non-public company as if they're near zero in value.
Because for most people, they will end up being worth exactly zero in value. Less if they went and exercised those options prior to a liquidity event that may never happen.
“Buying the startup” just means handing over megabucks to do-nothing investors. If Google isn’t buying any product or technology, why should investors get a talent fee?
Do nothing investors who enabled the company to reach this point? Employees who chose lower salaries in expectation of shares being worth something? Come on now.
There are many AI startups and we are just in the beginning of learning how to use them. There will be some stupid company like those you’ve listed that figures out a way to use AI that is far better than any other implementation, and Google, Meta, and Microsoft may go the way of Yahoo and AOL, but we’ll see
The “talent” is not very talented, trust me. These are the short term whims of very large, increasingly bloated organizations. A leaner startup that knows what it has will not sell so quickly. At least, the odds will soon be in favor of whoever first decides to take that bet.
If by Documentary, you mean a new Silicon Valley sitcom, yes , all the ingredients are there: The AGI believers, the doomers, the "cure all diseases" people, the board drama, the money grabbers, the VC dance , the poaching, the lawsuits for copyrights ... there s a whole new universe of caricatures
I went from Emacs to VS Code, then to Cursor, next to Claude Code, which is so good that I feel like I am having half a dozen junior devs at my fingertips, 24/7.
Since Claude Code is cli based, I reviewed my cli toolset: Migrated from iTerm2 to Ghostty and Tmux, from Cursor to NeoVim (my God is it good!).
Just had a 14h workday with this tooling. It’s so good that I complete the work of weeks and months within days! Absolutely beast.
At this point I am thinking IDEs do not reflect the changing reality of software development. They are designed for navigating project folders, writing / changing files. But I don’t review files that much anymore. I rather write prompts, watch Claude Code create a plan, implement it, even write meaningful commit messages.
Yes I can navigate the project with neovim, yes I can make commits in git and in lazygit, but my task is best spent in designing, planning, prompting, reviewing and testing.
I work at a company with over 700 employees. And there are tons of use cases where a simple CRUD app is sufficient. Or where glue code needs to be written / changed for legacy systems. Or where an OS system like Camunda is deployed and needs to be configured, workflows developed, etc
The reality of companies out there is much simpler than the challenges of a startup that needs to build systems that are state of the art, scale for millions of users, etc There are companies out there that make millions, in areas you‘ve never heard of, and their core business does not depend on software development best practices.
In our company we have an IT team with the median age of fifty, team members who never have developed software, just maintain systems, delegate hard work to expensive consultants.
Now in that setting someone coming from a startup background is like someone coming from the future. I feel like a wizard who can solve problems in days, instead of weeks or months waiting for a consultant to solve.
Not the OP but this is smth I've vibecoded using cursor: https://bestphoto.ai/ MRR ~$150. It basically started as a clone of my other site: https://aieasypic.com (MRR 2.5k, 5-8k/mo rev) since I was having trouble keeping code context in mind and claude was pretty bad at doing full features with the tech stack I used for that site(Django BE, NextJS FE) making adding new features a pain, so I completely switched to a stack that claude is very good at NextJS fullstack(trpc BE) and now it can basically one-shot a feature request.
Just putting this here because a lot of times AI coding seems to be dismissed as smth that can't do actual work ie generate revenue, while its more like making money as a solo dev is already pretty rare and if you're working in a corp. instead you're not going to just post your company name when asked for examples on what you're using AI for.
AI is often used to pump out sites and apps that scam users, SEO spam, etc. So there is definitely a revenue stream that makes scammers and grifters excited for AI. These tools have increased the scope and reach of their scams, and provide a huge boost to their productivity.
That's partly why I'm curious about OP's work. Nobody who's using these tools while following best software engineering practices would claim that they're making them that much more productive. Reviewing the generated code and fixing issues counteracts whatever time is saved by generating code. But if they're vibe coding and don't even look at the code...
At work we’re encouraged to use AI, so I do. For me the one thing that works well is using it to write one off scripts that do stuff and would be a chore to write.
Usually in 2-3 prompts I can get a python or shell script that reads some file list somewhere, reads some json/csv elsewhere. Combines it in various ways and spits out some output to be ingested by some other pipeline.
I just test this code if it works it’s good.
Never in my life would I put this in a critical system though. When I review these files they are full of tiny errors that would blow up in spectacular manner if the input was slightly off somewhere.
It’s good for what it is. But I’m honestly afraid of production code being vibe coded by these tools.
yes vibecoding is addicting like that. but if you are not reviewing any code and simply vibing then
in my expreience you'll eventually get stuck in "its still not working" loops beause you have no other context or insight to provide it other than that. Then you have either accept what you have or throw the whole thing out and/or actually read the code . kind of rules out last option because code is now just too far gone with too many special cases hardcoded because AI sucks at abstraction or real software engineering.
Well you can't risk Claude quitting overnight. It forgets everything it did the day before and now you have to start over ... must ... finish ... tonight ... within ... context ... window.
Fortunately LLMs are stateless thus not affected by passage of time - your context stays exactly as it was while the tool maintaining it is running.
(Prompt caches are another thing; leaving it for the night and resuming the next day will cost you a little extra on resume, if you're using models via API pay-as-you-go billing.)
Works out for Google and the C-suite. Horrible for the employees. These fake-acquisitions are effectively arbitrage against employees, who get left holding nothing. Should be illegal and regulated.
Not sure how the VCs get their cut. I'm guessing that Google can balance it out by participating in rounds for other startups in that VC's porfolio.
Cursor's Accept / Reject feature for each change it makes in each file is nice whereas I have to use a diff tool to review the changes in Claude Code.
Also, if I go down a prompt alley that's a dead end, Cursor has the Restore Checkpoint feature to get back to the original prompt and try a different path. With Claude Code, you had better have committed the code to git, otherwise you end up with a mess you didn't want.
My company pays for both, but I mostly use Cursor unless I know I am doing a new project or some proof of concept, which Claude Code might have an edge on with a more mature TODO list feature.
I got burned too many times from that Restore Checkpoint thing not working right, maybe it's been fixed by now but seems silly to rely on something thats not a literal tool built for the job (version control), not a good shortcut.
> Google will instead hire Windsurf CEO Varun Mohan, cofounder Douglas Chen, and some of Windsurf’s R&D employees and bring them onto the Google DeepMind team, [...] Google will not have any control over nor a stake in Windsurf, but it will take a non-exclusive license to some of Windsurf’s technology. [...] Google didn’t share how much it was paying to bring on the team. OpenAI was previously reported to be buying Windsurf for $3 billion.
Why not an acquisition?
How did Google get Windsurf and investors to agree to this maneuver that decapitated the leadership and key talent, without a big exit event for everyone?
My read of the article: "Here's x% of what OpenAI offered you, you waive legal challenges while we cherry-pick your people and license the tech in their heads, and you can keep the company, and everyone left behind can promote themselves to fill the vacancies."
I'm not surprised. I started using Windsurf when it came out because I liked its UX better than Cursor's.
However, while Cursor and GH Copilot improved, Windsurf went in the opposite direction. On each update, I started to get more and more issues. The agent often tried to run shell commands, and it hung up, or I found minor UI bugs. One day, I decided to give GH Copilot another chance, and I was surprised by how it evolved, to the point that it worked better than Windsurf for my usage. I don’t know what happened internally at Windsurf, but I notice the degradation as a user. If my case indicates what happened to other users, maybe OpenAI saw declining subscriptions and canceled the deal.
> OpenAI’s deal to buy Windsurf is off, and Google will instead hire Windsurf CEO Varun Mohan, cofounder Douglas Chen, and some of Windsurf’s R&D employees and bring them onto the Google DeepMind team, Google and Windsurf announced Friday.
> Mohan and the Windsurf employees will focus on agentic coding efforts at Google DeepMind and work largely on Gemini. Google will not have any control over nor a stake in Windsurf, but it will take a non-exclusive license to some of Windsurf’s technology.
Sounds to me like they're "hiring" them like one "hires" a consultant?
It's unclear if OpenAI cancelled the deal, or Google poached them? Either way, this season of "OpenAI Drama" is wild. First Meta, now Google. Your turn Amazon / Microsoft.
Apparently OpenAI allowed the deal to expire; likely Google had already been in discussion with Windsurf as I'm sure they knew the deal was likely to die well before today.
I don't know anyone who heard or used Windsurf outside the Bay Area. Even Cursor feels very Bay Area bubbly (although that is the market to go after if you're in ai dev tools).
Cursor does add value but it's just a thin layer on top of VSCode so companies could just build that in-house and don't need to acquire. There's no moat there.
Cursor has custom tab and embedding models. And has a lot of distribution / paying users already.
Arguably they have the strongest product moat, and I wouldn’t be surprised if they beat OpenAI in a vertical coding model from that. Easy for them to have users generate evals and have model product feedback loop here.
The tab completion is fast and the best available right now but is still so garbage that I turn it off 99% of the time because the suggestions are mostly noise.
I have the opposite experience, it’s at least 90% correct. For example, if I start writing the name of a function that I just added in a different file, tab will suggest the function, then jump to the top of the file to import it. If I’m changing the way something is called in 5 places, if I change it in the first place, tab will jump to make the same change in the other places. It’s honestly pretty spot on.
Zed tab is a lot worse in comparison (partly because it’s slow)
I have been using Windsurf for few months. They even have their own AI model SWE-1 model. I really liked using Windsurf. They also have integrations with other IDEs ex: jetbrains, VS code, etc.
This week I have been using Claude Code and Windsurf side by side. I would make change with one, stash it, ask the other for similar change and then would diff it.
Overall Windsurf was pretty on a par with Claude code.
I did not see this coming. Wow. The game of thrones in SV.
I wonder what happened with the OpenAI deal. Anyone have any guesses? My first guess is "Look at Claude Code, we can do this ourselves." But, I am likely thinking too simply.
edit: does this mean that Windsurf and its users will stop being iced-out by Anthropic? Or, is this the end of Windsurf?
I'm a rank and file dev at a non-big tech company and I got a call from a Windsurf sales rep this week who I had connected with on LinkedIn the day before (I never gave them my number). They told me my company was in talks with Windsurf about a licensing deal but that they would give me a 30 day trial of an enterprise account for use on personal projects to let me try it in advance. I guess the idea for them is to build enthusiasm among devs in the company?
Is this a standard sales strategy for products like this? It seems pretty aggressive to me but I'm just an engineer so I wouldn't know.
Very standard yep. Sales folks are sort of trained /indoctrinated into telling white lies like that in order to get in the door. There are loads of examples of using fake momentum to close deals. If its a senior person it’s “My CEO asked me to personally reach out to you” or a fake email from the CEO forwarded by the rep. If one person at the company uses it, it’s “we’re negotiating a company wide license” or “we already have a group license with extra seats” or “one of your teammates sent us a list of priority teammates” yada yada.
The deal for the founders may not have been as good as what Google offered. They may only hold 10% after those rounds, a serious part of the acquisition price could go to liquidation preferences of the VCs and the deal is mostly in OpenAI stock instead of cash. Not that hard to imagine the Google option offering them much more actual cash right now.
Perhaps it as combination of how much founders were diluted and how much they are being offered upfront. We are hearing about $100M signing bonuses.
It is hard to say no when Google/Meta gives you say $100M upfront and hundreds more if not Billion+ in RSUs. After 3 rounds it is not unreasonable to have only 5-10%.
10% of a company worth a few billion burning a lot of cash, that needs to keep raising more rounds i.e more dilution, may have less value than RSUs from multi-trillion dollar publicly traded liquid tech company today.
It is also quite hard to raise $5-10+Billion in cash. There are only handful of startups which have ever done so
Very few funds/investors can afford to do so large rounds. This was SoftBank's thesis for most of last decade, compete by just outfunding competing products in a market.
It's something you should never assume is true until the wire hits your account. I had a deal where I was going to make $15 million called off 36 hours before closing.
For the love of God, can we get a reboot of the Silicon Valley television show? Just on AI. Like when they wrapped it, they wrapped it on AI usage. So, it's got the perfect arc for a reboot that focuses perfectly on AI.
My favorite thing about that series was watching it with friends who weren't from the Bay Area. Often they'd be laughing at the sheer absurdity of a situation, and I'd get to point out that it was barely exaggerated from real life.
That's what my friends not from SF said. "This is insane, this would never happen"
Dude, I saw a lot crazier things happen on a monthly basis. And don't even get me started on the personal lives and partying that the show didn't display.
I don't know if you noticed but cursors language server aspect that runs the coding edits and stuff like that from a server to the workstation is a lot better than windsurf.
Windsurf phone's home on every code edit that you have and takes on 30% load on your servers or on your workstation depending on what you're running.
I would strongly discourage the use of windsurf on your systems.
Case in point their AI model that they just built.
This certainly aligns with my own usage. I'm currently using OpenAI's own Codex 50:1 compared to Windsurf. For me, I'd rather take some time to create a good quality prompt and have it work away for a few minutes and create a material delta. It isn't always perfect, and I often have to make a few tweaks myself, but it is much nicer and waiting around and watching Windsurf bang around on a tiny part of the solution. Windsurf is still nice to use for quick UI iteration however.
Are these "acquihire & license" the new M&A...? I recall hearing that this was a "hack" to avoid DOJ and FTC scrutiny over acquisitions, but I have no clue how such deals are structured. Anyone care to chime in?
Not Windsurf… OpenAI. And OpenAI cares because they’re competing (in part) against Copilot, so if Msoft gets all the benefits of Windsurf then OpenAI would effectively be paying 3B to feed their competitor.
Honestly there's no value that windsurf, cursor and all the other VSCode forks provide that couldn't be provided as an extension and even then - none of them perform as well for agentic coding as Cline / Roo Code (debates about the subscription pricing aside due to people often not realising their model limits, public US only based APIs, pay for useful API limits etc aside).
This isn’t a great look for OpenAI, but acquisitions fall through all the time.
The issue isn’t an acquisition not working out, it’s that the founding/exec team felt it appropriate to arrange their own exits and abandon their team before even communicating that their “successful exit” wasn’t actually happening.
Windsurf's value to OpenAI was for the latter to "see the whole chessboard" of context, which is helpful when you're training models to be good at coding.
But codex (and Claude Code) fulfill this from the CLI, and it's a first-party utility, not an acquisition.
I wonder if this is a result of the previously reported clashes between OpenAI and Microsoft over access to the Windsurf IP (under their investment agreement)
So the result of aggressively scrutinizing big tech acquisitions is acquihires, not a more competitive tech ecosystem with say more IPO’s.
The libertarian spin on this would be government should have never scrutinized acquisitions and the result is just worse for everyone.
The progressive spin would be to now ban acquihires somehow, and then whatever new legal invention will be created next. I can imagine the next step being, creating a consulting company out of your startup and then selling yourself as consultants to big techs. Now you are neither acquired nor technically acqui-hired and the whackamole continues.
At some point, we need to realize the solution is the culture of people involved. If the government could just ask to reduce acquisitions to make the ecosystem more competitive and companies tried following it in spirit to the best of their ability, we might have much better results than whatever we have now. When culture degrades, the govt can’t trust companies, the companies can’t trust the govt, everything just gets worse, regardless of what rules you write and enforce.
The culture of the people involved got us to this point, I’m not sure it’s the solution to the problem.
> The progressive spin would be to now ban acquihires somehow, and then whatever new legal invention will be created next.
Progressive has become a moving target, but the pro-competition view would be to break up the massively concentrated companies that are further consolidating markets. Thats what the Khan FTC was trying to do, but we need a Congress interested in a competitive marketplace, which we haven’t had in a while.
This wasn't a result of regulator scrutiny. The issue was that MS (owner of Copilot) was demanding access to the IP (due to their existing agreement with OpenAI), and OpenAI was resisting. In addition, Claude blocked access to Windsurf, which also damaged them as an acquisition target.
I find this hard to believe considering all the recent acquihires that happened recently like Character AI, Inflection, Covariant AI, Scale AI, context AI and so on. Maybe you’re right about the specifics of this situation, but my prior for this being an acquihire is very high and I would need to see very compelling evidence that that is not the case.
The founders fucked over the employees and the investors and sold out. I guess they don’t care if they are worth $200M each but they fucked every employee that poured their heart out into that company.
This deal always looked strange in the first place. The usage of Windsurf was significantly lower than Cursor and Copilot and somehow it was worth $3B.
Given the release of Claude Code, it was already over for them.
@dang - The title’s wording suggest that OpenAI’s CEO is leaving, not Windsurf. A more accurate title might be: “Windsurf’s deal with OpenAI is off, and its CEO is going to Google”
Is Lina Khan to blame for this new acquihire meta? She was very aggressive in blocking any tech acquisition during her time and ever since we have seen more and more acquihires which I believe these companies are using to prevent themselves from getting sued.
Google is having a hard time acquiring Wiz for 32b, and if it's blocked they owe 3.2b to Wiz. So why risk it when you can just spend the money to hire the talent behind it and spend a few month building out a new product.
Let's review the current state of things:
- Terminal CLI agents are several orders of magnitude less $$$ to develop than forking an entire IDE.
- CC is dead simple to onboard (use whatever IDE you're using now, with a simple extension for some UX improvements).
- Anthropic is free to aggressively undercut their own API margins (and middlemen like Cursor) in exchange for more predictable subscription revenue + training data access.
What does Cursor/Windsurf offer over VS Code + CC?
- Tab completion model (Cursor's remaining moat)
- Some UI niceties like "add selection to chat", and etc.
Personally I think this is a harbinger of where things are going. Cursor was fastest to $900M ARR and IMO will be fastest back down again.
I think only way Cursor and other UX wrappers still win is if on device models or at least open source models catch up in the next 2 years. Then i can see a big push for UX if models are truly a commodity. But as long as claude is much better then yes they hold all the cards. (And don't have a bigger company to have a civil war with like openai)
Do you have a citation for this?
It might be at a loss, but I don’t think it is that extravagant.
also you probably talking about distilled deepseek model
I am on Max and I can work 5 hrs+ a day easily. It does fall back to Sonnet pretty fast, but I don't seem to notice any big differece.
The reason they are talking about building new nuclear power plants in the US isn't just for a few training runs, its for inference. At scale the AI tools are going to be extremely expensive.
Also note China produces twice as much electricity as the United States. Software development and agent demand is going to be competitive across industries. You may think, oh I can just use a few hours of this a day and I got a week of work done (happens to me some days), but you are going to end up needing to match what your competitors are doing - not what you got comfortable with. This is the recurring trap of new technology (no capitalism required.)
There is a danger to independent developers becoming reliant on models. $100-$200 is a customer acquisition cost giveaway. The state of the art models probably will end up costing hourly what a human developer costs. There is also the speed and batching part. How willing is the developer to, for example, get 50% off but maybe wait twice as long for the output. Hopefully the good dev models end up only costing $1000-$2000 a month in a year. At least that will be more accessible.
Somewhere in the future these good models will run on device and just cost the price of your hardware. Will it be the AGI models? We will find out.
I wonder how this comment will age, will look back at it in 5 or 10 years.
I don’t see how that can be true, but if it is…
Either you, or I are definitely use Claude Code incorrectly.
And, since there is no global super-state, the world economy is a market economy, so even if every state were a state-owned planned economy, North Korea style, still there would exist this type of competition between states.
So yeah it basically comes down to your definition of "worker-owned". What fraction of worker ownership is necessary? Do C-level execs count as workers? Can it be "worker-owned" if the "workers" are people working elsewhere?
Beyond the "worker-owned" terminology, why is this distinction supposed to matter exactly? Supposing there was an SV startup that was relatively generous with equity compensation, so over 50% of equity is owned by non-C-level employees. What would you expect to change, if anything, if that threshold was passed?
Estimating productivity gains is a flame war I don’t want to start, but as a signal: if the CC Max plan goes up 10x in price, I’m still keeping my subscription.
I maintain top-tier subscription to every frontier service (~$1k/mo) and throughout the week spend multiple hours with each of Cursor, Amp, Augment, Windsurf, Codex CLI, Gemini CLI, but keep on defaulting to Claude Code.
Are you doing front end backend full stack or model development itself?
Are you destilling models for training your own?
I have never heard someone using so much subscription?
Is this for your full time job or startup?
Why not use qwen or deep seek and host it yourself?
I am impressed with what you are doing.
As to “why”: I’ve been coding for 25 years, and LLMs is the first technology that has a non-linear impact on my output. It’s simultaneously moronic and jaw-dropping. I’m good at what I do (eg, merged fixes into Node) and Claude/o3 regularly finds material edge cases in my code that I was confident in. Then they add a test case (as per our style), write a fix, and update docs/examples within two minutes.
I love coding and the art&craft of software development. I’ve written millions of lines of revenue generating code, and made millions doing it. If someone forced me to stop using LLMs in my production process, I’d quit on the spot.
Why not self host: open source models are a generation behind SOTA. R1 is just not in the same league as the pro commercial models.
Which frameworks & libraries have you found work well in this (agentic) context? I feel much of the js lib. landscape does not do enough to enforce an easily-understood project structure that would "constrain" the architecture and force modularity. (I might have this bias from my many years of work with Rails that is highly opinionated in this regard).
Yup 100% agree. I’d rather try to convince them of the benefits than go back to what feels like an unnecessarily inefficient process of writing all code by hand again.
And I’ve got 25+ years of solid coding experience. Never going back.
It’s so stupid fast to get running that you aren’t out anything if you don’t like it.
There was no way I was going to switch to a different IDE.
My app builds and runs fine on Termux, so my CLAUDE.md says to always run unit tests after making changes. So I punch in a request, close my phone for a bit, then check back later and review the diff. Usually takes one or two follow-up asks to get right, but since it always builds and passes tests, I never get complete garbage back.
There are some tasks that I never give it. Most of that is just intuition. Anything I need to understand deeply or care about the implementation of I do myself. And the app was originally hand-built by me, which I think is important - I would not trust CC to design the entire thing from scratch. It's much easier to review changes when you understand the overall architecture deeply.
i found opus is significantly more capable in coding than sonnet, especcially for the task that is poorly defined, thinking mode can fulfill alot of missing detail and you just need to edit a little before let it code.
"Agentic" workflows burn through tokens like there's no tomorrow, and the new Opus model is so expensive per-token that the Max plan pays itself back in one or two days of moderate usage. When people reports their Claude Code sessions costing $100+ per day, I read that as the API price equivalent - it makes no sense to actually "pay as you go" with Claude right now.
This is arguably the cheapest option available on the market right now in terms of results per dollar, but only if you can afford the subscription itself. There's also time/value component here: on Max x5, it's quite easy to hit the usage limits of Opus (fortunately the limit is per 5 hours or so); Max x20 is only twice the price of Max x5 but gives you 4x more Opus; better model = less time spent fighting with and cleaning up after the AI. It's expensive to be poor, unfortunately.
It is also BYOA or you can buy a subscription from Zed themselves and help them out. I currently use it with my free Copilot+ subscription (GitHub hands it out to pretty much any free/open source dev).
Since they announced that you can use the Pro subscription with Claude Code, I've been using it much more and I've never ever been rate limited.
The basic concept is out there.
Lots of smart people studying hard to catch up to also be poached. No shortage of those I assume.
Good trainingsdata still seems the most important to me.
(and lots of hardware)
Or does the specific training still involves lots of smart decisions all the time?
And those small or big decisions make all the difference?
We’d probably see more companies training their own models if it was cheaper, for sure. Maybe some of them would do very well. But even having a lot of money to throw at this doesn’t guarantee success, e.g. Meta’s Llama 4 was a big disappointment.
That said, it’s not impossible to catch up to close to state-of-the-art, as Deepseek showed.
The basic concept is out there: run very fast.
Lots of people running every day who could be poached. No shortage of those I assume.
Good running shoes still seem the most important to me.
2. Cost to train is also prohibitive. Grok data centre has 200,000 H100 Graphics cards. Impossible for a startup to compete with this.
its funny to me since xAI literally the "youngest" in this space and recently made an Grok4 that surpass all frontier model
it literally not impossible
I assume startup here means the average one, that has a little bit less of funding and connections.
money is "less" important factor, I don't say they don't matters but much less than you would think
It’s the difference between running a marathon (impressive) and winning a marathon (here’s a giant sponsorship check).
Coding startups also try to fine-tune OSS models to their own ends. But this is also very difficult, and usually just done as a cost optimization, not as a way to get better functionality.
Uh, the irony is that this is exactly what Windsurf tried.
But the chat UX is so simple it doesn't take up any extra brain-cycles. It's easier to alt-tab to and from; it feels like slacking a coworker. I can have one or more terminal windows open with agents I'm managing, and still monitor/intervene in my editor as they work. Fits much nicer with my brain, and accelerates my flow instead of disrupting it
There's something starkly different for me about not having to think about exactly what context to feed to the tool, which text to highlight or tabs to open, which predefined agent to select, which IDE button to press
Just formulate my concepts and intent and then express those in words. If I need to be more precise in my words then I will be, but I stay in a concepts + words headspace. That's very important for conserving my own mental context window
Their base is $20/mth. That would equal 3.75M people paying a sub to Cursor.
If literally everyone is on their $200/mth plan, then that would be 375K paid users.
There’s 50M VS Code + VS users (May 2025). [1] 7% of all VS Code users having switched to Cursor does not match my personal circle of developers. 0.7% . . . Maybe? But, that would be if everyone using Cursor were paying $200/month.
Seems impossibly high, especially given the number of other AI subscription options as well.
[1] https://devblogs.microsoft.com/blog/celebrating-50-million-d...
Last disclosed revenue from Cursor was $500mil. https://www.bloomberg.com/news/articles/2025-06-05/anysphere...
- Anthropic doesn't use the inputs for training.
- Cursor doesn't have $900M ARR. That was the raise. Their ARR is ~$500m [1].
- Claude Code already support the niceties, including "add selection to chat", accessing IDE's realtime warnings and errors (built-in tool 'ideDiagnostics'), and using IDE's native diff viewer for reviewing the edits.
[1] https://techcrunch.com/2025/06/05/cursors-anysphere-nabs-9-9...
And so I’d say this isn’t a harbinger of the death of Cursor, instead proof that there’s a future in the market they were just recently winning.
They either need to create their own model and compete on cost, or hope that token costs come down dramatically so as to be too cheap to meter.
A lot of devs are not superstar devs.
They don't want a terminal tool, or anything they have to configure.
A IDE you can just download and 'it just works' has value. And there are companies that will pay.
The bigger issue is the advantage Anthropic, Google and OpenAI have in developing and deploying their own models. It wasn't that long ago that Cursor was reading 50 lines of code at a time to save on token costs. Anthropic just came out and yolo'd the context window because they could afford to, and it blew everything else away.
Cursor could release a cli tomorrow but it wouldn't help them compete when Anthropic and Google can always be multiples cheaper
[0] https://cursor.com/blog/agent-web
I don’t think this is true at all. The reason CC is so good is that they’re very deliberate about what goes in the context. CC often spends ages reading 5 LOC snippets, but afterwards it only has relevant stuff in context.
Prompt: https://gist.github.com/transitive-bullshit/487c9cb52c75a970...
- AI is not good enough yet to abandon the traditional IDE experierence if you're doing anything non-trivial. Hard finding use cases for this right now.
- There's no moat here. There are already a dozen "Claude Code UI" OSS projects with similar basic functionality.
Plus recently launched OpenCode, open source CC is gaining traction fast.
There was always very little moat in the model wrapper.
The main value of CC is the free tool built by people who understand all the internals of their own models.
Auto-regressive nature of these things mean that errors accumulate, and IDEs are well placed to give that observability to the human, than a coding agent. I can course correct more easily in an IDE with clear diffs, coding navigation, than following a terminal timeline.
CC has some integration with VSC it is not all or nothing.
I resisted moving from Roo in VS Code to CC for this reason, and then tried it for a day, and didn't go back.
Claude Code - Agentic/Autonomous coding usecases.
Both have their own place in programming, though there are overlaps.
I am genuinely curious if any Cursor or Windsurf users who have also tried Claude Code could speak to why they prefer the IDE-fork tools? I’ve only ever used Claude Code myself - what am I missing?
While Zed's model is not as good the UI is so much better IMO.
It's quite interesting how little the Cursor power users use tab. Majority of the posts are some insane number of agent edits and close to (or exactly) 0 tabs.
It's interesting when I see videos or reddit posts about cursor and people getting rate limited and being super angry. In my experience tab is the number one feature, and I feel like most people using agent are probably overusing it tasks that would honestly take less time to do myself or using models way smarter than they need to be for the task at hand.
I haven't tried Claude Code VS Code extension. Did anyone replaced Cursor with this setup?
Besides that, the IDE seems poorly designed - some navigation options are confusing and it makes way too many intrusive changes (ex: automatically finishing strings).
I've since gone back to VS Code - with Cline (with OpenRouter and super cheap Qwen Coder models, Windsurf FREE, Claude Code with $20 per month) and I get great mileage from all of them.
I honestly don't know how great that is, because it just reiterates what I was planning anyways, and I can't tell if it's just glazing, or it's just drawing the same general conclusions. Seriously though, it does a decent job, and you can discuss / ruminate over approaches.
I assume you can do all the same things in an editor. I'm just comfortable with a shell is all, and as a hardcore Vi user, I don't really want to use Visual Studio.
Does anyone have a comparison between this and OpenAI Codex? I find OpenAI's thing really good actually (vastly better workflow that Windsurf). Maybe I am missing out however.
Occasionally they lose their connection to the terminal in VSCode, but I’ve got no other integration complaints.
And I really prefer the bring-your-own-key model as opposed to letting the IDE be my middleman.
I can do most of what I want with cline, and I've gone back from large changes to just small changes and been moving much quicker. Large refactors/changes start to deviate from what you actually want to accomplish unless you have written a dissertation, and even then they fail.
I've been using Augment for over a year with IntelliJ, and never understood why my colleagues were all raving about Cursor and Windsurf. I gave Cursor a real try, but it wasn't any better, and the value proposition of having to adopt a dedicated IDE wasn't attractive to me.
A plugin to leverage your existing tools makes a lot more sense than an IDE. Or at least until/if AI agents get so smart that you don't need most of the IDE's functionality, which might change what kinds of tooling are needed when you're in the passenger seat rather than the driver's seat.
Not if you want custom UI. There are a lot of things you can do in extension land (continue, cline, roocode, kilocode, etc. are good examples) but there are some things you can't.
One thing I would have thought would be really cool to try is to integrate it at the LSP level, and use all that good stuff, but apparently people trying (I think there was a company from .il trying) either went closed or didn't release anything note worthy...
So an extension will never be able to compete with Copilot.
What are the UX improvements?
I was using the Pycharm plugin and didn’t notice any actual integration.
I had problems with pycharm’s terminal—not least of which was default 5k line scroll back which while easy to change was worst part of CC for me at first.
I finally jumped to using iterm and then using pycharm separately to do code review, visual git workflows, some run config etc.
But the actual value of Pycharm—-and I’ve been a real booster of that IDE has shrank due to CC and moving out of the built in terminal is a threat to usage of the product for me.
If the plugin offered some big value I might stick with it but I’m not sure what they could even do.
My local ollama + continue + Qwen 2.5 coder gives good tab completion with minimal latency; how much better is Cursor’s tab completion model?
I’m still weary of letting LLM edit my code so my local setup gives me sufficient assistance with tab completion and occasional chat.
They're likely artificially holding it back either because its a loss leader they want to use a very specific way, or because they're planning the next big boom/launch (maybe with a new model to build hype?).
Cursor's @Docs is still unparalleled and no MCP server for documentation fetching even comes close. That is the only reason why I still use Cursor, sometimes I have esoteric packages that must be used in my code and other IDEs will simply hallucinate due to not having such a robust docs feature, if any, which is useless to me, and I believe Claude Code also falls into that bucket.
I strongly disagree. It will put the wrong doc snippets into context 99% of the time. If the docs are slightly long then forget it, it’ll be even worse.
I never use it because of this.
During the evaluation at a previous job, we found that windsurf is waaaay better than anything else. They were expensive (to train on our source code directly) but the solution the offered was outperforming others.
That said, the creator of Claude Code jumped to Cursor so they must see a there there.
> - Tab completion model (Cursor's remaining moat)
What is that? I have Gemini Code Assist installed in VSCode and I'm getting tab completion. (yes, LLM based tab completion)
Which, as an aside I find useful when it works but also often extremely confusing to read. Like say in C++ I type
The editor might show And it's nearly impossible to tell that I didn't enter that `;` so I move on to the next line instead of pressing tab only to find the `;` wasn't really there. That's also probably an easy example. Literally it feels like 1 of 6 lines I type I can't tell what is actually in the file and what is being suggested. Any tips? Maybe I just need to set some special background color for text being suggested.and PS: that tiny example is not an example of a great tab completion. A better one is when I start editing 1 of 10 similar lines, I edit the first one, it sees the pattern and auto does the other 9. Can also do the "type a comment and it fills in the code" thing. Just trying to be clear I'm getting LLM tab completion and not using Cursor
I get all AI or none, so it’s always obvious what’s happening.
Completions are OK, but I did not enjoy the feeling of both us having a hand on the wheel and trying to type at the same time.
gemini cli is very expensive.
https://blog.google/technology/developers/introducing-gemini...
There are IDE integrations where you can run it in a terminal session while perusing the files through your IDE, but it's not powering any autocomplete there AFAIK.
I love CC, but letting it auto-write changes is, at best, a waste of time trying to find the bugs after they start compounding.
Trivial/easy stuff - let it make a PR at the end and review in GitHub. It rarely gets this stuff wrong IME or does anything stupid.
Moderately complex stuff - let it code away, review/test it in my IDE and make any changes myself and tell claude what I've changed (and get it to do a quick review of my code)
Complex stuff - watch it like a hawk as it is thinking and interrupt it constantly asking questions/telling it what to do, then review in my IDE.
agentic tool + anthropic subsidized pricing.
Second part is why it has "exploded"
- > curl -fsSL http://claude.ai/install.sh | bash
- > claude
- > OAuth to your Anthropic account
Done. Now you have a SOTA agentic AI with pretty forgiving usage limits up and running immediately. This is why it's capturing developer mindshare. The simplicity of getting up and going with it is a selling point.
I doubt MS has ever made $900M off of VS Code.
What a harsh time to work for an AI startup as a rank and file employee! I wonder how the founders justify going along with it inside their mind.
[0] Character.ai CEO Noam Shazeer Returns to Google https://news.ycombinator.com/item?id=41141112 - 11 months ago (87 comments)
Edit: Thank you @jonny_eh for the clarification. I can't imagine it feels awesome being a leftover but at least you vested out. "Take the money and leave" is still a bit raw when the founders and researchers are now getting the initial payout + generous Google RSU's.
Hopefully Windsurf employees are treated well here.
Note: I worked at Character until recently.
[1] https://www.theverge.com/2024/8/2/24212348/google-hires-char...
$2.4 billion.
Character.ai reached out to me for an opportunity, but they've already been carved up.
I think it's great that the rank and file got some of their equity cash-out (based on the other comment), but I imagine it isn't an attractive prospect as a start-up to join at this point.
I just ignored the recruiter. I can't imagine their would be a second liquidity event.
Source: I was in GDM when character was acquired.
Otherwise why not merge all of engineering into ElGoog?
Windsurf’s value didn’t go to $0 overnight. The company will continue and their equity is likely still worth a decent amount wherever the company ends up.
Obviously a disappointing outcome for the people who thought life changing money was right around the corner, but they didn’t lose everything.
Edit: the people downvoting this clearly can't read, I made the exact same point as jonny_eh.
High interest rates make VC funding more expensive and now bigtech can swoop in, poach all the necessary staff and deprive investors of an exit.
What is the point any more?
Were I a Windsurf investor, I'd be pissed right now and calling my lawyer.
the only reason he'd walk away is because he thinks other opportunities are higher EV. if he believes this, a) the investors investment is likely worth virtually 0 anyway and b) if it's not, removing a leader who doesn't want to be there probably increases P(success) for the company and further increases the value of the investment.
founder departure isn't good for the narrative, but it's a symptom of an investment going bad, not often a cause.
But in any case, I just can't see how AI code editors like Windsurf or Cursor, without any proprietary model, can be valued at billions. What's the underlying IP that justifies these valuations?
The moat is paper thin.
GitHub has open sourced copilot.
The open source community is working hard on their own projects.
No doubt Cursor is moving fast to create amazing innovations, but if the competition only focuses on thin wrappers they are not worth the billion dollar valuations.
I love watching this space as it is moving extremely fast.
Because they didn't do their jobs properly?
What happened?
i like cursor fine, but check out the forum/subreddit to see people talking like addicts, pissed their fix is getting more expensive
i think this aggressive reaction is more pronounced for non-programmers who are making things for the first time. they tasted a new power and they don't want it taken away.
Look no further than founders in the sports betting space, like the fanduel founders. Borrow a bunch of money at huge valuations because of hype and ignore the fact, that despite it being exciting and popular, the margins are like <5%. Fanduel founders sold for 400 something million, walked away with nothing. Its now a multibillion dollar company when the new owners realized the product was marketing, not the vig. These AI companies are shifting towards their "marketing" eras.
Will users feel that a $200 subscription is worth it or not?
I'm not an extreme user of Cursor. It has become an essential part of my workflow, but I also probably on the lower/medium section of users. I know that a lot of my friends were spending $XXX amounts/month on extra usage with them, while I've never gone beyond 50% included premium credits usage.
After their changes I'm getting hit with throttling multiple times a day, which likely means that the same thing happens to almost every Cursor user. So that means one or more of:
- They are jacking up the prices, to squeeze out more profit, so it looks good in the VC game
- They had to jack up the prices, so that they aren't running at a loss anymore (that would be a bad indicator regarding profitability for the whole field)
- They are really incompetent about simulating/estimating the impact of their pricing decisions, which also isn't a good future indicator for their customers
Whilst profits aren't important you also can't burn all your current capital, so if the burn rate gets too high you have to put up prices, which seems likely to be what Cursor is doing.
The thesis is that once you’re paying $200 a month, you’re beholden and won’t pay and compare it with anything else.
Good thing for consumers who use AI coding tools is that there is no lock-in like in Photoshop or similar software where you hone your skills for years to use particular tool. Switching from Cursor to any other platform would literally take 10 minutes.
Seems harsh and cultish to assume malice. He didnt say you parents have false credentials
I would say calling out people and institutions like that is important so as to keep them honest, and if they arent honest and are trying to grift/defraud people then they deserve the reputation loss
> He is losing close to zero by blocking you, but preventing a potential big loss.
Thats great for gary, but the rest of the world isnt there waiting to be optimized for his benefit. If people trust YC to incubate good talent, but feel its becoming a hub for grifters, then some accountability is in order. Institutions are beholden to their public stakeholders, even private institutions, because they still have people who are using and supporting them
At least in the Scale case there seemed to be some form of payout to employees and equity holders, but this takes it a whole lot further by just throwing out all other employees.
There is supposed to be the concept that “all common stock is the same”. These fake-acquisitions completely undermine that.
I am still a paid subscriber but most of my usage is claude code now becaue Windsurf does not Sonnet 4 included in their plan.
if it was better it would have survived.
There are some alternatives like continue.dev or Jetbrains own AI offering but no Cursor or Claude Code ( Sonnet 3.7/4) you can get through Jetbrains plugin or others, but Anthropic does not provide support same with cursor.
Thanks for the share !
...which no one talks about anymore. Okay I guess you have a point.
Nice plan I guess. Kind of obvious to spot though.
They’re just not enforceable against “rank and file” employees.
Maybe there’s more to the story.
Gentle reminder that more startups die by suicide than homicide, and that an early-stage startup is a total crapshoot.
This really is a whole new level of getting screwed.
Founders get a big pay day and leave within a couple years while 100 employees share a 1% of the company between themselves.
Because for most people, they will end up being worth exactly zero in value. Less if they went and exercised those options prior to a liquidity event that may never happen.
You sweet summer child.
Were you under the impression that venture capital is anything more than rent-seeking?
And AI applied to biomedicine arguably already delivered some acceleration.
Don’t forget all the GPUs. Nvidia always gets its cut.
AI growth has slowed to a crawl, and it's priced it self out vs cost of compute.
NVIDIA feels a lot like SUN.
> amazing documentary
Been there, done that: 2001, Startup Dot Com https://www.youtube.com/watch?v=cP4PGjnZwJE
Since Claude Code is cli based, I reviewed my cli toolset: Migrated from iTerm2 to Ghostty and Tmux, from Cursor to NeoVim (my God is it good!).
Just had a 14h workday with this tooling. It’s so good that I complete the work of weeks and months within days! Absolutely beast.
At this point I am thinking IDEs do not reflect the changing reality of software development. They are designed for navigating project folders, writing / changing files. But I don’t review files that much anymore. I rather write prompts, watch Claude Code create a plan, implement it, even write meaningful commit messages.
Yes I can navigate the project with neovim, yes I can make commits in git and in lazygit, but my task is best spent in designing, planning, prompting, reviewing and testing.
The reality of companies out there is much simpler than the challenges of a startup that needs to build systems that are state of the art, scale for millions of users, etc There are companies out there that make millions, in areas you‘ve never heard of, and their core business does not depend on software development best practices.
In our company we have an IT team with the median age of fifty, team members who never have developed software, just maintain systems, delegate hard work to expensive consultants.
Now in that setting someone coming from a startup background is like someone coming from the future. I feel like a wizard who can solve problems in days, instead of weeks or months waiting for a consultant to solve.
Just putting this here because a lot of times AI coding seems to be dismissed as smth that can't do actual work ie generate revenue, while its more like making money as a solo dev is already pretty rare and if you're working in a corp. instead you're not going to just post your company name when asked for examples on what you're using AI for.
I'd consider that a liability, not an asset, but they were pretty happy with it.
AI is often used to pump out sites and apps that scam users, SEO spam, etc. So there is definitely a revenue stream that makes scammers and grifters excited for AI. These tools have increased the scope and reach of their scams, and provide a huge boost to their productivity.
That's partly why I'm curious about OP's work. Nobody who's using these tools while following best software engineering practices would claim that they're making them that much more productive. Reviewing the generated code and fixing issues counteracts whatever time is saved by generating code. But if they're vibe coding and don't even look at the code...
Say no more.
You don't review the code? Just test it works?
Usually in 2-3 prompts I can get a python or shell script that reads some file list somewhere, reads some json/csv elsewhere. Combines it in various ways and spits out some output to be ingested by some other pipeline.
I just test this code if it works it’s good.
Never in my life would I put this in a critical system though. When I review these files they are full of tiny errors that would blow up in spectacular manner if the input was slightly off somewhere.
It’s good for what it is. But I’m honestly afraid of production code being vibe coded by these tools.
and yet you're pulling 14 hour workdays..
(Prompt caches are another thing; leaving it for the night and resuming the next day will cost you a little extra on resume, if you're using models via API pay-as-you-go billing.)
Not sure how the VCs get their cut. I'm guessing that Google can balance it out by participating in rounds for other startups in that VC's porfolio.
Cursor's Accept / Reject feature for each change it makes in each file is nice whereas I have to use a diff tool to review the changes in Claude Code.
Also, if I go down a prompt alley that's a dead end, Cursor has the Restore Checkpoint feature to get back to the original prompt and try a different path. With Claude Code, you had better have committed the code to git, otherwise you end up with a mess you didn't want.
My company pays for both, but I mostly use Cursor unless I know I am doing a new project or some proof of concept, which Claude Code might have an edge on with a more mature TODO list feature.
> Google hires Windsurf CEO Varun Mohan, others in $2.4 billion AI talent deal
https://www.cnbc.com/2025/07/11/google-windsurf-ceo-varun-mo...
Why not an acquisition?
How did Google get Windsurf and investors to agree to this maneuver that decapitated the leadership and key talent, without a big exit event for everyone?
My read of the article: "Here's x% of what OpenAI offered you, you waive legal challenges while we cherry-pick your people and license the tech in their heads, and you can keep the company, and everyone left behind can promote themselves to fill the vacancies."
If the people instead just quit their jobs and start working at Google … nothing to see here.
However, while Cursor and GH Copilot improved, Windsurf went in the opposite direction. On each update, I started to get more and more issues. The agent often tried to run shell commands, and it hung up, or I found minor UI bugs. One day, I decided to give GH Copilot another chance, and I was surprised by how it evolved, to the point that it worked better than Windsurf for my usage. I don’t know what happened internally at Windsurf, but I notice the degradation as a user. If my case indicates what happened to other users, maybe OpenAI saw declining subscriptions and canceled the deal.
> Mohan and the Windsurf employees will focus on agentic coding efforts at Google DeepMind and work largely on Gemini. Google will not have any control over nor a stake in Windsurf, but it will take a non-exclusive license to some of Windsurf’s technology.
Sounds to me like they're "hiring" them like one "hires" a consultant?
Arguably they have the strongest product moat, and I wouldn’t be surprised if they beat OpenAI in a vertical coding model from that. Easy for them to have users generate evals and have model product feedback loop here.
Zed tab is a lot worse in comparison (partly because it’s slow)
This week I have been using Claude Code and Windsurf side by side. I would make change with one, stash it, ask the other for similar change and then would diff it.
Overall Windsurf was pretty on a par with Claude code.
I wonder what happened with the OpenAI deal. Anyone have any guesses? My first guess is "Look at Claude Code, we can do this ourselves." But, I am likely thinking too simply.
edit: does this mean that Windsurf and its users will stop being iced-out by Anthropic? Or, is this the end of Windsurf?
You must be new around here.
UI is also worse compared to Claude.
They still have some work to do if they want to compete with Claude TBH.
I'm a rank and file dev at a non-big tech company and I got a call from a Windsurf sales rep this week who I had connected with on LinkedIn the day before (I never gave them my number). They told me my company was in talks with Windsurf about a licensing deal but that they would give me a 30 day trial of an enterprise account for use on personal projects to let me try it in advance. I guess the idea for them is to build enthusiasm among devs in the company?
Is this a standard sales strategy for products like this? It seems pretty aggressive to me but I'm just an engineer so I wouldn't know.
They raised A, B, and C round (according to CrunchBase), and then the founders just walk away and get a job/deal at Google?
It is hard to say no when Google/Meta gives you say $100M upfront and hundreds more if not Billion+ in RSUs. After 3 rounds it is not unreasonable to have only 5-10%.
10% of a company worth a few billion burning a lot of cash, that needs to keep raising more rounds i.e more dilution, may have less value than RSUs from multi-trillion dollar publicly traded liquid tech company today.
It is also quite hard to raise $5-10+Billion in cash. There are only handful of startups which have ever done so
Very few funds/investors can afford to do so large rounds. This was SoftBank's thesis for most of last decade, compete by just outfunding competing products in a market.
The same set of rules that apply to you and me are not universal.
Dude, I saw a lot crazier things happen on a monthly basis. And don't even get me started on the personal lives and partying that the show didn't display.
Windsurf phone's home on every code edit that you have and takes on 30% load on your servers or on your workstation depending on what you're running.
I would strongly discourage the use of windsurf on your systems.
Case in point their AI model that they just built.
And it was a crazy deal to begin with, for reference JetBrains who's building IDEs for 24 years are evaluated at $7 billions
The issue isn’t an acquisition not working out, it’s that the founding/exec team felt it appropriate to arrange their own exits and abandon their team before even communicating that their “successful exit” wasn’t actually happening.
Windsurf's value to OpenAI was for the latter to "see the whole chessboard" of context, which is helpful when you're training models to be good at coding.
But codex (and Claude Code) fulfill this from the CLI, and it's a first-party utility, not an acquisition.
Imagine backing this startup and the founder team takes a parachute...
https://news.ycombinator.com/item?id=44538243
The libertarian spin on this would be government should have never scrutinized acquisitions and the result is just worse for everyone.
The progressive spin would be to now ban acquihires somehow, and then whatever new legal invention will be created next. I can imagine the next step being, creating a consulting company out of your startup and then selling yourself as consultants to big techs. Now you are neither acquired nor technically acqui-hired and the whackamole continues.
At some point, we need to realize the solution is the culture of people involved. If the government could just ask to reduce acquisitions to make the ecosystem more competitive and companies tried following it in spirit to the best of their ability, we might have much better results than whatever we have now. When culture degrades, the govt can’t trust companies, the companies can’t trust the govt, everything just gets worse, regardless of what rules you write and enforce.
> The progressive spin would be to now ban acquihires somehow, and then whatever new legal invention will be created next.
Progressive has become a moving target, but the pro-competition view would be to break up the massively concentrated companies that are further consolidating markets. Thats what the Khan FTC was trying to do, but we need a Congress interested in a competitive marketplace, which we haven’t had in a while.
Nothing to do with regulators.
I hope no one works for them again.
Given the release of Claude Code, it was already over for them.
loool dead
I commented on the OG thread something like "weird since MSFT owns VS Code" and got downvoted to oblivion.
Yet here we are, always right :).
Google is having a hard time acquiring Wiz for 32b, and if it's blocked they owe 3.2b to Wiz. So why risk it when you can just spend the money to hire the talent behind it and spend a few month building out a new product.