There are too many AI coding tools.
Cursor, GitHub Copilot, Claude Code, Cline, Aider, Continue, Tabnine, Codeium, Bolt, v0, Lovable, Replit Agent, Windsurf, Zed AI, and at least a dozen new ones launched while I was writing this sentence.
Every developer I talk to is trying to figure out which ones are worth the subscription and which ones are hype. The answer is messier than the marketing suggests, because the right tool depends on what you are actually doing.
Here is the honest version of which tools earn their place in my daily workflow, which ones I stopped using, and why.
What I Use Every Day
Claude Code for the Real Work
When I am building a feature that touches multiple files, Claude Code is what I reach for. It runs in the terminal, reads the codebase, understands the context, and makes changes across files in a way that no IDE-based tool I have used does as well.
The reason it earns daily use is because it does not try to do too much. It does not auto-suggest while I type. It does not interrupt my thinking with completions I did not ask for. I tell it what I want, it does the work, I review the diff, and I move on.
For full features, refactors, and anything that requires the model to understand more than one file, Claude Code is where I get the most leverage. The Figma-to-React workflow we use, where a 3-hour build becomes 25 minutes, runs almost entirely through Claude Code with a clear, structured prompt.
Cursor for the In-Editor Work
Cursor is my IDE. Not because the AI features are revolutionary, but because the integration is right. Tab to autocomplete, Cmd+K to edit a selected block, Cmd+L to chat with the codebase open in context.
What I use Cursor for is the small stuff. Renaming a variable across a file. Writing a quick utility function. Asking what a piece of unfamiliar code does. Generating boilerplate I do not want to type.
What I do not use Cursor for is the same thing I use Claude Code for. The two tools overlap in capability but diverge in workflow. Cursor is fast for inline work. Claude Code is better for structural work.
What I Tried and Stopped Using
GitHub Copilot
Copilot was the first tool I used seriously, back when it was the only option. It is fine. The autocomplete is reliable. The chat is decent. But once I started using Cursor and Claude Code together, Copilot did not have a unique role.
The autocomplete in Cursor is at least as good. The chat in Claude Code is better. There is no specific task where Copilot is the right answer for me. I cancelled the subscription a year ago and have not missed it.
This is not a criticism of Copilot. It is the reality of how the market has moved. Copilot was the leader for a long time and the rest of the field caught up.
Bolt and v0 for Production Work
I use v0 sometimes for quick UI mockups. It is good at generating React components from a prompt and giving me something visual to react to.
But I stopped using v0 (and Bolt) for any production work. The output is generic. The component patterns do not match how I actually build. Every component generated by these tools needs significant rework before it can ship.
What they are good for is exploration. If I need to see what a layout might look like before committing to building it, v0 is faster than starting from scratch. The moment I want to actually ship the component, I rebuild it from the design with Claude Code.
Multiple AI Tools in One IDE
I tried running Cursor with Cline as a side panel. I tried adding Continue on top of that. I tried Aider in a second terminal while Claude Code ran in another.
Every time, I ended up using one of them and the others sat there. More tools did not mean more productivity. They meant more decisions about which tool to reach for, more context-switching, and more inconsistency in how the AI was approaching my code.
The lesson was simple. Pick one tool for one job. If two tools do similar things, pick the better one and uninstall the other.
What Actually Matters in an AI Coding Tool
After two years of trying every tool that gets released, here is what I look for now.
Does It Understand My Codebase
The single biggest differentiator. A tool that has read the relevant files and understands my conventions will produce code I can use. A tool that is guessing will produce code that needs to be rewritten.
This is why Claude Code is so valuable. It actually reads files. It checks types. It runs tests. It understands the structure before it writes.
Tools that work on a single file with no awareness of the rest of the codebase are useful for narrow tasks. They are not useful for serious work.
Does It Stay Out of My Way When I Do Not Want It
Auto-suggesting completions every time I pause typing is not helpful. It is interrupting.
The tools I have stuck with let me work in silence and step in when I ask. The tools I dropped tried to predict what I needed and got it wrong often enough that I ended up dismissing more suggestions than I accepted.
Does It Show Me What It Did
A diff. A summary of changes. A list of files touched. The minimum information I need to verify the AI did the right thing.
Tools that just commit changes and move on do not earn my trust. Tools that show me their work do.
Does It Work Where I Already Work
I work in the terminal and in the editor. Tools that require me to leave both (a separate web app, a desktop window, a browser tab) get used less. Friction kills daily use.
This is why Cursor and Claude Code together cover most of my needs. They live where I already live.
My Actual Daily Stack
For anyone wondering what the daily reality looks like, this is it.
Cursor is my IDE. I use the autocomplete and the inline edit features constantly. I rarely use the chat panel.
Claude Code runs in a terminal next to my editor. Anything that crosses files, restructures code, or requires understanding context goes through it.
Figma stays open when I am doing UI work. The Figma-to-React workflow is a structured Claude Code prompt that takes the Figma export and produces production React with our component library.
Browser DevTools and Vercel logs are the tools I reach for when something is not working. AI tools are bad at debugging actual production issues. They are good at writing code, not at investigating systems.
That is it. Two AI tools and a few utilities. The minimum stack that does the most work.
The Honest Reality
The AI tooling space moves fast. The list above will probably change in six months. New tools will earn a place. Existing tools will get replaced.
The thing that will not change is the principle. The right AI tool is the one that produces work you can ship without rewriting it. Anything else is a demo, not a tool.
If you are paying for three AI tools and only using one of them, cancel the other two. If a new tool launches and the marketing makes big claims, wait until someone you trust uses it on real work.
The goal is not to use the most AI tools. The goal is to ship better code faster. The tools are means, not ends.
The studios and agencies that are getting real leverage from AI are the ones who picked one or two tools, learned them deeply, and built workflows around them. Not the ones who keep chasing the next thing.
That is the boring answer. It is also the right one.