Skip to content
Back to AI & LLM
AI & LLM5 min read

AI Coding Tools Ranked: What I Actually Use

aicoding-toolsclaude-codecursorcopilot
Share

I've used every major AI coding tool in production. Not for a weekend trial — for months of real product development. Building AssessAI (200+ files, 630+ tests), building voice AI at Avoca, and maintaining several side projects. Here's my honest ranking.

1. Claude Code (daily driver)

This is what I use 80% of the time. Not because it's the most polished — it runs in a terminal with no GUI. Because it's the most capable.

What makes it different:

  • Full codebase awareness. It reads files, searches across the project, understands the architecture before making changes.
  • Agentic execution. "Add authentication to the API routes" isn't a prompt — it's a task. Claude Code plans, implements, writes tests, and creates the commit.
  • Extensible with agents, skills, and MCP connections. I have 15 agents that specialize in different tasks. A code reviewer, a TDD enforcer, a security scanner — all running in the same interface.

The rough edges:

  • Terminal-only. No inline diffs, no visual code review, no click-to-accept.
  • Token-hungry. Complex tasks can burn through context fast. I autocompact at 80%.
  • Steep learning curve. Getting the most out of it requires configuring agents, commands, and skills. Out of the box, it's good. Configured, it's a multiplier.

Best for: Sustained product development. Building features end-to-end. Refactoring across multiple files. Anything that needs architectural understanding.

2. Cursor

The best GUI-based AI coding experience. Period.

What it does well:

  • Inline diffs. You see exactly what the AI wants to change before accepting. This feedback loop is fast and natural.
  • Cmd+K for targeted edits. Select code, describe the change, review the diff. Simple and effective.
  • Composer for multi-file changes. Not as capable as Claude Code's agentic mode, but the visual presentation is better.
  • Tab completion that actually understands context. Better than Copilot's completions in my experience.

Where it falls short:

  • The AI model swapping is confusing. Claude, GPT-4o, their own fine-tuned model — it's not always clear which you're using.
  • Composer struggles with large refactors. It works well for 3-5 file changes but gets confused past 10 files.
  • Expensive. $20/month plus API costs if you use your own keys for premium models.

Best for: Day-to-day editing. Code review. Quick feature additions. When you want visual diffs and don't need full agentic execution.

3. GitHub Copilot

The safe choice. Everyone has it. It's fine.

What it does well:

  • Tab completion is solid and fast. The latency between typing and suggestion is lower than competitors.
  • Copilot Chat in VS Code works well for quick questions. "What does this function do?" "Write a test for this."
  • Copilot Workspace is interesting for issue-to-PR workflows. It's getting better.
  • Lowest friction to start. Already integrated into most dev environments.

Where it falls short:

  • Completions lack architectural awareness. It suggests the next line, not the right line for this codebase.
  • Chat doesn't have deep codebase understanding. It sees the current file and maybe a few related ones. Ask it about a pattern used across 15 files and it guesses.
  • No agentic capabilities worth mentioning. It's reactive, not proactive.

Best for: Teams standardizing on one tool. Junior developers who benefit from constant suggestions. Codebases where the patterns are well-established.

4. Windsurf (Codeium)

Promising but not there yet.

What it does well:

  • Cascade (their agentic mode) shows the most improvement over the past 6 months. The multi-file editing is getting competitive.
  • Free tier is generous. Useful for evaluating without commitment.
  • Good at following existing code patterns in the current file.

Where it falls short:

  • Cascade hallucinates file paths and imports more than competitors.
  • The model quality is inconsistent. Some sessions feel like Claude-level output, others feel like GPT-3.5.
  • Less community, fewer resources, smaller ecosystem. When you hit a problem, you're more on your own.

Best for: Budget-conscious developers. Teams that want agentic features without Claude Code's learning curve.

the honest comparison table

| Feature | Claude Code | Cursor | Copilot | Windsurf | |---------|------------|--------|---------|----------| | Code understanding | Best | Good | OK | OK | | Multi-file refactoring | Best | Good | Weak | Improving | | Inline completions | N/A (terminal) | Best | Good | Good | | Visual diffs | No | Best | OK | Good | | Agentic execution | Best | Improving | Weak | Improving | | Ease of setup | Hard | Easy | Easiest | Easy | | Cost | $100-200/mo | $20/mo + API | $10-19/mo | Free-$15/mo |

what I actually run day-to-day

My setup:

  • Claude Code in the terminal for all feature work, refactoring, and architecture
  • Cursor open alongside for visual code review and quick inline edits
  • Copilot active in VS Code for tab completions when I'm in a different editor

Could I use just one? Yes — Claude Code. But combining them fills different parts of the workflow. Claude Code thinks. Cursor shows. Copilot suggests.

The tool matters less than how you use it. An engineer who knows their codebase and writes clear instructions will get great output from any of these. The tool just changes the ceiling and the speed.


Share

More in AI & LLM