Cline review
This post is initially generated by a language model, usually by summarizing a human conversation or expanding a human-written summary. The goal is not to populate the internet with yet another piece of uncalled for, AI-generated slop (in which, unfortunately, people working in AI are complicit). Rather, it is to enable lower friction in sharing and distilling information. I have worked on, and often significantly rewritten, the post to ensure it accurately reflects the underlying human intentions and experiences, but there may be inaccuracies and biases that remain.
Update June 29 2025
The situation changed a lot from the time I initially tried out Cline (or generally speaking, the more agentic AI coding tools). The first thing, of course, is that success rate massively improved with new models/tools. There are much more complex tasks I feel comfortable to rely on LMs now, and the easy “gotchas” happen less.
At the same time, I don’t think the big picture has changed: if you have an important project, you probably don’t want a codebase you don’t fully understand, so you can only scrap it when things stop working/LMs stop being able to work on it. Keep the important things in your control!
But I can also confidently say that Claude Code/Codex/Gemini CLIs offer a easier-to-setup and more customizable experience. And if you’re starting out, agent mode in VSCode’s Copilot and Cursor is more user friendly, and also has the benefit of not having to pay-per-use (Cline bill shoots up extremely quickly due to how much context it collects!). Since all of them basically use the same suite of models as backend, ultimately the choice is a choice of whose UX is better, which changes rather quickly, and whose prompts best suit your use cases. I’m guessing that there’s an intricate balance between telling the model too little so it makes the same obvious mistaks and telling it too much so its operating structure is too rigid, but that’s just my guess from the outsider’s perspective.
TL;DR
Cline is a VSCode extension offering a bring-your-own-key alternative to GitHub Copilot, with the ability to execute commands and plan multi-step edits. While promising in concept, its high token usage and several UX limitations make it difficult to recommend over Copilot Edit (as of Feb 2025). It’s also not as “smart” as you think it might be1. Cost per coding session can range from $0.5-3 depending on your model choice.
Also see the verdict section.
Overview
Cline operates as a chatbot-style interface within VS Code, capable of code generation, modification, and terminal command execution (which sounds more promising than it actually is). The default plan is completely free, with the only cost being your LM API calls.
Core Functionality
Unlike Copilot’s real-time completions, Cline works in a turn-based manner similar to Copilot Edit, where you request specific changes or additions and the AI responds with complete code snippets or modifications. The two most important features are the internal feedback loop and more generous access: Cline can execute code changes in steps following its own plan, and it can modify files/execute commands on your computer.
Pros
- Bring-your-own-key! Use any LM and provider you want
- Cline has a “plan” mode, in which it gathers information and makes a plan
- Can request access to files/execute terminal commands
- Offers checkpoint features for reverting changes
Cons
- Cline determines when a task is complete, not you. Once it declares the task is complete it’s done. I find this really weird.
- Very token-consuming: first request is often 10k+ tokens, hitting context limit is realistic. Each session can be $0.5-3 depending on your model, so expect to spend more than copilot/cursor if you let it run by itself.
- No effective code verification: Cline can, in principle, run commands and check outputs, but it doesn’t do it reliably and use command outputs productively.
- An example: I start a task telling Cline how to verify success (run the script with tests in it). Cline executes the command, and without checking the outputs, immediately declares the task is complete.
- In general it feels much like vanilla AI autocomplete: once Cline generates a plan, it executes it step-by-step, without verifying after steps or re-planning. Think about if your initial plan for a coding project every worked out completely!2
- Cannot revert to checkpoints before AI modifications (as of Feb 2025). This could be a really simple fix, but they don’t yet have it. You’d better have another copy/commit before Cline starts working on your code.
- Each session has its own context, so Cline always starts by gathering information. This can be frustrating if your codebase is complicated.
- Doesn’t feel as polished compared to Copilot Edit
Verdict
While Cline offers flexibility through custom API keys, it doesn’t eliminate what I think is the biggest bottleneck in coding — your thinking speed. It’s not reliable enough for you to only care about the high-level functions/designs3, so you still have to be in the loop, understand every line of code, and tell it specifically what to do. If you treat it as a human capable of executing on your high-level goals, you will be thoroughly disappointed. But if you treat it like a multi-turn Copilot Edit, it’s not too bad and can definitely be a productivity tool.
Footnotes
Especially given the model we use is ranked 18th in the world in programming. Yes, I tried O1. Yes, I tried DeepSeek R1. As of Feb 2025, you can’t code hands-off yet.↩︎
Spoiler alert: these executions don’t often work.↩︎
My general impression of what works/what doesn’t work in AI coding: describing only the high level input-output behavior equals diaster. Giving pseudocode or a complete description of the implementation works (and saves you a lot of time), but usually you need to debug it yourself.↩︎