I don't know much about the factors that determine why one AI coding harness is better than others. Is it system prompts? Or just personal preference in terms of the UX and there isn't actually a better output between using CC or Pi?
So what makes Pi better than CC? Is it better than OpenCode?
My experience with harnesses is entirely about UX, personally. You could just use an LLM directly and pipe its output directly into your source files, but that would produce terrible results in practice. Harnesses / agents are just better versions of “curl https://llm.com > source.{py/js/cpp/etc}”, imo
Long term I’m bullish on an open source harness “winning” the foot race, in a similar way that Linux “won” over Windows and MacOS (that is, debatably)
You mention the client-server architecture of opencode. Is that a local server or is it calling home to opencode servers?
One of the ideas I like about opencode is the ability to prompt and such from a web browser. So I'm curious if that is the client-server architecture you are talking about, or if it's something else.
For reference, I used replit for some vibe-ish coding for a little bit and really liked that I could easily prompt and view output on my phone when hanging out away from my computer. Or while waiting at the airport for example.
(RIP OG replit by the way. They've pretty much completely pivoted from a REPL to AI, which is pretty hilarious to me given the company name xD)
> One of the ideas I like about opencode is the ability to prompt and such from a web browser. So I'm curious if that is the client-server architecture you are talking about, or if it's something else.
Yes, this is what I meant. And yes it's ok that you like this about opencode :)
> For reference, I used replit for some vibe-ish coding for a little bit and really liked that I could easily prompt and view output on my phone when hanging out away from my computer. Or while waiting at the airport for example.
I use Google Jules and also appreciate being able to nudge it forward when not at the computer. In general I often appreciate when things run on other people's machines. However, if I'm to run a thing on my machines, it better be minimalist!
The harness or the tool is ok but all the defaults as part of the paid pieces of the tool have really bad privacy decisions. So they offer Zen as a pay as you use credit system with access to the models they think work best with the harness. Their own stealth model in it along with a number of the leading new models are always-on sharing data for training purposes. They don’t make this immediately obvious either you have to click through links on their website to see the breakdown.
I am not usually super privacy minded but if you already made it nonobvious this is happening I don’t really trust the underlying tool.
Above is the link. The front page says your privacy is important and says they don’t do training on your data except for the following exceptions which links to this page. Then even their own model is training on your data except there is no opt out. So if you pay for zen and you select one of these models in the tool you have no clue it’s auto training on your data.
Your blog is a treasure trove - thanks for sharing.
Do you still cut your own hair ;) ?
But yes us folks in the creative world can learn a few things from the corporate world when it comes to contracts and payment schedules. Mike Monteiro's talk 'F*ck you, pay me' comes to mind.
The scoring layer sits between ingestion and storage. Incoming items get evaluated on a few axes: source reliability (did the agent observe this directly or was it told?), semantic distance from existing memories, and recency weighting for time-sensitive facts.
Contradiction detection runs as a separate step - we embed the incoming memory, similarity-search against existing ones, and score the pair for logical consistency. If it trips a threshold, it gets stored with a conflict flag and a link to the contradicting memory rather than silently overwriting.
The agent sees both during retrieval and reasons about which to trust in context. Sounds like overhead but it's fast — the scoring is a simple feedforward pass, not another LLM call.
So my understanding - from a friend at WPP who told me the same and from a freakonomics episode - is that advertising was wildly oversold before digital.
When the metrics arrived with digital, they saw that advertising, in some ways, was just not as effective as they’d hoped. In some ways the ROI wasn’t there. Seth Godin agrees. He says that advertising in the digital era could be as simple as just having a good product. I think this is Tesla’s position on it - make the best product and the internet takes care of it.
Legacy companies have kept large ad budgets but those are diminishing. From what I spoke with my friend at WPP, he said their data science team showed that outside of a new product or a product that is not recognised by consumers, the actual outcomes from ads are marginal or incremental. Thats what he told me. If your product is already known to consumers, the ROI is questionable.
Always felt suspicious to me that so much of company dynamics are basically about selling yourself to management...and there's one team in the company who's full-time job is selling? Wonder how that will turn out.
None of my coworkers could figure out why I was laid off, and were shocked because I was important to getting the work done, but management made it clear I hadn't been selling myself to management.
My exit is storytelling. I think that’s the only thing that will remain. I suspect humans will still want to hear stories about and from other humans.
There’s something about AIs that feels wrong for storytelling. I just don’t think people will want AIs to tell them stories. And if they do… Well, I believe in human storytelling.
So what makes Pi better than CC? Is it better than OpenCode?
reply