For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | tiku's commentsregister

Yeah for dev purposes perhaps. Production would be another story.

Im using z.ai when I hit my Claude limit after a few questions..drops in easily in Claude code.

I hate that my M5 with 24 gb has so much trouble with these models. Not getting any good speeds, even with simple models.

A lot of trial and error. I've built graphical tools with GD in PHP, the difficult part for me what that the coordinates where inverted.. I only knew how to draw lines and pixels, but I got the job done.

I remember the LinkedIn app that got all your contacts from your phone and tried to add them to your network. I had random people from internet-deals (local craigslist) that where popping up. So strange that this was allowed.

Thanks, I was already distracted after the first sentence, hoping there would be a good explanation.


I still hate Claude for turning down limits. I use z.ai in Claude code now, haven't hit the limit yet.


Perhaps time to dig out a channel to cut across the land there? To add some second path.


Personally I'm so disappointed about the state of local AI. Only old models run "decent" but decent is way to slow to be usable.


This is exactly the problem we're trying to solve. The models themselves have gotten surprisingly capable at small sizes, Qwen3.5 4B with 262K context, LFM2 1.2B for fast tool calling, but the inference infrastructure hasn't kept up.

When people say "local AI is too slow," they usually mean the engine is too slow, not the model. A 4B model at 186 tok/s (MetalRT on M4 Max) feels genuinely responsive for interactive chat. The same model at 87 tok/s (llama.cpp) feels sluggish. Same weights, same quality, 2x the speed, that's a usability cliff.

We think the gap between cloud and on-device inference is a infrastructure problem, not a model problem. That's what we're working on.


Depends on your measurements. If you measure with 1 cm it is longer than if measure with 10 cm.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You