For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | ali_chherawalla's commentsregister

fair point, let me know work on that. That's a small lift.


lol thanks buddy!


ofcourse ofcourse,

I've documented everything here: https://github.com/alichherawalla/off-grid-mobile-ai/blob/ma...

llama.cpp compiled as a native Android library via the NDK, linked into React Native through a custom JSI bridge. GGUF models loaded straight into memory. On Snapdragon devices we use QNN (Qualcomm Neural Network) for hardware acceleration. OpenCL GPU fallback on everything else. CPU-only as a last resort.

Image gen is Stable Diffusion running on the NPU where available. Vision uses SmolVLM and Qwen3-VL. Voice is on-device Whisper.

The model browser filters by your device's RAM so you never download something your phone can't run. The whole thing is MIT licensed - happy to answer anything about the architecture.


Any roadmap to add Mediatek NPU support?


I'm working on that we speak. Shouldn't not be that difficult of a lift and should be able to do that tonight or in the next couple of nights


Happy to test, I have poco X6 pro, 12gb ram model


awesome. I'll let you know once thats in


you're doing the marketing for me ahaha! appreciate it though. thank you


dude thats awesome to hear! I literally added support for web search, tool calling and KV cache optimization support + usage of all 99 GPU layers a few hours ago!

Those changes are not live on the play store / app store but its available on GH. I'll make a release later today.

Let me know what you think!




Off Grid is live on the playstore now. Check it out: https://play.google.com/store/apps/details?id=ai.offgridmobi...


it doesn't need internet to generate an image. it needs it to download the model, and to be fair if you've got the zip for the model already you can just import that


I have download the model... only then I have disabled network access. It loops with an error message needing access to DSP ?

Anyway, I like what you did, as I do not want to train AI with my inputs this is perfect.


I think thats a bug, I'm guessing you're trying one of the NPU models? If you drop to CPU for now it should hold.

Only phones with qualcomm chips are able to use the NPU. I'm working towards changing that.

but yeah just to be clear there is no internet needed to run any of this. Infact I'm so averse to it, I've not even added analytics for this one. So flying pretty blind here.


sure, I'll take a look at it. Thank you for the recommendation


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You