My 50c - ollama cloud 20$. GLM5 and kimi are really competitive models, Ollama usage limits insane high, no limits where to use (has normal APIs), privacy and no logging
Interesting. I've always been turned off by how vague the descriptions of Ollama's limits are for their paid tiers. What sort of work have you been doing with it?
The worst I saw - multiple parallel agents (opencode & pi-coding agents), with Kimi and glm, almost non stop development during the work day - 15-20% session consumption (I think it’s 2h bucket) max. Never hit the limit.
In contrast, 20$ Claude in the similar mode I consumed after just few hours of work.
yeah? why do you like that over using GLM5 in a VPS that charges by token use? $20 still cheaper and seamless to set up? how are the tokens per second?
I have roughly 20-40M token usage per day for GLM only (more if count other models). Using API pricing from OR it means ollama more profitable for me after day (few days if count cache properly).
For several models like Kimi and glm they have b300 and performance really good. At launch I got closer to 90-100 tps. Nowadays it’s around 60 tps stable across most models I used (utility models < 120B almost instant)
In my experience, promotion depends on: (a) is cheaper (including on boarding) to fire and hire or pay more, (b) will it cause domino effect so others will also ask for promotions, (c) will employee find an offer outside (matter of personal traits and market conditions).
Personal performance and achievements are usually secondary.
I tried BIND, Dnsmasq, unbound, adguard.
At the end I picked CoreDNS outside kubernetes. It can do everything (including wildcards and ACL), requires few resources and easy to configure.
Few years ago I felt the same and created trusted-cgi.
However, through the years I learned:
- yes, forks and in general processes are fast
- yes, it saves memory and CPU on low load sites
- yes, it’s simple protocol and can be used even in shell
However,
- splitting functions (mimic serverless) as different binaries/scripts creates mess of cross scripts communication
- deployment is not that simple
- security wise, you need to run manager as root and use unique users for each script or use cgroups (or at least chroot). At that moment the main question is why not use containers asis
Also, compute wise, even huge Go app with hundreds endpoints can fit just few megabytes of RAM - there is no much sense to save so few memory.
At worst - just create single binary and run on demand for different endpoints
This project emerged from a crucial need to store highly sensitive information while maintaining the flexibility to integrate with various automation solutions. I've explored options like Baserow, ApiTable, and many others, but they all raised significant concerns for me.
One, these platforms often come with proprietary data structures, making it challenging to reuse existing database tables or ensuring data rescue in case the application is discontinued.
Two, the focus on generating revenue sometimes compromises data security. It's hard to guarantee that our data won't be used, even in an aggregated or anonymous form, for purposes we might not consent to, especially when dealing with exceptionally sensitive information.
Additionally, several features are locked behind paywalls, and certain licenses may prohibit the use of the open-source edition for commercial purposes.
Lastly, the abundance of telemetry is a constant headache, requiring vigilant monitoring of each release to disable or block intrusive data collection.
WebForms seems to address these concerns, offering a self-hosted, privacy-centric, and GitOps-friendly solution for HTML forms. It empowers users to control their data, integrate it into automation workflows, and maintain privacy in a world where data security is paramount. And it's Free and Open Source under MPL-2.0.
I felt exactly the same as the author. That's why I created https://trusted-cgi.reddec.net/ (currently on the way to v1 and under the heavy refactoring)
reply