For the best experience on desktop, install the Chrome extension to track your reading on news.ycombinator.com
Hacker Newsnew | past | comments | ask | show | jobs | submit | history | more enether's commentsregister

What an awful vulnerability. The most interesting fact is that this has been there since the PR that introduced it in 2017[1].

I'm not sure how Mongo's review process works, but it seems like this one had zero review.

[1] - https://github.com/mongodb/mongo/pull/1152


Warpstream (by Confluent (an IBM company))


For Rust-based Kafka alternatives, I like Tansu[1]. It at least provides Kafka API parity, and critically also gives users a pluggable backend (embedded SQLite, S3 for low cost diskless type workloads and Postgres because just use Postgres)

It’s nice to try and out innovate Kafka, but I fear the network effect can’t be beaten unless the alternative is 10x better.

Something like Warpstream’s architecture[2] had a shot at dethroning Kafka, but critically even they adopted the Kafka API. Sure enough, Apache Kafka introduced a competing feature[3] within two years of warpstreams launch too.

[1] - https://github.com/tansu-io/tansu [2] - https://www.warpstream.com/ [3] - https://topicpartition.io/blog/kip-1150-diskless-topics-in-a...


That is hilarious. The product looks like garbage to me. The business opportunity looks like it can be incredibly lucrative if it does catches on.


Aren't nanoplastics a larger concern?

All microplastics out in the open will degrade to nanoplastics at some point, and those find it much easier to infiltrate the human body. They penetrate the blood-brain barrier.

This leads me to believe that it's literally impossible to avoid. The air/water supply must be getting more poluted by the minute with these things


I wonder if plastics, being non-polar, might accumulate in organs which are a bit on the "oily" side - the liver, adipose tissue, and the brain.


how is node failure handled? is this using KRaft or ZK?


These are real trade offs!

Client impl. can be re-created once in a library/extension and be done with.

The network effect of Kafka is the hardest to topple. The API is standard. If you need that connectivity, then use Kafka.

An alternative idea I did not mention is to use a Kafka API proxy on top of Postgres. Tansu is developing this (amongst other pluggable backends) - https://github.com/tansu-io/tansu; That solution would retain both the client implementations and tools (network effect). But it comes at extra complexity


Exactly. "a properly-configured Kafka cluster" implies you have very properly configured your clients too, which is almost never the case because it's practically very hard to do in the messy reality of a large-scale organization.

Even if you somehow get everyone to follow best-practices, you most likely still won't get to saturate the network on "minimal hardware". The number of client connections and requests per second will likely saturate your "minimal CPU".

It's true that minimal hardware on Kafka can saturate the network, but this mostly happens in low-digit client scenarios. In practice, orgs pushing serious data have serious client counts.


Can you give some examples? I'm super curious about single-node Kafka use cases in general


I believe most setups using DB+{Queue,Kafka} don't truly deal with it fwiw.


Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search:

HN For You