Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Any reason you can't shove it into Kafka?


Too many network calls. Technically it's feasible, operationally it's expensive for Telemetry usecase. Ex: Imagine we are capturing API telemetry. If there are 1000 API calls per minute per node, then we will end up somewhere 1000*10 calls per minute to Kafka. It's not efficient.


I can assure you from deep experience working on telemetry products that Kafka will handle this load like a champ.

Batching sends under the covers to reduce network round trips is all baked in.

This is also one of the things that most existing telemetry clients handle for you ie batching telemetry in memory and shipping it out on an interval, so there's a great deal of existing work you can draw from if not outright copy.


I didn't catch the part where "Parent wants storage at the source node.".

So if the goal is to eventually have the timeseries data merged back to a timeseries DB, and latency isn't too much of a concern then wouldn't batch writing to Kafka (Kinesis, etc) be tolerable?


Parent wants storage at the source node.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: