We serve a few hundred concurrently online users of our web dashboard which visualises data from our ~100k online IoT device fleet.
With respect to our time series systems:
We store around 1M fields (kv pairs) per second, around 10k rows per second i.e. almost a trillion rows per day and almost a hundred trillion new data points per day.
We compress this down to around 1TB of data per day using Clickhouse and store and make available the last 30 days of data for “online analysis”.
We also store all the data we’ve ever ingested in BigQuery for M/L purposes. I shudder to think of how much this costs.
Our queries are all pretty simple and fast, just scanning and aggregating with filtering, all under 100ms.
This costs around 30k USD per month on AWS, not including networking which is too hard for me to separate from our non time series workloads.
Our infrastructure is all python and microservices, using Kafka, and under 1000 cores for the whole pipeline excluding databases.
Our databases are I believe low tens of the largest graviton instances that were available the last time we upgraded.
This is maintained by a backend team of around 15 people, and a total engineering team including hardware, data and software of around 45 people.