It’s not a straightforward question to answer, the infra needs to be tweaked depending on how you are querying the data, creating dashboards and alerts etc. As of now, we don’t have any such public benchmarks for that w.r.t infra
There is a logs benchmark that we published sometime back but it’s a bit outdated right now https://signoz.io/blog/logs-performance-benchmark/
nitya-signoz
09/09/2024, 4:27 PM
@Prashant Shahi any suggestion for a good starting point for minimum resources w.r.t to 1TB/day logs ingestion.
s
Sahil Arora
09/11/2024, 8:46 AM
Thanks @nitya-signoz. All I am looking for is hardware requirements for ingesting 1TB logs a day. The search queries should be simple, like:
1. Find a token in this timeframe
2. Find the number of occurences of a token
3. Find regex matches
The log retention is 30+days. I want to understand how Signoz will scale to this, and see whether self hosted is the solution or if we need a cloud license.
CC: @Prashant Shahi
p
Prashant Shahi
09/12/2024, 6:53 AM
The can widely vary on your nature of data, peak spikes and other requirements.
You can start with allocating more resources to clickhouse and otel-collectors. Rest of the components will not be taking up too much resources.
And update the allotted resource as you see.
For very high production workload, it is recommend have message queuing tool like Kafka/Redpanda or any other compatiable tools.
This way, you can handle random spikes better and have a reliable set up with no data loss even when SigNoz clickhouse/collectors are down.
SigNoz is an open-source APM. It helps developers monitor their applications & troubleshoot problems, an open-source alternative to DataDog, NewRelic, etc.