Hi, Could someone here look into this? <https://s...
# support
s
n
It’s not a straightforward question to answer, the infra needs to be tweaked depending on how you are querying the data, creating dashboards and alerts etc. As of now, we don’t have any such public benchmarks for that w.r.t infra There is a logs benchmark that we published sometime back but it’s a bit outdated right now https://signoz.io/blog/logs-performance-benchmark/
@Prashant Shahi any suggestion for a good starting point for minimum resources w.r.t to 1TB/day logs ingestion.
s
Thanks @nitya-signoz. All I am looking for is hardware requirements for ingesting 1TB logs a day. The search queries should be simple, like: 1. Find a token in this timeframe 2. Find the number of occurences of a token 3. Find regex matches The log retention is 30+days. I want to understand how Signoz will scale to this, and see whether self hosted is the solution or if we need a cloud license. CC: @Prashant Shahi
p
The can widely vary on your nature of data, peak spikes and other requirements. You can start with allocating more resources to clickhouse and otel-collectors. Rest of the components will not be taking up too much resources. And update the allotted resource as you see. For very high production workload, it is recommend have message queuing tool like Kafka/Redpanda or any other compatiable tools. This way, you can handle random spikes better and have a reliable set up with no data loss even when SigNoz clickhouse/collectors are down.