Hi all, I'm trying to deploy the `signoz-0.89.0` Helm chart to an EKS auto mode cluster. The deploym...
b

Ben Evans

27 days ago
Hi all, I'm trying to deploy the
signoz-0.89.0
Helm chart to an EKS auto mode cluster. The deployment appears to succeed but the
signoz-0
pod keeps logging the error
{"level":"error","timestamp":"2025-08-22T05:34:16.103Z","caller":"opamp/opamp_server.go:117","msg":"Failed to find or create agent","agentID":"0198cff4-aeac-705d-90ef-b8db1ce6e534","error":"cannot create agent without orgId","errorVerbose":"cannot create agent without orgId\ngithub.com/SigNoz/signoz/pkg/query-service/app/opamp/model.(*Agents).FindOrCreateAgent\n\t/home/runner/work/signoz/signoz/pkg/query-service/app/opamp/model/agents.go:91\ngithub.com/SigNoz/signoz/pkg/query-service/app/opamp.(*Server).OnMessage\n\t/home/runner/work/signoz/signoz/pkg/query-service/app/opamp/opamp_server.go:115\ngithub.com/open-telemetry/opamp-go/server.(*server).handleWSConnection\n\t/home/runner/go/pkg/mod/github.com/open-telemetry/opamp-go@v0.19.0/server/serverimpl.go:253\nruntime.goexit\n\t/opt/hostedtoolcache/go/1.23.11/x64/src/runtime/asm_amd64.s:1700","stacktrace":"github.com/SigNoz/signoz/pkg/query-service/app/opamp.(*Server).OnMessage\n\t/home/runner/work/signoz/signoz/pkg/query-service/app/opamp/opamp_server.go:117\ngithub.com/open-telemetry/opamp-go/server.(*server).handleWSConnection\n\t/home/runner/go/pkg/mod/github.com/open-telemetry/opamp-go@v0.19.0/server/serverimpl.go:253"}
...and also on the
signoz-otel-collector
pod there are these errors (I assume related):
{"level":"error","timestamp":"2025-08-22T05:00:16.057Z","caller":"opamp/server_client.go:146","msg":"Server returned an error response: %v","component":"opamp-server-client","":"","stacktrace":"github.com/SigNoz/signoz-otel-collector/opamp.(*serverClient).Start.func3\n\t/home/runner/work/signoz-otel-collector/signoz-otel-collector/opamp/server_client.go:146\ngithub.com/open-telemetry/opamp-go/client/internal.(*receivedProcessor).processErrorResponse\n\t/home/runner/go/pkg/mod/github.com/open-telemetry/opamp-go@v0.19.0/client/internal/receivedprocessor.go:247\ngithub.com/open-telemetry/opamp-go/client/internal.(*receivedProcessor).ProcessReceivedMessage\n\t/home/runner/go/pkg/mod/github.com/open-telemetry/opamp-go@v0.19.0/client/internal/receivedprocessor.go:170\ngithub.com/open-telemetry/opamp-go/client/internal.(*wsReceiver).ReceiverLoop\n\t/home/runner/go/pkg/mod/github.com/open-telemetry/opamp-go@v0.19.0/client/internal/wsreceiver.go:94"}
The helm chart deployed fine on a local k3s instance, I only see this issue on EKS. Any ideas how I can fix or investigate further? Thanks. (edit: corrected the first error message)
Hi Team, I'm new to SigNoz and currently exploring exception-based alerts using ClickHouse queries,...
y

Yash Z

4 months ago
Hi Team, I'm new to SigNoz and currently exploring exception-based alerts using ClickHouse queries, but I'm encountering an issue. I would really appreciate your support. Thank you in advance! Here's the additional environment configuration I have made for email for signoz service in docker-compose,
- SIGNOZ_ALERTMANAGER_SIGNOZ_GLOBAL_SMTP__SMARTHOST=<http://scluster.stablehost.com:587|scluster.stablehost.com:587>
- SIGNOZ_ALERTMANAGER_SIGNOZ_GLOBAL_SMTP__AUTH__USERNAME=name@website.com
- SIGNOZ_ALERTMANAGER_SIGNOZ_GLOBAL_SMTP__AUTH__PASSWORD=password      
- SIGNOZ_ALERTMANAGER_SIGNOZ_GLOBAL_SMTP__FROM=support@website.com
- SIGNOZ_ALERTMANAGER_SIGNOZ_GLOBAL_SMTP__REQUIRE__TLS=false
- SIGNOZ_ALERTMANAGER_SIGNOZ_GLOBAL_SMTP__TLS__INSECURE__SKIP__VERIFY=true
Issue: I'm able to receive the test email from the alert channel configuration, but not from test notifications or actual alerts. When I trigger a test notification, I see a success message and an entry with the status "Firing" in the Triggered Alerts tab. However, I encounter an error in the SigNoz service. Error:
2025-05-12 16:30:49 {"level":"info","timestamp":"2025-05-12T11:00:49.950Z","logger":"<http://go.signoz.io/pkg/http/middleware|go.signoz.io/pkg/http/middleware>","caller":"middleware/logging.go:88","msg":"::RECEIVED-REQUEST::","client.address":"172.18.0.1:35196","user_agent.original":"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/133.0.0.0 Safari/537.36","server.address":"localhost","server.port":"8080","http.request.size":0,"http.route":"/api/v1/rules","http.response.status_code":200,"http.server.request.duration":0.015405533}
2025-05-12 16:31:24 {"level":"info","timestamp":"2025-05-12T11:01:24.860Z","caller":"rules/threshold_rule.go:109","msg":"prepareQueryRange","ts":1747047684854,"evalWindow":300000,"evalDelay":120000}
2025-05-12 16:31:24 {"level":"info","timestamp":"2025-05-12T11:01:24.908Z","caller":"utils/time.go:17","msg":"Elapsed time","func_name":"GetTimeSeriesResultV3","duration":0.047875804,"query":"SELECT\n    toStartOfMinute(min_timestamp) AS minute,\n    count() AS trace_count\nFROM\n(\n    SELECT\n        trace_id,\n        min(timestamp) AS min_timestamp,\n        groupArray(`resource_string_service$$name`) AS services\n    --WHERE timestamp BETWEEN toDateTime(1747047240) AND toDateTime(1747047540)\n    FROM signoz_traces.signoz_index_v3\n    GROUP BY trace_id\n)\nWHERE (has(services, 'service-one') = 1) AND (has(services, 'debezium-connect') = 1) AND (has(services, 'service-two') = 0)\nGROUP BY minute\nORDER BY minute","source":"alerts","client":"query-service","alertID":"0196a520-e782-78bc-b30c-49f40e68395e"}
2025-05-12 16:31:24 {"level":"info","timestamp":"2025-05-12T11:01:24.908Z","caller":"rules/threshold_rule.go:483","msg":"number of alerts found","name":"Missing service alert","count":1}
2025-05-12 16:31:49 {"level":"info","timestamp":"2025-05-12T11:01:49.906Z","logger":"<http://go.signoz.io/pkg/http/middleware|go.signoz.io/pkg/http/middleware>","caller":"middleware/logging.go:88","msg":"::RECEIVED-REQUEST::","client.address":"172.18.0.1:35196","user_agent.original":"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/133.0.0.0 Safari/537.36","server.address":"localhost","server.port":"8080","http.request.size":0,"http.route":"/api/v1/rules","http.response.status_code":200,"http.server.request.duration":0.00284774}
2025-05-12 16:31:54 {"timestamp":"2025-05-12T11:01:54.927349855Z","level":"WARN","code":{"function":"<http://github.com/prometheus/alertmanager/notify.RetryStage.exec|github.com/prometheus/alertmanager/notify.RetryStage.exec>","file":"/home/runner/go/pkg/mod/github.com/prometheus/alertmanager@v0.28.0/notify/notify.go","line":866},"msg":"Notify attempt failed, will retry later","logger":"<http://github.com/SigNoz/signoz/pkg/alertmanager/signozalertmanager|github.com/SigNoz/signoz/pkg/alertmanager/signozalertmanager>","pkg":"<http://go.signoz.io/pkg/alertmanager/alertmanagerserver|go.signoz.io/pkg/alertmanager/alertmanagerserver>","component":"dispatcher","receiver":"Pruthvi NB Email","integration":"email[0]","aggrGroup":"{}/{ruleId=~\"-1|0196a520-e782-78bc-b30c-49f40e68395e\"}:{alertname=\"Missing service alert\"}","attempts":1,"err":"establish connection to server: dial tcp [::1]:25: connect: connection refused"}
Hi Team Using docker compose to setup signoz Otel migrator is not successful Clickhouse server error...
d

Deepsikha kar

about 1 year ago
Hi Team Using docker compose to setup signoz Otel migrator is not successful Clickhouse server error
0. Poco::Net::SocketImpl::error(int, String const&) @ 0x00000000153a1b5f in /usr/bin/clickhouse
1. Poco::Net::SocketImpl::sendBytes(void const*, int, int) @ 0x00000000153a2bbd in /usr/bin/clickhouse
2. Poco::Net::StreamSocketImpl::sendBytes(void const*, int, int) @ 0x00000000153a5296 in /usr/bin/clickhouse
3. Poco::Net::HTTPSession::write(char const*, long) @ 0x00000000153908b3 in /usr/bin/clickhouse
4. Poco::Net::HTTPHeaderIOS::~HTTPHeaderIOS() @ 0x000000001538bbdb in /usr/bin/clickhouse
5. Poco::Net::HTTPHeaderOutputStream::~HTTPHeaderOutputStream() @ 0x000000001538bf1f in /usr/bin/clickhouse
6. DB::HTTPServerResponse::send() @ 0x0000000012942988 in /usr/bin/clickhouse
7. DB::HTTPServerConnection::sendErrorResponse(Poco::Net::HTTPServerSession&, Poco::Net::HTTPResponse::HTTPStatus) @ 0x000000001293ecda in /usr/bin/clickhouse
8. DB::HTTPServerConnection::run() @ 0x000000001293e97b in /usr/bin/clickhouse
9. Poco::Net::TCPServerConnection::start() @ 0x00000000153a5a72 in /usr/bin/clickhouse
10. Poco::Net::TCPServerDispatcher::run() @ 0x00000000153a6871 in /usr/bin/clickhouse
11. Poco::PooledThread::run() @ 0x000000001549f047 in /usr/bin/clickhouse
12. Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001549d67d in /usr/bin/clickhouse
13. ? @ 0x00007f3bb0645609
14. ? @ 0x00007f3bb056a353
 (version 24.1.2.5 (official build))
2024.08.05 07:45:03.078225 [ 47 ] {} <Error> ServerErrorHandler: Poco::Exception. Code: 1000, e.code() = 32, I/O error: Broken pipe, Stack trace (when copying this message, always include the lines below):