Help: Error sending GroupCoordinatorRequest_v0 to node
See original GitHub issueUpgraded to kafka-python 1.2.2 running against a 3-node Kafka 0.10 cluster. We’re receiving this error once on every subscribe action:
ERROR:kafka.coordinator:Error sending GroupCoordinatorRequest_v0 to node XXX [XXX]
The node specified appears to be random.
Issue Analytics
- State:
- Created 7 years ago
- Comments:18 (11 by maintainers)
Top Results From Across the Web
The group coordinator is not available-Kafka - Stack Overflow
I faced a similar issue. The problem was when you start your Kafka broker there is a property associated with it, "KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR".
Read more >Kafka cluster health check fails with the error "Group ...
This article provides a workaround for the Kafka health check error "Group coordinator lookup failed: The coordinator is not available."
Read more >Solving My Weird Kafka Rebalancing Problems & Explaining ...
First, every consumer sends a JoinGroup request to the coordinator while not revoking their partitions. The group leader reassigns all ...
Read more >Kafka 3.3 Documentation
To connect, store, and make available data produced by different divisions of a company. To serve as the foundation for data platforms, event-driven ......
Read more >Coordinator in Apache Kafka - Waiting For Code
But it's not always the case. For heartbeating, coordinator sends requests with specified delay. Regarding to the group management actions, ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Found the root cause of the problem. The reason for the issue was that the session timeout was larger than the request timeout. Whenever we would restart a process, the previous process’ session would live for too long relative to the request timeout, leading to the request timeout error.
When looking at the standard java based kafka client, there is actually an exception being thrown when the session timeout is larger than the request timeout, or when the fetch-max-wait is larger than the request timeout. See code Here.
I’ve created a pull request that throws an error if these constraints are violated #986 . Ran the tests manually through pytest tough, since tox stuff wasn’t working for me. Send me any comments on it if needed.
Thanks Harel
Joining @harelba - It happened to me and was a nightmare to solve until I understood the problem, there is no sense in having session_timeout_ms > request_timeout_ms by definition.