ClusterMetadata¶
- class kafka.cluster.ClusterMetadata(**configs)[source]¶
A class to manage kafka cluster metadata.
This class does not perform any IO. It simply updates internal state given API responses (MetadataResponse, GroupCoordinatorResponse).
- Keyword Arguments:
- retry_backoff_ms (int): Milliseconds to backoff when retrying on
errors. Default: 100.
- metadata_max_age_ms (int): The period of time in milliseconds after
which we force a refresh of metadata even if we haven’t seen any partition leadership changes to proactively discover any new brokers or partitions. Default: 300000
- bootstrap_servers: ‘host[:port]’ string (or list of ‘host[:port]’
strings) that the client should contact to bootstrap initial cluster metadata. This does not have to be the full node list. It just needs to have at least one broker that will respond to a Metadata API Request. Default port is 9092. If no servers are specified, will default to localhost:9092.
- add_group_coordinator(group, response)[source]¶
Update with metadata for a group coordinator
- Arguments:
group (str): name of group from GroupCoordinatorRequest response (GroupCoordinatorResponse): broker response
- Returns:
string: coordinator node_id if metadata is updated, None on error
- available_partitions_for_topic(topic)[source]¶
Return set of partitions with known leaders
- Arguments:
topic (str): topic to check for partitions
- Returns:
set: {partition (int), …} None if topic not found.
- broker_metadata(broker_id)[source]¶
Get BrokerMetadata
- Arguments:
broker_id (int): node_id for a broker to check
- Returns:
BrokerMetadata or None if not found
- coordinator_for_group(group)[source]¶
Return node_id of group coordinator.
- Arguments:
group (str): name of consumer group
- Returns:
int: node_id for group coordinator None if the group does not exist.
- partitions_for_broker(broker_id)[source]¶
Return TopicPartitions for which the broker is a leader.
- Arguments:
broker_id (int): node id for a broker
- Returns:
set: {TopicPartition, …} None if the broker either has no partitions or does not exist.
- partitions_for_topic(topic)[source]¶
Return set of all partitions for topic (whether available or not)
- Arguments:
topic (str): topic to check for partitions
- Returns:
set: {partition (int), …}
- request_update()[source]¶
Flags metadata for update, return Future()
Actual update must be handled separately. This method will only change the reported ttl()
- Returns:
kafka.future.Future (value will be the cluster object after update)
- topics(exclude_internal_topics=True)[source]¶
Get set of known topics.
- Arguments:
- exclude_internal_topics (bool): Whether records from internal topics
(such as offsets) should be exposed to the consumer. If set to True the only way to receive records from an internal topic is subscribing to it. Default True
- Returns:
set: {topic (str), …}