| | from __future__ import absolute_import |
| |
|
| | import uuid |
| |
|
| | import pytest |
| |
|
| | from test.testutil import env_kafka_version, random_string |
| | from test.fixtures import KafkaFixture, ZookeeperFixture |
| |
|
| | @pytest.fixture(scope="module") |
| | def zookeeper(): |
| | """Return a Zookeeper fixture""" |
| | zk_instance = ZookeeperFixture.instance() |
| | yield zk_instance |
| | zk_instance.close() |
| |
|
| |
|
| | @pytest.fixture(scope="module") |
| | def kafka_broker(kafka_broker_factory): |
| | """Return a Kafka broker fixture""" |
| | return kafka_broker_factory()[0] |
| |
|
| |
|
| | @pytest.fixture(scope="module") |
| | def kafka_broker_factory(zookeeper): |
| | """Return a Kafka broker fixture factory""" |
| | assert env_kafka_version(), 'KAFKA_VERSION must be specified to run integration tests' |
| |
|
| | _brokers = [] |
| | def factory(**broker_params): |
| | params = {} if broker_params is None else broker_params.copy() |
| | params.setdefault('partitions', 4) |
| | num_brokers = params.pop('num_brokers', 1) |
| | brokers = tuple(KafkaFixture.instance(x, zookeeper, **params) |
| | for x in range(num_brokers)) |
| | _brokers.extend(brokers) |
| | return brokers |
| |
|
| | yield factory |
| |
|
| | for broker in _brokers: |
| | broker.close() |
| |
|
| |
|
| | @pytest.fixture |
| | def kafka_client(kafka_broker, request): |
| | """Return a KafkaClient fixture""" |
| | (client,) = kafka_broker.get_clients(cnt=1, client_id='%s_client' % (request.node.name,)) |
| | yield client |
| | client.close() |
| |
|
| |
|
| | @pytest.fixture |
| | def kafka_consumer(kafka_consumer_factory): |
| | """Return a KafkaConsumer fixture""" |
| | return kafka_consumer_factory() |
| |
|
| |
|
| | @pytest.fixture |
| | def kafka_consumer_factory(kafka_broker, topic, request): |
| | """Return a KafkaConsumer factory fixture""" |
| | _consumer = [None] |
| |
|
| | def factory(**kafka_consumer_params): |
| | params = {} if kafka_consumer_params is None else kafka_consumer_params.copy() |
| | params.setdefault('client_id', 'consumer_%s' % (request.node.name,)) |
| | params.setdefault('auto_offset_reset', 'earliest') |
| | _consumer[0] = next(kafka_broker.get_consumers(cnt=1, topics=[topic], **params)) |
| | return _consumer[0] |
| |
|
| | yield factory |
| |
|
| | if _consumer[0]: |
| | _consumer[0].close() |
| |
|
| |
|
| | @pytest.fixture |
| | def kafka_producer(kafka_producer_factory): |
| | """Return a KafkaProducer fixture""" |
| | yield kafka_producer_factory() |
| |
|
| |
|
| | @pytest.fixture |
| | def kafka_producer_factory(kafka_broker, request): |
| | """Return a KafkaProduce factory fixture""" |
| | _producer = [None] |
| |
|
| | def factory(**kafka_producer_params): |
| | params = {} if kafka_producer_params is None else kafka_producer_params.copy() |
| | params.setdefault('client_id', 'producer_%s' % (request.node.name,)) |
| | _producer[0] = next(kafka_broker.get_producers(cnt=1, **params)) |
| | return _producer[0] |
| |
|
| | yield factory |
| |
|
| | if _producer[0]: |
| | _producer[0].close() |
| |
|
| | @pytest.fixture |
| | def kafka_admin_client(kafka_admin_client_factory): |
| | """Return a KafkaAdminClient fixture""" |
| | yield kafka_admin_client_factory() |
| |
|
| | @pytest.fixture |
| | def kafka_admin_client_factory(kafka_broker): |
| | """Return a KafkaAdminClient factory fixture""" |
| | _admin_client = [None] |
| |
|
| | def factory(**kafka_admin_client_params): |
| | params = {} if kafka_admin_client_params is None else kafka_admin_client_params.copy() |
| | _admin_client[0] = next(kafka_broker.get_admin_clients(cnt=1, **params)) |
| | return _admin_client[0] |
| |
|
| | yield factory |
| |
|
| | if _admin_client[0]: |
| | _admin_client[0].close() |
| |
|
| | @pytest.fixture |
| | def topic(kafka_broker, request): |
| | """Return a topic fixture""" |
| | topic_name = '%s_%s' % (request.node.name, random_string(10)) |
| | kafka_broker.create_topics([topic_name]) |
| | return topic_name |
| |
|
| |
|
| | @pytest.fixture |
| | def conn(mocker): |
| | """Return a connection mocker fixture""" |
| | from kafka.conn import ConnectionStates |
| | from kafka.future import Future |
| | from kafka.protocol.metadata import MetadataResponse |
| | conn = mocker.patch('kafka.client_async.BrokerConnection') |
| | conn.return_value = conn |
| | conn.state = ConnectionStates.CONNECTED |
| | conn.send.return_value = Future().success( |
| | MetadataResponse[0]( |
| | [(0, 'foo', 12), (1, 'bar', 34)], |
| | [])) |
| | conn.blacked_out.return_value = False |
| | def _set_conn_state(state): |
| | conn.state = state |
| | return state |
| | conn._set_conn_state = _set_conn_state |
| | conn.connect.side_effect = lambda: conn.state |
| | conn.connect_blocking.return_value = True |
| | conn.connecting = lambda: conn.state in (ConnectionStates.CONNECTING, |
| | ConnectionStates.HANDSHAKE) |
| | conn.connected = lambda: conn.state is ConnectionStates.CONNECTED |
| | conn.disconnected = lambda: conn.state is ConnectionStates.DISCONNECTED |
| | return conn |
| |
|
| |
|
| | @pytest.fixture() |
| | def send_messages(topic, kafka_producer, request): |
| | """A factory that returns a send_messages function with a pre-populated |
| | topic topic / producer.""" |
| |
|
| | def _send_messages(number_range, partition=0, topic=topic, producer=kafka_producer, request=request): |
| | """ |
| | messages is typically `range(0,100)` |
| | partition is an int |
| | """ |
| | messages_and_futures = [] |
| | for i in number_range: |
| | |
| | encoded_msg = '{}-{}-{}'.format(i, request.node.name, uuid.uuid4()).encode('utf-8') |
| | future = kafka_producer.send(topic, value=encoded_msg, partition=partition) |
| | messages_and_futures.append((encoded_msg, future)) |
| | kafka_producer.flush() |
| | for (msg, f) in messages_and_futures: |
| | assert f.succeeded() |
| | return [msg for (msg, f) in messages_and_futures] |
| |
|
| | return _send_messages |
| |
|