@@ -26,12 +26,10 @@ def test_buffer_pool():
26
26
@pytest .mark .parametrize ("compression" , [None , 'gzip' , 'snappy' , 'lz4' , 'zstd' ])
27
27
def test_end_to_end (kafka_broker , compression ):
28
28
if compression == 'lz4' :
29
- # LZ4 requires 0.8.2
30
29
if env_kafka_version () < (0 , 8 , 2 ):
31
- return
32
- # python-lz4 crashes on older versions of pypy
30
+ pytest .skip ('LZ4 requires 0.8.2' )
33
31
elif platform .python_implementation () == 'PyPy' :
34
- return
32
+ pytest . skip ( 'python-lz4 crashes on older versions of pypy' )
35
33
36
34
if compression == 'zstd' and env_kafka_version () < (2 , 1 , 0 ):
37
35
return
@@ -87,7 +85,7 @@ def test_kafka_producer_gc_cleanup():
87
85
@pytest .mark .parametrize ("compression" , [None , 'gzip' , 'snappy' , 'lz4' , 'zstd' ])
88
86
def test_kafka_producer_proper_record_metadata (kafka_broker , compression ):
89
87
if compression == 'zstd' and env_kafka_version () < (2 , 1 , 0 ):
90
- return
88
+ pytest . skip ( 'zstd requires 2.1.0 or more' )
91
89
connect_str = ':' .join ([kafka_broker .host , str (kafka_broker .port )])
92
90
producer = KafkaProducer (bootstrap_servers = connect_str ,
93
91
retries = 5 ,
@@ -130,10 +128,8 @@ def test_kafka_producer_proper_record_metadata(kafka_broker, compression):
130
128
if headers :
131
129
assert record .serialized_header_size == 22
132
130
133
- # generated timestamp case is skipped for broker 0.9 and below
134
131
if magic == 0 :
135
- return
136
-
132
+ pytest .skip ('generated timestamp case is skipped for broker 0.9 and below' )
137
133
send_time = time .time () * 1000
138
134
future = producer .send (
139
135
topic ,
0 commit comments