From 27db4e85667bc8a12b6b00d8afa86b9629cb7cf3 Mon Sep 17 00:00:00 2001 From: Kevin Tindall Date: Wed, 31 Jan 2018 13:29:12 -0600 Subject: [PATCH] use absolute imports everywhere --- kafka/__init__.py | 2 +- kafka/client_async.py | 24 +++++++++---------- kafka/cluster.py | 6 ++--- kafka/conn.py | 6 ++--- kafka/consumer/__init__.py | 6 ++--- kafka/consumer/multiprocess.py | 6 ++--- kafka/consumer/simple.py | 4 ++-- kafka/coordinator/assignors/range.py | 4 ++-- kafka/coordinator/assignors/roundrobin.py | 6 ++--- kafka/coordinator/base.py | 14 ++++++------ kafka/coordinator/consumer.py | 22 +++++++++--------- kafka/metrics/__init__.py | 16 ++++++------- kafka/metrics/stats/__init__.py | 20 ++++++++-------- kafka/partitioner/__init__.py | 6 ++--- kafka/partitioner/default.py | 2 +- kafka/partitioner/hashed.py | 2 +- kafka/partitioner/roundrobin.py | 2 +- kafka/producer/__init__.py | 6 ++--- kafka/producer/buffer.py | 2 +- kafka/producer/future.py | 4 ++-- kafka/producer/kafka.py | 28 +++++++++++------------ kafka/producer/keyed.py | 4 ++-- kafka/producer/record_accumulator.py | 8 +++---- kafka/producer/sender.py | 12 +++++----- kafka/producer/simple.py | 2 +- kafka/protocol/__init__.py | 2 +- kafka/protocol/admin.py | 4 ++-- kafka/protocol/api.py | 4 ++-- kafka/protocol/commit.py | 4 ++-- kafka/protocol/fetch.py | 4 ++-- kafka/protocol/group.py | 6 ++--- kafka/protocol/message.py | 10 ++++---- kafka/protocol/metadata.py | 4 ++-- kafka/protocol/offset.py | 4 ++-- kafka/protocol/produce.py | 4 ++-- kafka/protocol/struct.py | 6 ++--- kafka/protocol/types.py | 2 +- kafka/record/__init__.py | 2 +- kafka/record/default_records.py | 4 ++-- kafka/record/legacy_records.py | 4 ++-- kafka/record/memory_records.py | 6 ++--- kafka/record/util.py | 2 +- kafka/serializer/__init__.py | 2 +- 43 files changed, 144 insertions(+), 144 deletions(-) diff --git a/kafka/__init__.py b/kafka/__init__.py index 6a8041825..f108eff1c 100644 --- a/kafka/__init__.py +++ b/kafka/__init__.py @@ -1,7 +1,7 @@ from __future__ import absolute_import __title__ = 'kafka' -from .version import __version__ +from kafka.version import __version__ __author__ = 'Dana Powers' __license__ = 'Apache License 2.0' __copyright__ = 'Copyright 2016 Dana Powers, David Arthur, and Contributors' diff --git a/kafka/client_async.py b/kafka/client_async.py index 2e78e3b3a..4962d9f1d 100644 --- a/kafka/client_async.py +++ b/kafka/client_async.py @@ -13,26 +13,26 @@ import selectors # pylint: disable=import-error except ImportError: # vendored backport module - from .vendor import selectors34 as selectors + from kafka.vendor import selectors34 as selectors import socket import time from kafka.vendor import six -from .cluster import ClusterMetadata -from .conn import BrokerConnection, ConnectionStates, collect_hosts, get_ip_port_afi -from . import errors as Errors -from .future import Future -from .metrics import AnonMeasurable -from .metrics.stats import Avg, Count, Rate -from .metrics.stats.rate import TimeUnit -from .protocol.metadata import MetadataRequest -from .util import Dict, WeakMethod +from kafka.cluster import ClusterMetadata +from kafka.conn import BrokerConnection, ConnectionStates, collect_hosts, get_ip_port_afi +from kafka import errors as Errors +from kafka.future import Future +from kafka.metrics import AnonMeasurable +from kafka.metrics.stats import Avg, Count, Rate +from kafka.metrics.stats.rate import TimeUnit +from kafka.protocol.metadata import MetadataRequest +from kafka.util import Dict, WeakMethod # Although this looks unused, it actually monkey-patches socket.socketpair() # and should be left in as long as we're using socket.socketpair() in this file -from .vendor import socketpair -from .version import __version__ +from kafka.vendor import socketpair +from kafka.version import __version__ if six.PY2: ConnectionError = None diff --git a/kafka/cluster.py b/kafka/cluster.py index d646fdfee..416852eed 100644 --- a/kafka/cluster.py +++ b/kafka/cluster.py @@ -8,9 +8,9 @@ from kafka.vendor import six -from . import errors as Errors -from .future import Future -from .structs import BrokerMetadata, PartitionMetadata, TopicPartition +from kafka import errors as Errors +from kafka.future import Future +from kafka.structs import BrokerMetadata, PartitionMetadata, TopicPartition log = logging.getLogger(__name__) diff --git a/kafka/conn.py b/kafka/conn.py index 5ff27d5ce..7d4b12b4c 100644 --- a/kafka/conn.py +++ b/kafka/conn.py @@ -12,7 +12,7 @@ import selectors # pylint: disable=import-error except ImportError: # vendored backport module - from .vendor import selectors34 as selectors + from kafka.vendor import selectors34 as selectors import socket import struct @@ -857,8 +857,8 @@ def check_version(self, timeout=2, strict=False): # vanilla MetadataRequest. If the server did not recognize the first # request, both will be failed with a ConnectionError that wraps # socket.error (32, 54, or 104) - from .protocol.admin import ApiVersionRequest, ListGroupsRequest - from .protocol.commit import OffsetFetchRequest, GroupCoordinatorRequest + from kafka.protocol.admin import ApiVersionRequest, ListGroupsRequest + from kafka.protocol.commit import OffsetFetchRequest, GroupCoordinatorRequest # Socket errors are logged as exceptions and can alarm users. Mute them from logging import Filter diff --git a/kafka/consumer/__init__.py b/kafka/consumer/__init__.py index 36c8ff094..4b900ac8c 100644 --- a/kafka/consumer/__init__.py +++ b/kafka/consumer/__init__.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .simple import SimpleConsumer -from .multiprocess import MultiProcessConsumer -from .group import KafkaConsumer +from kafka.consumer.simple import SimpleConsumer +from kafka.consumer.multiprocess import MultiProcessConsumer +from kafka.consumer.group import KafkaConsumer __all__ = [ 'SimpleConsumer', 'MultiProcessConsumer', 'KafkaConsumer' diff --git a/kafka/consumer/multiprocess.py b/kafka/consumer/multiprocess.py index 58e3e07bd..1da4a3353 100644 --- a/kafka/consumer/multiprocess.py +++ b/kafka/consumer/multiprocess.py @@ -8,15 +8,15 @@ from kafka.vendor.six.moves import queue # pylint: disable=import-error -from ..common import KafkaError -from .base import ( +from kafka.common import KafkaError +from kafka.consumer.base import ( Consumer, AUTO_COMMIT_MSG_COUNT, AUTO_COMMIT_INTERVAL, NO_MESSAGES_WAIT_TIME_SECONDS, FULL_QUEUE_WAIT_TIME_SECONDS, MAX_BACKOFF_SECONDS, ) -from .simple import SimpleConsumer +from kafka.consumer.simple import SimpleConsumer log = logging.getLogger(__name__) diff --git a/kafka/consumer/simple.py b/kafka/consumer/simple.py index f5b6a99cf..c0c1b1ed3 100644 --- a/kafka/consumer/simple.py +++ b/kafka/consumer/simple.py @@ -12,7 +12,7 @@ from kafka.vendor import six from kafka.vendor.six.moves import queue # pylint: disable=import-error -from .base import ( +from kafka.consumer.base import ( Consumer, FETCH_DEFAULT_BLOCK_TIMEOUT, AUTO_COMMIT_MSG_COUNT, @@ -24,7 +24,7 @@ ITER_TIMEOUT_SECONDS, NO_MESSAGES_WAIT_TIME_SECONDS ) -from ..common import ( +from kafka.common import ( FetchRequestPayload, KafkaError, OffsetRequestPayload, ConsumerFetchSizeTooSmall, UnknownTopicOrPartitionError, NotLeaderForPartitionError, diff --git a/kafka/coordinator/assignors/range.py b/kafka/coordinator/assignors/range.py index cbf411e5c..c232d9e41 100644 --- a/kafka/coordinator/assignors/range.py +++ b/kafka/coordinator/assignors/range.py @@ -5,8 +5,8 @@ from kafka.vendor import six -from .abstract import AbstractPartitionAssignor -from ..protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment +from kafka.coordinator.assignors.abstract import AbstractPartitionAssignor +from kafka.coordinator.protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment log = logging.getLogger(__name__) diff --git a/kafka/coordinator/assignors/roundrobin.py b/kafka/coordinator/assignors/roundrobin.py index c24168599..a8310338c 100644 --- a/kafka/coordinator/assignors/roundrobin.py +++ b/kafka/coordinator/assignors/roundrobin.py @@ -6,9 +6,9 @@ from kafka.vendor import six -from .abstract import AbstractPartitionAssignor -from ...common import TopicPartition -from ..protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment +from kafka.coordinator.assignors.abstract import AbstractPartitionAssignor +from kafka.common import TopicPartition +from kafka.coordinator.protocol import ConsumerProtocolMemberMetadata, ConsumerProtocolMemberAssignment log = logging.getLogger(__name__) diff --git a/kafka/coordinator/base.py b/kafka/coordinator/base.py index 301c06d67..a41112bb3 100644 --- a/kafka/coordinator/base.py +++ b/kafka/coordinator/base.py @@ -10,13 +10,13 @@ from kafka.vendor import six -from .heartbeat import Heartbeat -from .. import errors as Errors -from ..future import Future -from ..metrics import AnonMeasurable -from ..metrics.stats import Avg, Count, Max, Rate -from ..protocol.commit import GroupCoordinatorRequest, OffsetCommitRequest -from ..protocol.group import (HeartbeatRequest, JoinGroupRequest, +from kafka.coordinator.heartbeat import Heartbeat +from kafka import errors as Errors +from kafka.future import Future +from kafka.metrics import AnonMeasurable +from kafka.metrics.stats import Avg, Count, Max, Rate +from kafka.protocol.commit import GroupCoordinatorRequest, OffsetCommitRequest +from kafka.protocol.group import (HeartbeatRequest, JoinGroupRequest, LeaveGroupRequest, SyncGroupRequest) log = logging.getLogger('kafka.coordinator') diff --git a/kafka/coordinator/consumer.py b/kafka/coordinator/consumer.py index ab30883c8..651b4212f 100644 --- a/kafka/coordinator/consumer.py +++ b/kafka/coordinator/consumer.py @@ -7,17 +7,17 @@ from kafka.vendor import six -from .base import BaseCoordinator, Generation -from .assignors.range import RangePartitionAssignor -from .assignors.roundrobin import RoundRobinPartitionAssignor -from .protocol import ConsumerProtocol -from .. import errors as Errors -from ..future import Future -from ..metrics import AnonMeasurable -from ..metrics.stats import Avg, Count, Max, Rate -from ..protocol.commit import OffsetCommitRequest, OffsetFetchRequest -from ..structs import OffsetAndMetadata, TopicPartition -from ..util import WeakMethod +from kafka.coordinator.base import BaseCoordinator, Generation +from kafka.coordinator.assignors.range import RangePartitionAssignor +from kafka.coordinator.assignors.roundrobin import RoundRobinPartitionAssignor +from kafka.coordinator.protocol import ConsumerProtocol +from kafka import errors as Errors +from kafka.future import Future +from kafka.metrics import AnonMeasurable +from kafka.metrics.stats import Avg, Count, Max, Rate +from kafka.protocol.commit import OffsetCommitRequest, OffsetFetchRequest +from kafka.structs import OffsetAndMetadata, TopicPartition +from kafka.util import WeakMethod log = logging.getLogger(__name__) diff --git a/kafka/metrics/__init__.py b/kafka/metrics/__init__.py index 6055142a6..2a62d6334 100644 --- a/kafka/metrics/__init__.py +++ b/kafka/metrics/__init__.py @@ -1,13 +1,13 @@ from __future__ import absolute_import -from .compound_stat import NamedMeasurable -from .dict_reporter import DictReporter -from .kafka_metric import KafkaMetric -from .measurable import AnonMeasurable -from .metric_config import MetricConfig -from .metric_name import MetricName -from .metrics import Metrics -from .quota import Quota +from kafka.metrics.compound_stat import NamedMeasurable +from kafka.metrics.dict_reporter import DictReporter +from kafka.metrics.kafka_metric import KafkaMetric +from kafka.metrics.measurable import AnonMeasurable +from kafka.metrics.metric_config import MetricConfig +from kafka.metrics.metric_name import MetricName +from kafka.metrics.metrics import Metrics +from kafka.metrics.quota import Quota __all__ = [ 'AnonMeasurable', 'DictReporter', 'KafkaMetric', 'MetricConfig', diff --git a/kafka/metrics/stats/__init__.py b/kafka/metrics/stats/__init__.py index ab1fb715f..a3d535dfd 100644 --- a/kafka/metrics/stats/__init__.py +++ b/kafka/metrics/stats/__init__.py @@ -1,15 +1,15 @@ from __future__ import absolute_import -from .avg import Avg -from .count import Count -from .histogram import Histogram -from .max_stat import Max -from .min_stat import Min -from .percentile import Percentile -from .percentiles import Percentiles -from .rate import Rate -from .sensor import Sensor -from .total import Total +from kafka.metrics.stats.avg import Avg +from kafka.metrics.stats.count import Count +from kafka.metrics.stats.histogram import Histogram +from kafka.metrics.stats.max_stat import Max +from kafka.metrics.stats.min_stat import Min +from kafka.metrics.stats.percentile import Percentile +from kafka.metrics.stats.percentiles import Percentiles +from kafka.metrics.stats.rate import Rate +from kafka.metrics.stats.sensor import Sensor +from kafka.metrics.stats.total import Total __all__ = [ 'Avg', 'Count', 'Histogram', 'Max', 'Min', 'Percentile', 'Percentiles', diff --git a/kafka/partitioner/__init__.py b/kafka/partitioner/__init__.py index 299b485d9..a9dbbdccb 100644 --- a/kafka/partitioner/__init__.py +++ b/kafka/partitioner/__init__.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .default import DefaultPartitioner -from .hashed import HashedPartitioner, Murmur2Partitioner, LegacyPartitioner -from .roundrobin import RoundRobinPartitioner +from kafka.partitioner.default import DefaultPartitioner +from kafka.partitioner.hashed import HashedPartitioner, Murmur2Partitioner, LegacyPartitioner +from kafka.partitioner.roundrobin import RoundRobinPartitioner __all__ = [ 'DefaultPartitioner', 'RoundRobinPartitioner', 'HashedPartitioner', diff --git a/kafka/partitioner/default.py b/kafka/partitioner/default.py index 087166c0f..e4d9df5dc 100644 --- a/kafka/partitioner/default.py +++ b/kafka/partitioner/default.py @@ -2,7 +2,7 @@ import random -from .hashed import murmur2 +from kafka.partitioner.hashed import murmur2 class DefaultPartitioner(object): diff --git a/kafka/partitioner/hashed.py b/kafka/partitioner/hashed.py index 06307f08d..be92daffa 100644 --- a/kafka/partitioner/hashed.py +++ b/kafka/partitioner/hashed.py @@ -2,7 +2,7 @@ from kafka.vendor import six -from .base import Partitioner +from kafka.partitioner.base import Partitioner class Murmur2Partitioner(Partitioner): diff --git a/kafka/partitioner/roundrobin.py b/kafka/partitioner/roundrobin.py index 9ac2ed0cd..e68c37242 100644 --- a/kafka/partitioner/roundrobin.py +++ b/kafka/partitioner/roundrobin.py @@ -1,6 +1,6 @@ from __future__ import absolute_import -from .base import Partitioner +from kafka.partitioner.base import Partitioner class RoundRobinPartitioner(Partitioner): diff --git a/kafka/producer/__init__.py b/kafka/producer/__init__.py index 5213fe818..54fd8d2ae 100644 --- a/kafka/producer/__init__.py +++ b/kafka/producer/__init__.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .kafka import KafkaProducer -from .simple import SimpleProducer -from .keyed import KeyedProducer +from kafka.producer.kafka import KafkaProducer +from kafka.producer.simple import SimpleProducer +from kafka.producer.keyed import KeyedProducer __all__ = [ 'KafkaProducer', diff --git a/kafka/producer/buffer.py b/kafka/producer/buffer.py index 19ea7322e..8a8d7174c 100644 --- a/kafka/producer/buffer.py +++ b/kafka/producer/buffer.py @@ -5,7 +5,7 @@ import threading import time -from ..metrics.stats import Rate +from kafka.metrics.stats import Rate import kafka.errors as Errors diff --git a/kafka/producer/future.py b/kafka/producer/future.py index e39a0a97a..aa216c4e5 100644 --- a/kafka/producer/future.py +++ b/kafka/producer/future.py @@ -3,8 +3,8 @@ import collections import threading -from .. import errors as Errors -from ..future import Future +from kafka import errors as Errors +from kafka.future import Future class FutureProduceResult(Future): diff --git a/kafka/producer/kafka.py b/kafka/producer/kafka.py index 5e2fb6900..ae2877920 100644 --- a/kafka/producer/kafka.py +++ b/kafka/producer/kafka.py @@ -8,20 +8,20 @@ import time import weakref -from ..vendor import six - -from .. import errors as Errors -from ..client_async import KafkaClient, selectors -from ..codec import has_gzip, has_snappy, has_lz4 -from ..metrics import MetricConfig, Metrics -from ..partitioner.default import DefaultPartitioner -from ..record.default_records import DefaultRecordBatchBuilder -from ..record.legacy_records import LegacyRecordBatchBuilder -from ..serializer import Serializer -from ..structs import TopicPartition -from .future import FutureRecordMetadata, FutureProduceResult -from .record_accumulator import AtomicInteger, RecordAccumulator -from .sender import Sender +from kafka.vendor import six + +from kafka import errors as Errors +from kafka.client_async import KafkaClient, selectors +from kafka.codec import has_gzip, has_snappy, has_lz4 +from kafka.metrics import MetricConfig, Metrics +from kafka.partitioner.default import DefaultPartitioner +from kafka.record.default_records import DefaultRecordBatchBuilder +from kafka.record.legacy_records import LegacyRecordBatchBuilder +from kafka.serializer import Serializer +from kafka.structs import TopicPartition +from kafka.producer.future import FutureRecordMetadata, FutureProduceResult +from kafka.producer.record_accumulator import AtomicInteger, RecordAccumulator +from kafka.producer.sender import Sender log = logging.getLogger(__name__) diff --git a/kafka/producer/keyed.py b/kafka/producer/keyed.py index 9fba33bbf..8de3ad80f 100644 --- a/kafka/producer/keyed.py +++ b/kafka/producer/keyed.py @@ -3,8 +3,8 @@ import logging import warnings -from .base import Producer -from ..partitioner import HashedPartitioner +from kafka.producer.base import Producer +from kafka.partitioner import HashedPartitioner log = logging.getLogger(__name__) diff --git a/kafka/producer/record_accumulator.py b/kafka/producer/record_accumulator.py index 5158474f8..61f1e0e2a 100644 --- a/kafka/producer/record_accumulator.py +++ b/kafka/producer/record_accumulator.py @@ -6,10 +6,10 @@ import threading import time -from .. import errors as Errors -from .buffer import SimpleBufferPool -from .future import FutureRecordMetadata, FutureProduceResult -from ..structs import TopicPartition +from kafka import errors as Errors +from kafka.producer.buffer import SimpleBufferPool +from kafka.producer.future import FutureRecordMetadata, FutureProduceResult +from kafka.structs import TopicPartition from kafka.record.memory_records import MemoryRecordsBuilder from kafka.record.legacy_records import LegacyRecordBatchBuilder diff --git a/kafka/producer/sender.py b/kafka/producer/sender.py index 48ad06e64..895045da6 100644 --- a/kafka/producer/sender.py +++ b/kafka/producer/sender.py @@ -8,12 +8,12 @@ from kafka.vendor import six -from .. import errors as Errors -from ..metrics.measurable import AnonMeasurable -from ..metrics.stats import Avg, Max, Rate -from ..protocol.produce import ProduceRequest -from ..structs import TopicPartition -from ..version import __version__ +from kafka import errors as Errors +from kafka.metrics.measurable import AnonMeasurable +from kafka.metrics.stats import Avg, Max, Rate +from kafka.protocol.produce import ProduceRequest +from kafka.structs import TopicPartition +from kafka.version import __version__ log = logging.getLogger(__name__) diff --git a/kafka/producer/simple.py b/kafka/producer/simple.py index 90b3d4a7d..589363c93 100644 --- a/kafka/producer/simple.py +++ b/kafka/producer/simple.py @@ -6,7 +6,7 @@ from kafka.vendor.six.moves import xrange # pylint: disable=import-error -from .base import Producer +from kafka.producer.base import Producer log = logging.getLogger(__name__) diff --git a/kafka/protocol/__init__.py b/kafka/protocol/__init__.py index 4dcf4a4eb..55ecb867f 100644 --- a/kafka/protocol/__init__.py +++ b/kafka/protocol/__init__.py @@ -1,6 +1,6 @@ from __future__ import absolute_import -from .legacy import ( +from kafka.protocol.legacy import ( create_message, create_gzip_message, create_snappy_message, create_message_set, CODEC_NONE, CODEC_GZIP, CODEC_SNAPPY, ALL_CODECS, diff --git a/kafka/protocol/admin.py b/kafka/protocol/admin.py index 09746bf5e..f535ba119 100644 --- a/kafka/protocol/admin.py +++ b/kafka/protocol/admin.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Boolean, Bytes, Int16, Int32, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Boolean, Bytes, Int16, Int32, Schema, String class ApiVersionResponse_v0(Response): diff --git a/kafka/protocol/api.py b/kafka/protocol/api.py index ec24a3993..efaf63ea2 100644 --- a/kafka/protocol/api.py +++ b/kafka/protocol/api.py @@ -2,8 +2,8 @@ import abc -from .struct import Struct -from .types import Int16, Int32, String, Schema +from kafka.protocol.struct import Struct +from kafka.protocol.types import Int16, Int32, String, Schema class RequestHeader(Struct): diff --git a/kafka/protocol/commit.py b/kafka/protocol/commit.py index 9d744c782..31fc23707 100644 --- a/kafka/protocol/commit.py +++ b/kafka/protocol/commit.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Int8, Int16, Int32, Int64, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String class OffsetCommitResponse_v0(Response): diff --git a/kafka/protocol/fetch.py b/kafka/protocol/fetch.py index 0b03845ee..3904b9a73 100644 --- a/kafka/protocol/fetch.py +++ b/kafka/protocol/fetch.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Int8, Int16, Int32, Int64, Schema, String, Bytes +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String, Bytes class FetchResponse_v0(Response): diff --git a/kafka/protocol/group.py b/kafka/protocol/group.py index c6acca83f..60b3f9e6f 100644 --- a/kafka/protocol/group.py +++ b/kafka/protocol/group.py @@ -1,8 +1,8 @@ from __future__ import absolute_import -from .api import Request, Response -from .struct import Struct -from .types import Array, Bytes, Int16, Int32, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.struct import Struct +from kafka.protocol.types import Array, Bytes, Int16, Int32, Schema, String class JoinGroupResponse_v0(Response): diff --git a/kafka/protocol/message.py b/kafka/protocol/message.py index a330ed805..19dcbd9de 100644 --- a/kafka/protocol/message.py +++ b/kafka/protocol/message.py @@ -3,15 +3,15 @@ import io import time -from ..codec import (has_gzip, has_snappy, has_lz4, +from kafka.codec import (has_gzip, has_snappy, has_lz4, gzip_decode, snappy_decode, lz4_decode, lz4_decode_old_kafka) -from .frame import KafkaBytes -from .struct import Struct -from .types import ( +from kafka.protocol.frame import KafkaBytes +from kafka.protocol.struct import Struct +from kafka.protocol.types import ( Int8, Int32, Int64, Bytes, Schema, AbstractType ) -from ..util import crc32, WeakMethod +from kafka.util import crc32, WeakMethod class Message(Struct): diff --git a/kafka/protocol/metadata.py b/kafka/protocol/metadata.py index 2be82090e..e13d6a7d7 100644 --- a/kafka/protocol/metadata.py +++ b/kafka/protocol/metadata.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Boolean, Int16, Int32, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Boolean, Int16, Int32, Schema, String class MetadataResponse_v0(Response): diff --git a/kafka/protocol/offset.py b/kafka/protocol/offset.py index 517965836..3c254de40 100644 --- a/kafka/protocol/offset.py +++ b/kafka/protocol/offset.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Array, Int8, Int16, Int32, Int64, Schema, String +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Array, Int8, Int16, Int32, Int64, Schema, String UNKNOWN_OFFSET = -1 diff --git a/kafka/protocol/produce.py b/kafka/protocol/produce.py index 34ff949ef..b65de58e2 100644 --- a/kafka/protocol/produce.py +++ b/kafka/protocol/produce.py @@ -1,7 +1,7 @@ from __future__ import absolute_import -from .api import Request, Response -from .types import Int16, Int32, Int64, String, Array, Schema, Bytes +from kafka.protocol.api import Request, Response +from kafka.protocol.types import Int16, Int32, Int64, String, Array, Schema, Bytes class ProduceResponse_v0(Response): diff --git a/kafka/protocol/struct.py b/kafka/protocol/struct.py index 3288172cf..676de1ba4 100644 --- a/kafka/protocol/struct.py +++ b/kafka/protocol/struct.py @@ -2,10 +2,10 @@ from io import BytesIO -from .abstract import AbstractType -from .types import Schema +from kafka.protocol.abstract import AbstractType +from kafka.protocol.types import Schema -from ..util import WeakMethod +from kafka.util import WeakMethod class Struct(AbstractType): diff --git a/kafka/protocol/types.py b/kafka/protocol/types.py index 6a6e89e41..d5e446a10 100644 --- a/kafka/protocol/types.py +++ b/kafka/protocol/types.py @@ -2,7 +2,7 @@ from struct import pack, unpack, error -from .abstract import AbstractType +from kafka.protocol.abstract import AbstractType def _pack(f, value): diff --git a/kafka/record/__init__.py b/kafka/record/__init__.py index 4c75acb13..cbd70d93a 100644 --- a/kafka/record/__init__.py +++ b/kafka/record/__init__.py @@ -1,3 +1,3 @@ -from .memory_records import MemoryRecords +from kafka.record.memory_records import MemoryRecords __all__ = ["MemoryRecords"] diff --git a/kafka/record/default_records.py b/kafka/record/default_records.py index 3d517af25..2bbd47e9c 100644 --- a/kafka/record/default_records.py +++ b/kafka/record/default_records.py @@ -57,8 +57,8 @@ import io import struct import time -from .abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder -from .util import decode_varint, encode_varint, calc_crc32c, size_of_varint +from kafka.record.abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder +from kafka.record.util import decode_varint, encode_varint, calc_crc32c, size_of_varint from kafka.errors import CorruptRecordException from kafka.codec import ( diff --git a/kafka/record/legacy_records.py b/kafka/record/legacy_records.py index 8c0791ef9..036e6c45c 100644 --- a/kafka/record/legacy_records.py +++ b/kafka/record/legacy_records.py @@ -44,8 +44,8 @@ import struct import time -from .abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder -from .util import calc_crc32 +from kafka.record.abc import ABCRecord, ABCRecordBatch, ABCRecordBatchBuilder +from kafka.record.util import calc_crc32 from kafka.codec import ( gzip_encode, snappy_encode, lz4_encode, lz4_encode_old_kafka, diff --git a/kafka/record/memory_records.py b/kafka/record/memory_records.py index 56aa51faa..cb1cc01b4 100644 --- a/kafka/record/memory_records.py +++ b/kafka/record/memory_records.py @@ -22,9 +22,9 @@ import struct from kafka.errors import CorruptRecordException -from .abc import ABCRecords -from .legacy_records import LegacyRecordBatch, LegacyRecordBatchBuilder -from .default_records import DefaultRecordBatch, DefaultRecordBatchBuilder +from kafka.record.abc import ABCRecords +from kafka.record.legacy_records import LegacyRecordBatch, LegacyRecordBatchBuilder +from kafka.record.default_records import DefaultRecordBatch, DefaultRecordBatchBuilder class MemoryRecords(ABCRecords): diff --git a/kafka/record/util.py b/kafka/record/util.py index 88135f1a7..55d7adbd0 100644 --- a/kafka/record/util.py +++ b/kafka/record/util.py @@ -1,6 +1,6 @@ import binascii -from ._crc32c import crc as crc32c_py +from kafka.record._crc32c import crc as crc32c_py def encode_varint(value, write): diff --git a/kafka/serializer/__init__.py b/kafka/serializer/__init__.py index c08cffe89..90cd93ab2 100644 --- a/kafka/serializer/__init__.py +++ b/kafka/serializer/__init__.py @@ -1,3 +1,3 @@ from __future__ import absolute_import -from .abstract import Serializer, Deserializer +from kafka.serializer.abstract import Serializer, Deserializer