Gift from folks at LinkedIn
high-performance | low-latency commit log storage | replication | propagation |
---|
ZooKeeper is Centralized service for maintaining:
compression.codec
and compressed.topic
can be used for setting up the compression: Gzip replica.fetch.max.bytes
for each partition they replicatereplica.fetch.max.bytes
is set to 1 MiB, and you have 1000 partitions, about 1 GiB of RAM is requiredfetch.message.max.bytes
settingzookeeper.session.timeout.ms
message.max.bytes
:
Maximum message size the broker will accept.
log.segment.bytes
:
Size of a Kafka data file. Must be larger than any single message.
Default: 1GB
replica.fetch.max.bytes
:
Maximum message size a broker can replicate. Must be larger than message.max.bytes
or a broker can accept messages it cannot replicate, potentially resulting in data loss.
Default: 1MB
max.partition.fetch.bytes
:
The maximum amount of data per-partition the server will return.
Default: 10MB
fetch.max.bytes
:
The maximum amount of data the server should return for a fetch request.
Default: 50MB
fetch.message.max.bytes
:
Maximum message size a consumer can read. Must be at least as large as message.max.bytes
Default: 1MB
send()
command, the result returned is a future
batch.size
: measures batch size in total bytes instead of the number of messages linger.ms
: maximum time to buffer data in asynchronous modelinger.ms
to 5 and send more messages in one batch.replica.high.watermark.checkpoint.interval.ms
:
If you have to go back and locate missing data,
you have a checkpoint from which to move forward without having to reread prior data.