Gift from folks at LinkedIn
| high-performance | low-latency commit log storage | replication | propagation |
|---|
ZooKeeper is Centralized service for maintaining:
compression.codec and compressed.topic can be used for setting up the compression: Gzip replica.fetch.max.bytes for each partition they replicatereplica.fetch.max.bytes is set to 1 MiB, and you have 1000 partitions, about 1 GiB of RAM is requiredfetch.message.max.bytes settingzookeeper.session.timeout.ms
message.max.bytes :
Maximum message size the broker will accept.
log.segment.bytes :
Size of a Kafka data file. Must be larger than any single message.
Default: 1GB
replica.fetch.max.bytes :
Maximum message size a broker can replicate. Must be larger than message.max.bytes
or a broker can accept messages it cannot replicate, potentially resulting in data loss.
Default: 1MB
max.partition.fetch.bytes:
The maximum amount of data per-partition the server will return.
Default: 10MB
fetch.max.bytes :
The maximum amount of data the server should return for a fetch request.
Default: 50MB
fetch.message.max.bytes :
Maximum message size a consumer can read. Must be at least as large as message.max.bytes
Default: 1MB
send() command, the result returned is a futurebatch.size : measures batch size in total bytes instead of the number of messages linger.ms : maximum time to buffer data in asynchronous modelinger.ms to 5 and send more messages in one batch.
replica.high.watermark.checkpoint.interval.ms :
If you have to go back and locate missing data,
you have a checkpoint from which to move forward without having to reread prior data.