Pulsar adaptor for Apache Kafka
Pulsar provides an easy option for applications that are currently written using the Apache Kafka Java client API.
Use the Pulsar Kafka compatibility wrapper
In an existing application, change the regular Kafka client dependency and replace it with the Pulsar Kafka wrapper. Remove the following dependency in pom.xml
:
<dependency>
<groupId>org.apache.kafka</groupId>
<artifactId>kafka-clients</artifactId>
<version>0.10.2.1</version>
</dependency>
Then include this dependency for the Pulsar Kafka wrapper:
<dependency>
<groupId>org.apache.pulsar</groupId>
<artifactId>pulsar-client-kafka</artifactId>
<version>3.1.1</version>
</dependency>
With the new dependency, the existing code works without any changes. You need to adjust the configuration, and make sure it points the producers and consumers to Pulsar service rather than Kafka, and uses a particular Pulsar topic.
Use the Pulsar Kafka compatibility wrapper together with existing Kafka client
When migrating from Kafka to Pulsar, the application might use the original Kafka client and the Pulsar Kafka wrapper together during migration. You should consider using the unshaded Pulsar Kafka client wrapper.
<dependency>
<groupId>org.apache.pulsar</groupId>
<artifactId>pulsar-client-kafka-original</artifactId>
<version>3.1.1</version>
</dependency>
When using this dependency, construct producers using org.apache.kafka.clients.producer.PulsarKafkaProducer
instead of org.apache.kafka.clients.producer.KafkaProducer
and org.apache.kafka.clients.producer.PulsarKafkaConsumer
for consumers.
Producer example
// Topic needs to be a regular Pulsar topic
String topic = "persistent://public/default/my-topic";
Properties props = new Properties();
// Point to a Pulsar service
props.put("bootstrap.servers", "pulsar://localhost:6650");
props.put("key.serializer", IntegerSerializer.class.getName());
props.put("value.serializer", StringSerializer.class.getName());
Producer<Integer, String> producer = new KafkaProducer(props);
for (int i = 0; i < 10; i++) {
producer.send(new ProducerRecord<Integer, String>(topic, i, "hello-" + i));
log.info("Message {} sent successfully", i);
}
producer.close();
Consumer example
String topic = "persistent://public/default/my-topic";
Properties props = new Properties();
// Point to a Pulsar service
props.put("bootstrap.servers", "pulsar://localhost:6650");
props.put("group.id", "my-subscription-name");
props.put("enable.auto.commit", "false");
props.put("key.deserializer", IntegerDeserializer.class.getName());
props.put("value.deserializer", StringDeserializer.class.getName());
Consumer<Integer, String> consumer = new KafkaConsumer(props);
consumer.subscribe(Arrays.asList(topic));
while (true) {
ConsumerRecords<Integer, String> records = consumer.poll(100);
records.forEach(record -> {
log.info("Received record: {}", record);
});
// Commit last offset
consumer.commitSync();
}
Complete Examples
You can find the complete producer and consumer examples here.
Compatibility matrix
Currently, the Pulsar Kafka wrapper supports most of the operations offered by the Kafka API.
Producer
APIs:
Producer Method | Supported | Notes |
---|---|---|
Future<RecordMetadata> send(ProducerRecord<K, V> record) | Yes | |
Future<RecordMetadata> send(ProducerRecord<K, V> record, Callback callback) | Yes | |
void flush() | Yes | |
List<PartitionInfo> partitionsFor(String topic) | No | |
Map<MetricName, ? extends Metric> metrics() | No | |
void close() | Yes | |
void close(long timeout, TimeUnit unit) | Yes |
Properties:
Config property | Supported | Notes |
---|---|---|
acks | Ignored | Durability and quorum writes are configured at the namespace level |
auto.offset.reset | Yes | It uses a default value of earliest if you do not give a specific setting. |
batch.size | Ignored | |
bootstrap.servers | Yes | |
buffer.memory | Ignored | |
client.id | Ignored | |
compression.type | Yes | Allows gzip and lz4 . No snappy . |
connections.max.idle.ms | Yes | Only support up to 2,147,483,647,000(Integer.MAX_VALUE * 1000) ms of idle time |
interceptor.classes | Yes | |
key.serializer | Yes | |
linger.ms | Yes | Controls the group commit time when batching messages |
max.block.ms | Ignored | |
max.in.flight.requests.per.connection | Ignored | In Pulsar ordering is maintained even with multiple requests in flight |
max.request.size | Ignored | |
metric.reporters | Ignored | |
metrics.num.samples | Ignored | |
metrics.sample.window.ms | Ignored | |
partitioner.class | Yes | |
receive.buffer.bytes | Ignored | |
reconnect.backoff.ms | Ignored | |
request.timeout.ms | Ignored | |
retries | Ignored | Pulsar client retries with exponential backoff until the send timeout expires. |
send.buffer.bytes | Ignored | |
timeout.ms | Yes | |
value.serializer | Yes |
Consumer
The following table lists consumer APIs.
Consumer Method | Supported | Notes |
---|---|---|
Set<TopicPartition> assignment() | No | |
Set<String> subscription() | Yes | |
void subscribe(Collection<String> topics) | Yes | |
void subscribe(Collection<String> topics, ConsumerRebalanceListener callback) | No | |
void assign(Collection<TopicPartition> partitions) | No | |
void subscribe(Pattern pattern, ConsumerRebalanceListener callback) | No | |
void unsubscribe() | Yes | |
ConsumerRecords<K, V> poll(long timeoutMillis) | Yes | |
void commitSync() | Yes | |
void commitSync(Map<TopicPartition, OffsetAndMetadata> offsets) | Yes | |
void commitAsync() | Yes | |
void commitAsync(OffsetCommitCallback callback) | Yes | |
void commitAsync(Map<TopicPartition, OffsetAndMetadata> offsets, OffsetCommitCallback callback) | Yes | |
void seek(TopicPartition partition, long offset) | Yes | |
void seekToBeginning(Collection<TopicPartition> partitions) | Yes | |
void seekToEnd(Collection<TopicPartition> partitions) | Yes | |
long position(TopicPartition partition) | Yes | |
OffsetAndMetadata committed(TopicPartition partition) | Yes | |
Map<MetricName, ? extends Metric> metrics() | No | |
List<PartitionInfo> partitionsFor(String topic) | No | |
Map<String, List<PartitionInfo>> listTopics() | No | |
Set<TopicPartition> paused() | No | |
void pause(Collection<TopicPartition> partitions) | No | |
void resume(Collection<TopicPartition> partitions) | No | |
Map<TopicPartition, OffsetAndTimestamp> offsetsForTimes(Map<TopicPartition, Long> timestampsToSearch) | No | |
Map<TopicPartition, Long> beginningOffsets(Collection<TopicPartition> partitions) | No | |
Map<TopicPartition, Long> endOffsets(Collection<TopicPartition> partitions) | No | |
void close() | Yes | |
void close(long timeout, TimeUnit unit) | Yes | |
void wakeup() | No |
Properties:
Config property | Supported | Notes |
---|---|---|
group.id | Yes | Maps to a Pulsar subscription name |
max.poll.records | Yes | |
max.poll.interval.ms | Ignored | Messages are “pushed” from broker |
session.timeout.ms | Ignored | |
heartbeat.interval.ms | Ignored | |
bootstrap.servers | Yes | Needs to point to a single Pulsar service URL |
enable.auto.commit | Yes | |
auto.commit.interval.ms | Ignored | With auto-commit, acks are sent immediately to broker |
partition.assignment.strategy | Ignored | |
auto.offset.reset | Yes | Only support earliest and latest. |
fetch.min.bytes | Ignored | |
fetch.max.bytes | Ignored | |
fetch.max.wait.ms | Ignored | |
interceptor.classes | Yes | |
metadata.max.age.ms | Ignored | |
max.partition.fetch.bytes | Ignored | |
send.buffer.bytes | Ignored | |
receive.buffer.bytes | Ignored | |
client.id | Ignored |
Customize Pulsar configurations
You can configure Pulsar authentication provider directly from the Kafka properties.
Pulsar client properties
Config property | Default | Notes |
---|---|---|
pulsar.authentication.class | Configure to auth provider. For example, org.apache.pulsar.client.impl.auth.AuthenticationTls . | |
pulsar.authentication.params.map | Map which represents parameters for the Authentication-Plugin. | |
pulsar.authentication.params.string | String which represents parameters for the Authentication-Plugin, for example, key1:val1,key2:val2 . | |
pulsar.use.tls | false | Enable TLS transport encryption. |
pulsar.tls.trust.certs.file.path | Path for the TLS trust certificate store. | |
pulsar.tls.allow.insecure.connection | false | Accept self-signed certificates from brokers. |
pulsar.operation.timeout.ms | 30000 | General operations timeout. |
pulsar.stats.interval.seconds | 60 | Pulsar client lib stats printing interval. |
pulsar.num.io.threads | 1 | The number of Netty IO threads to use. |
pulsar.connections.per.broker | 1 | The maximum number of connection to each broker. |
pulsar.use.tcp.nodelay | true | TCP no-delay. |
pulsar.concurrent.lookup.requests | 50000 | The maximum number of concurrent topic lookups. |
pulsar.max.number.rejected.request.per.connection | 50 | The threshold of errors to forcefully close a connection. |
pulsar.keepalive.interval.ms | 30000 | Keep alive interval for each client-broker-connection. |
Pulsar producer properties
Config property | Default | Notes |
---|---|---|
pulsar.producer.name | Specify the producer name. | |
pulsar.producer.initial.sequence.id | Specify baseline for sequence ID of this producer. | |
pulsar.producer.max.pending.messages | 1000 | Set the maximum size of the message queue pending to receive an acknowledgment from the broker. |
pulsar.producer.max.pending.messages.across.partitions | 50000 | Set the maximum number of pending messages across all the partitions. |
pulsar.producer.batching.enabled | true | Control whether automatic batching of messages is enabled for the producer. |
pulsar.producer.batching.max.messages | 1000 | The maximum number of messages in a batch. |
pulsar.block.if.producer.queue.full | Specify the block producer if queue is full. | |
pulsar.crypto.reader.factory.class.name | Specify the CryptoReader-Factory(CryptoKeyReaderFactory ) classname which allows producer to create CryptoKeyReader. |
Pulsar consumer Properties
Config property | Default | Notes |
---|---|---|
pulsar.consumer.name | Specify the consumer name. | |
pulsar.consumer.receiver.queue.size | 1000 | Set the size of the consumer receiver queue. |
pulsar.consumer.acknowledgments.group.time.millis | 100 | Set the maximum amount of group time for consumers to send the acknowledgments to the broker. |
pulsar.consumer.total.receiver.queue.size.across.partitions | 50000 | Set the maximum size of the total receiver queue across partitions. |
pulsar.consumer.subscription.topics.mode | PersistentOnly | Set the subscription topic mode for consumers. |
pulsar.crypto.reader.factory.class.name | Specify the CryptoReader-Factory(CryptoKeyReaderFactory ) classname which allows consumer to create CryptoKeyReader. |