在Java中使用Kafka

Producer部分

Producer在实例化后, 对外提供send方法, 用于将数据送到指定的topic和partition; 以及在退出时需要的destroy方法.

接口 KafkaProducer.java

import java.util.List;
import java.util.Properties;

public interface KafkaProducer<D> {

    default void init() {
    }
    default void destroy() {
    }
    boolean send(String topic, D data);
    boolean send(String topic, Integer partition, D data);
    boolean send(String topic, List<D> dataList);
    boolean send(String topic, Integer partition, List<D> dataList);

    /**
     * 默认配置
     */
    default Properties getDefaultProps() {
        Properties props = new Properties();
        props.put("acks", "1");
        props.put("retries", 1);
        props.put("batch.size", 16384);
        props.put("linger.ms", 1);
        props.put("buffer.memory", 32 * 1024 * 1024L);
        return props;
    }
}

参数说明

Properties props = new Properties(); 
props.put("bootstrap.servers", "localhost:9092"); 
// The acks config controls the criteria under which requests are considered complete. The "all" setting we have specified will result in blocking on the full commit of the record, the slowest but most durable setting. 
props.put("acks", "all"); 
// If the request fails, the producer can automatically retry, though since we have specified retries as 0 it won't. Enabling retries also opens up the possibility of duplicates (see the documentation on message delivery semantics for details). 
props.put("retries", 0); 
// The producer maintains buffers of unsent records for each partition. These buffers are of a size specified by the batch.size config. Making this larger can result in more batching, but requires more memory (since we will generally have one of these buffers for each active partition). 
props.put("batch.size", 16384); 
// By default a buffer is available to send immediately even if there is additional unused space in the buffer. However if you want to reduce the number of requests you can set linger.ms to something greater than 0. This will instruct the producer to wait up to that number of milliseconds before sending a request in hope that more records will arrive to fill up the same batch.
props.put("linger.ms", 1); 
// 生产者缓冲大小,当缓冲区耗尽后,额外的发送调用将被阻塞。时间超过max.block.ms将抛出TimeoutException 
props.put("buffer.memory", 33554432); 
// The key.serializer and value.serializer instruct how to turn the key and value objects the user provides with their ProducerRecord into bytes. You can use the included ByteArraySerializer or StringSerializer for simple string or byte types. 
props.put("key.serializer", "org.apache.kafka.common.serialization.StringSerializer"); 
props.put("value.serializer", "org.apache.kafka.common.serialization.StringSerializer"); 

实现 KafkaProducerImpl.java

import com.google.common.base.Strings;
import org.apache.kafka.clients.producer.Producer;
import org.apache.kafka.clients.producer.ProducerRecord;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.util.List;
import java.util.Map;
import java.util.Properties;

public class KafkaProducerImpl<D> implements KafkaProducer<D> {
    private static final Logger logger = LoggerFactory.getLogger(KafkaProducerImpl.class);
    private final Producer<D, D> producer;

    public KafkaProducerImpl() {
        Properties props = this.getDefaultProps();
        props.put("bootstrap.servers", servers);
        props.put("key.serializer", serializer);
        props.put("value.serializer", serializer);
        producer = new org.apache.kafka.clients.producer.KafkaProducer<>(props);
    }

    @Override
    public void destroy() {
        if (producer != null) {
            producer.close();
        }
    }

    @Override
    public boolean send(String topic, D data) {
        boolean isSuc = true;
        try {
            producer.send(new ProducerRecord<>(topic, data));
        } catch (Exception e) {
            isSuc = false;
            logger.error(String.format("KafkaStringProducer send error.topic:[%s],data:[%s]", topic, data), e);
        }
        return isSuc;
    }

    @Override
    public boolean send(String topic, Integer partition, D data) {
        boolean isSuc = true;
        try {
            producer.send(new ProducerRecord<>(topic, partition, null, data));
        } catch (Exception e) {
            isSuc = false;
            logger.error(String.format("KafkaStringProducer send error.topic:[%s],data:[%s]", topic, data), e);
        }
        return isSuc;
    }

    @Override
    public boolean send(String topic, List<D> dataList) {
        boolean isSuc = true;
        try {
            if (dataList != null) {
                dataList.forEach(item -> producer.send(new ProducerRecord<>(topic, item)));
            }
        } catch (Exception e) {
            isSuc = false;
            logger.error(String.format("KafkaStringProducer send error.topic:[%s],dataList:[%s]", topic, dataList), e);
        }
        return isSuc;
    }

    @Override
    public boolean send(String topic, Integer partition, List<D> dataList) {
        boolean isSuc = true;
        try {
            if (dataList != null) {
                dataList.forEach(item -> producer.send(new ProducerRecord<>(topic, partition, null, item)));
            }
        } catch (Exception e) {
            isSuc = false;
            logger.error(String.format("KafkaStringProducer send error.topic:[%s],partition[%s],dataList:[%s]", topic, partition, dataList), e);
        }
        return isSuc;
    }
}

Consumer 部分

Consumer 在实例化后, 负责将ConsumerListener添加到列表, 并订阅指定的topic, 启动一个阻塞的循环, 在收到消息后依次调用ConsumerListener进行处理

接口 KafkaConsumer.java

import java.util.Properties;

public interface KafkaConsumer {

    default void init() {
    }

    default void destroy() {
    }

    void start();

    /**
     * 默认配置
     */
    default Properties getDefaultProps() {
        Properties props = new Properties();
        props.put("enable.auto.commit", "true");
        props.put("auto.commit.interval.ms", "1000");
        props.put("session.timeout.ms", "30000");
        return props;
    }
}  

参数说明

Properties props = new Properties();
props.put("bootstrap.servers", "localhost:9092");
props.put("group.id", "test");
// Setting enable.auto.commit means that offsets are committed automatically with a frequency controlled by the config auto.commit.interval.ms.
props.put("enable.auto.commit", "true");
props.put("auto.commit.interval.ms", "1000");
// The deserializer settings specify how to turn bytes into objects. For example, by specifying string deserializers, we are saying that our record's key and value will just be simple strings. 
props.put("key.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
props.put("value.deserializer", "org.apache.kafka.common.serialization.StringDeserializer");
KafkaConsumer<String, String> consumer = new KafkaConsumer<>(props);
// This consumer is subscribing to the topics foo and bar as part of a group of consumers called test as configured with group.id. 
consumer.subscribe(Arrays.asList("foo", "bar"));
while (true) {
 ConsumerRecords<String, String> records = consumer.poll(100);
 for (ConsumerRecord<String, String> record : records)
     System.out.printf("offset = %d, key = %s, value = %s%n", record.offset(), record.key(), record.value());
}

实现 KafkaConsumerImpl.java

import com.google.common.base.Strings;
import org.apache.kafka.clients.consumer.Consumer;
import org.apache.kafka.clients.consumer.ConsumerRecord;
import org.apache.kafka.clients.consumer.ConsumerRecords;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.util.*;

public class KafkaConsumerImpl<K, V> implements KafkaConsumer {
    private static final Logger logger = LoggerFactory.getLogger(KafkaConsumerImpl.class);
    private final List<KafkaConsumerListener<K, V>> consumerListeners = new ArrayList<>();
    private Consumer<K, V> consumer;
    private boolean running = true;

    private final int waitingTimeout = 100;

    public KafkaConsumerImpl(String topic, String groupId, String deserializer) {
        Properties props = this.getDefaultProps();
        props.put("group.id", groupId);
        props.put("bootstrap.servers", servers);
        props.put("key.deserializer", deserializer);
        props.put("value.deserializer", deserializer);
        consumer = new org.apache.kafka.clients.consumer.KafkaConsumer<>(props);
        consumer.subscribe(Arrays.asList(topic));
    }

    public void setConsumerListeners(List<KafkaConsumerListener<K, V>> consumerListeners) {
        synchronized (this) {
            this.consumerListeners.clear();
            if (null != consumerListeners && 0 != consumerListeners.size()) {
                consumerListeners.forEach(this.consumerListeners::add);
            }
        }
    }

    public void addConsumerListener(KafkaConsumerListener<K, V> consumerListener) {
        synchronized (this) {
            if (null != consumerListener && !this.consumerListeners.contains(consumerListener)) {
                this.consumerListeners.add(consumerListener);
            }
        }
    }

    public void removeConsumerListener(KafkaConsumerListener<K, V> consumerListener) {
        synchronized (this) {
            if (null != consumerListener && this.consumerListeners.contains(consumerListener)) {
                this.consumerListeners.remove(consumerListener);
            }
        }
    }

    @Override
    public void init() {
        this.start();
    }

    @Override
    public void destroy() {
        running = false;
    }

    @Override
    public void start() {
        new Thread(() -> {
            while (running) {
                ConsumerRecords<K, V> records = consumer.poll(waitingTimeout);
                for (ConsumerRecord<K, V> record : records) {
                    if (consumerListeners != null) {
                        K key = record.key();
                        if (key == null)
                            consumerListeners.forEach(consumer -> consumer.consume(record.value()));
                        else
                            consumerListeners.forEach(consumer -> consumer.consume(record.key(), record.value()));
                    }
                }
            }
            //should use consumer in different thread, or it will throw ConcurrentModificationException
            if (consumer != null) {
                try {
                    logger.info("start to close consumer.");
                    consumer.close();
                } catch (Exception e) {
                    logger.error("close kafka consumer error.", e);
                }
                consumer = null;
            }
        }).start();
    }
}

接口 KafkaConsumerListener.java

public interface KafkaConsumerListener<K, V> {
    void consume(V value);

    default void consume(K key, V value) {
        consume(value);
    }
}

.

原文地址:https://www.cnblogs.com/milton/p/9805317.html