在springboot中對(duì)kafka進(jìn)行讀寫的示例代碼
springboot對(duì)kafka的client很好的實(shí)現(xiàn)了集成,使用非常方便,本文也實(shí)現(xiàn)了一個(gè)在springboot中實(shí)現(xiàn)操作kafka的demo。
1.POM配置
只需要在dependencies中增加 spring-kafka的配置即可。完整效果如下:
<parent> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-parent</artifactId> <version>1.5.4.RELEASE</version> </parent> <properties> <java.version>1.8</java.version> <spring-kafka.version>1.2.2.RELEASE</spring-kafka.version> <project.build.sourceEncoding>UTF-8</project.build.sourceEncoding> </properties> <dependencies> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-web</artifactId> </dependency> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter</artifactId> </dependency> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-test</artifactId> <scope>test</scope> </dependency> <dependency> <groupId>org.springframework.boot</groupId> <artifactId>spring-boot-starter-aop</artifactId> </dependency> <!-- spring-kafka --> <dependency> <groupId>org.springframework.kafka</groupId> <artifactId>spring-kafka</artifactId> <version>${spring-kafka.version}</version> </dependency> <dependency> <groupId>org.springframework.kafka</groupId> <artifactId>spring-kafka-test</artifactId> <version>${spring-kafka.version}</version> <scope>test</scope> </dependency> </dependencies>
2.生產(chǎn)者
參數(shù)配置類,其參數(shù)卸載yml文件中,通過(guò)@Value注入
package com.dhb.kafka.producer; import org.apache.kafka.clients.producer.ProducerConfig; import org.apache.kafka.common.serialization.StringSerializer; import org.springframework.beans.factory.annotation.Value; import org.springframework.context.annotation.Bean; import org.springframework.context.annotation.Configuration; import org.springframework.kafka.core.DefaultKafkaProducerFactory; import org.springframework.kafka.core.KafkaTemplate; import org.springframework.kafka.core.ProducerFactory; import java.util.HashMap; import java.util.Map; @Configuration public class SenderConfig { @Value("${kafka.bootstrap-servers}") private String bootstrapServers; @Bean public Map<String,Object> producerConfigs() { Map<String,Object> props = new HashMap<>(); props.put(ProducerConfig.BOOTSTRAP_SERVERS_CONFIG,this.bootstrapServers); props.put(ProducerConfig.KEY_SERIALIZER_CLASS_CONFIG, StringSerializer.class); props.put(ProducerConfig.VALUE_SERIALIZER_CLASS_CONFIG,StringSerializer.class); props.put(ProducerConfig.ACKS_CONFIG,"0"); return props; } @Bean public ProducerFactory<String,String> producerFactory() { return new DefaultKafkaProducerFactory<>(producerConfigs()); } @Bean public KafkaTemplate<String,String> kafkaTemplate() { return new KafkaTemplate<String, String>(producerFactory()); } @Bean public Sender sender() { return new Sender(); } }
消息發(fā)送類
package com.dhb.kafka.producer; import lombok.extern.slf4j.Slf4j; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.kafka.core.KafkaTemplate; @Slf4j public class Sender { @Autowired private KafkaTemplate<String,String> kafkaTemplate; public void send(String topic,String payload) { log.info("sending payload='{}' to topic='{}'",payload,topic); this.kafkaTemplate.send(topic,payload); } }
3.消費(fèi)者
參數(shù)配置類
package com.dhb.kafka.consumer; import org.apache.kafka.clients.consumer.ConsumerConfig; import org.apache.kafka.common.serialization.StringDeserializer; import org.springframework.beans.factory.annotation.Value; import org.springframework.context.annotation.Bean; import org.springframework.context.annotation.Configuration; import org.springframework.kafka.annotation.EnableKafka; import org.springframework.kafka.config.ConcurrentKafkaListenerContainerFactory; import org.springframework.kafka.core.ConsumerFactory; import org.springframework.kafka.core.DefaultKafkaConsumerFactory; import java.util.HashMap; import java.util.Map; @Configuration @EnableKafka public class ReceiverConfig { @Value("${kafka.bootstrap-servers}") private String bootstrapServers; public Map<String,Object> consumerConfigs() { Map<String,Object> props = new HashMap<>(); props.put(ConsumerConfig.BOOTSTRAP_SERVERS_CONFIG,bootstrapServers); props.put(ConsumerConfig.KEY_DESERIALIZER_CLASS_CONFIG, StringDeserializer.class); props.put(ConsumerConfig.VALUE_DESERIALIZER_CLASS_CONFIG,StringDeserializer.class); props.put(ConsumerConfig.GROUP_ID_CONFIG,"helloword"); return props; } @Bean public ConsumerFactory<String,String> consumerFactory() { return new DefaultKafkaConsumerFactory<>(consumerConfigs()); } @Bean public ConcurrentKafkaListenerContainerFactory<String,String> kafkaListenerContainerFactory() { ConcurrentKafkaListenerContainerFactory<String,String> factory = new ConcurrentKafkaListenerContainerFactory<>(); factory.setConsumerFactory(consumerFactory()); return factory; } @Bean public Receiver receiver() { return new Receiver(); } }
消息接受類
package com.dhb.kafka.consumer; import lombok.extern.slf4j.Slf4j; import org.springframework.kafka.annotation.KafkaListener; import java.util.concurrent.CountDownLatch; @Slf4j public class Receiver { private CountDownLatch latch = new CountDownLatch(1); public CountDownLatch getLatch() { return latch; } @KafkaListener(topics = "${kafka.topic.helloworld}") public void receive(String payload) { log.info("received payload='{}'",payload); latch.countDown(); } }
3.web測(cè)試類
定義了一個(gè)基于http的web測(cè)試接口
package com.dhb.kafka.web; import com.dhb.kafka.producer.Sender; import lombok.extern.slf4j.Slf4j; import org.springframework.beans.factory.annotation.Autowired; import org.springframework.web.bind.annotation.RequestMapping; import org.springframework.web.bind.annotation.RequestMethod; import org.springframework.web.bind.annotation.RestController; import javax.servlet.http.HttpServletRequest; import javax.servlet.http.HttpServletResponse; import java.io.IOException; @RestController @Slf4j public class KafkaProducer { @Autowired Sender sender; @RequestMapping(value = "/sender.action", method = RequestMethod.POST) public void exec(HttpServletRequest request, HttpServletResponse response,String data) throws IOException{ this.sender.send("testtopic",data); response.setCharacterEncoding("UTF-8"); response.setContentType("text/json"); response.getWriter().write("success"); response.getWriter().flush(); response.getWriter().close(); } }
4.啟動(dòng)類及配置
package com.dhb.kafka; import org.springframework.boot.SpringApplication; import org.springframework.boot.autoconfigure.SpringBootApplication; @SpringBootApplication public class KafkaApplication { public static void main(String[] args) { SpringApplication.run(KafkaApplication.class,args); } }
application.yml
kafka: bootstrap-servers: 192.168.162.239:9092 topic: helloworld: testtopic
程序結(jié)構(gòu):
包結(jié)構(gòu)
5.讀寫測(cè)試
通過(guò)執(zhí)行KafkaApplication的main方法啟動(dòng)程序。然后打開postman進(jìn)行測(cè)試:
運(yùn)行后返回success
生產(chǎn)者日志:
消費(fèi)者日志:
以上就是本文的全部?jī)?nèi)容,希望對(duì)大家的學(xué)習(xí)有所幫助,也希望大家多多支持腳本之家。
相關(guān)文章
基于mybatis高級(jí)映射多對(duì)多查詢的實(shí)現(xiàn)
下面小編就為大家?guī)?lái)一篇基于mybatis高級(jí)映射多對(duì)多查詢的實(shí)現(xiàn)。小編覺(jué)得挺不錯(cuò)的,現(xiàn)在就分享給大家,也給大家做個(gè)參考。一起跟隨小編過(guò)來(lái)看看吧2017-10-10Quarkus改造Pmml模型項(xiàng)目異常記錄及解決處理
這篇文章主要為大家介紹了Quarkus改造Pmml模型項(xiàng)目是遇到的異常記錄以及解決方法,有需要的同學(xué)可以借鑒參考下,希望能夠有所幫助,祝大家多多進(jìn)步2022-02-02Java代碼規(guī)范與質(zhì)量檢測(cè)插件SonarLint的使用
本文主要介紹了Java代碼規(guī)范與質(zhì)量檢測(cè)插件SonarLint的使用,文中通過(guò)示例代碼介紹的非常詳細(xì),對(duì)大家的學(xué)習(xí)或者工作具有一定的參考學(xué)習(xí)價(jià)值,需要的朋友們下面隨著小編來(lái)一起學(xué)習(xí)學(xué)習(xí)吧2022-08-08基于java實(shí)現(xiàn)websocket協(xié)議過(guò)程詳解
這篇文章主要介紹了基于java實(shí)現(xiàn)websocket協(xié)議過(guò)程詳解,文中通過(guò)示例代碼介紹的非常詳細(xì),對(duì)大家的學(xué)習(xí)或者工作具有一定的參考學(xué)習(xí)價(jià)值,需要的朋友可以參考下2020-09-09SpringBoot整合JWT的實(shí)現(xiàn)示例
JWT是目前比較流行的跨域認(rèn)證解決方案,本文主要介紹了SpringBoot整合JWT的實(shí)現(xiàn)示例,文中通過(guò)示例代碼介紹的非常詳細(xì),具有一定的參考價(jià)值,感興趣的小伙伴們可以參考一下2022-01-01