三节点kafka集群配置

这篇具有很好参考价值的文章主要介绍了三节点kafka集群配置。希望对大家有所帮助。如果存在错误或未考虑完全的地方,请大家不吝赐教,您也可以点击"举报违法"按钮提交疑问。

部署zookeeper

1. 创建数据目录

mkdir -p /data/zookeeper

2. 三台服务器创建myid文件,分别写入1 2 3

echo "1" > /data/zookeeper/myid
echo "2" > /data/zookeeper/myid
echo "3" > /data/zookeeper/myid

3. 配置hosts

vim /etc/hosts

192.168.44.180 node1
192.168.44.182 node2
192.168.44.184 node3

4. 配置zookeeper配置文件

cp zoo_sample.cfg zoo.cfg

vim zoo.cfg

tickTime=2000
dataDir=/data/zookeeper/
clientPort=2181
initLimit=5
syncLimit=2
server.1=node1:2888:3888
server.2=node2:2888:3888
server.3=node3:2888:3888

5. 启动zookeeper

 bin/zkServer.sh start

查看日志kafka1 kafka2 kafka3 是服务器的主机名

tail -200f logs/zookeeper-root-server-kafka1.out
tail -200f logs/zookeeper-root-server-kafka2.out
tail -200f logs/zookeeper-root-server-kafka3.out

6. 查看zookeeper状态

查看第一台

#依次启动时,第一台会先提示失败,启动第二天成功后,再次查看第一天会提示成功
[root@kafka1 bin]# ./zkServer.sh status
/usr/bin/java
ZooKeeper JMX enabled by default
Using config: /root/apache-zookeeper-3.7.1-bin/bin/../conf/zoo.cfg
Client port found: 2181. Client address: localhost. Client SSL: false.
Error contacting service. It is probably not running.

#启动第二天成功后再次查看
[root@kafka1 bin]# ./zkServer.sh status
/usr/bin/java
ZooKeeper JMX enabled by default
Using config: /root/apache-zookeeper-3.7.1-bin/bin/../conf/zoo.cfg
Client port found: 2181. Client address: localhost. Client SSL: false.
Mode: follower

查看第二台

[root@kafka2 bin]# ./zkServer.sh status
/usr/bin/java
ZooKeeper JMX enabled by default
Using config: /root/apache-zookeeper-3.7.1-bin/bin/../conf/zoo.cfg
Client port found: 2181. Client address: localhost. Client SSL: false.
Mode: leader

查看第三台

[root@kafka3 bin]# ./zkServer.sh status
ZooKeeper JMX enabled by default
Using config: /root/apache-zookeeper-3.7.1-bin/bin/../conf/zoo.cfg
Client port found: 2181. Client address: localhost. Client SSL: false.
Mode: follower

部署kafka

1. 创建数据目录

mkdir -p /data/kafka

2. hosts文件配置zookeeper时已配置

3.配置kafka配置文件

服务器1

 #当前服务在集群的唯一标识
 broker.id=0

#当前服务的ip和端口
 listeners=PLAINTEXT://192.168.44.180:9092

#数据存储目录 
log.dirs=/data/kafka/kafka-logs

#zookeeper地址
zookeeper.connect=node1:2181,node2:2181,node3:2181

服务器2

#当前服务在集群的唯一标识
broker.id=1

#当前服务的ip和端口
listeners=PLAINTEXT://192.168.44.182:9092

#数据存储目录 
log.dirs=/data/kafka/kafka-logs

#zookeeper地址
zookeeper.connect=node1:2181,node2:2181,node3:2181

服务器3

 #当前服务在集群的唯一标识
 broker.id=2

#当前服务的ip和端口
 listeners=PLAINTEXT://192.168.44.184:9092

#数据存储目录 
log.dirs=/data/kafka/kafka-logs

#zookeeper地址
zookeeper.connect=node1:2181,node2:2181,node3:2181

4. 启动kafka

bin/kafka-server-start.sh -daemon config/server.properties

查看kafka命令脚本

cd bin
ll
-rwxr-xr-x 1 root root  1423 21 02:02 connect-distributed.sh
-rwxr-xr-x 1 root root  1396 21 02:02 connect-mirror-maker.sh
-rwxr-xr-x 1 root root  1420 21 02:02 connect-standalone.sh
-rwxr-xr-x 1 root root   861 21 02:02 kafka-acls.sh
-rwxr-xr-x 1 root root   873 21 02:02 kafka-broker-api-versions.sh
-rwxr-xr-x 1 root root   860 21 02:02 kafka-cluster.sh
-rwxr-xr-x 1 root root   864 21 02:02 kafka-configs.sh
-rwxr-xr-x 1 root root   945 21 02:02 kafka-console-consumer.sh
-rwxr-xr-x 1 root root   944 21 02:02 kafka-console-producer.sh
-rwxr-xr-x 1 root root   871 21 02:02 kafka-consumer-groups.sh
-rwxr-xr-x 1 root root   948 21 02:02 kafka-consumer-perf-test.sh
-rwxr-xr-x 1 root root   871 21 02:02 kafka-delegation-tokens.sh
-rwxr-xr-x 1 root root   869 21 02:02 kafka-delete-records.sh
-rwxr-xr-x 1 root root   866 21 02:02 kafka-dump-log.sh
-rwxr-xr-x 1 root root   863 21 02:02 kafka-features.sh
-rwxr-xr-x 1 root root   865 21 02:02 kafka-get-offsets.sh
-rwxr-xr-x 1 root root   870 21 02:02 kafka-leader-election.sh
-rwxr-xr-x 1 root root   863 21 02:02 kafka-log-dirs.sh
-rwxr-xr-x 1 root root   881 21 02:02 kafka-metadata-quorum.sh
-rwxr-xr-x 1 root root   873 21 02:02 kafka-metadata-shell.sh
-rwxr-xr-x 1 root root   862 21 02:02 kafka-mirror-maker.sh
-rwxr-xr-x 1 root root   959 21 02:02 kafka-producer-perf-test.sh
-rwxr-xr-x 1 root root   874 21 02:02 kafka-reassign-partitions.sh
-rwxr-xr-x 1 root root   874 21 02:02 kafka-replica-verification.sh
-rwxr-xr-x 1 root root 10884 21 02:02 kafka-run-class.sh
-rwxr-xr-x 1 root root  1376 21 02:02 kafka-server-start.sh
-rwxr-xr-x 1 root root  1361 21 02:02 kafka-server-stop.sh
-rwxr-xr-x 1 root root   860 21 02:02 kafka-storage.sh
-rwxr-xr-x 1 root root   945 21 02:02 kafka-streams-application-reset.sh
-rwxr-xr-x 1 root root   863 21 02:02 kafka-topics.sh
-rwxr-xr-x 1 root root   879 21 02:02 kafka-transactions.sh
-rwxr-xr-x 1 root root   958 21 02:02 kafka-verifiable-consumer.sh
-rwxr-xr-x 1 root root   958 21 02:02 kafka-verifiable-producer.sh
-rwxr-xr-x 1 root root  1714 21 02:02 trogdor.sh
drwxr-xr-x 2 root root  4096 21 02:02 windows
-rwxr-xr-x 1 root root   867 21 02:02 zookeeper-security-migration.sh
-rwxr-xr-x 1 root root  1393 21 02:02 zookeeper-server-start.sh
-rwxr-xr-x 1 root root  1366 21 02:02 zookeeper-server-stop.sh
-rwxr-xr-x 1 root root  1019 21 02:02 zookeeper-shell.sh

5. 操作topic

#操作topic的脚本
kafka-topics.sh

[root@kafka1 bin]# ./kafka-topics.sh --help
This tool helps to create, delete, describe, or change a topic.
Option                                   Description
------                                   -----------
--alter                                  Alter the number of partitions,
                                           replica assignment, and/or
                                           configuration for the topic.
--at-min-isr-partitions                  if set when describing topics, only
                                           show partitions whose isr count is
                                           equal to the configured minimum.
--bootstrap-server <String: server to    REQUIRED: The Kafka server to connect
  connect to>                              to.
--command-config <String: command        Property file containing configs to be
  config property file>                    passed to Admin Client. This is used
                                           only with --bootstrap-server option
                                           for describing and altering broker
                                           configs.
--config <String: name=value>            A topic configuration override for the
                                           topic being created or altered. The
                                           following is a list of valid
                                           configurations:
                                                cleanup.policy
                                                compression.type
                                                delete.retention.ms
                                                file.delete.delay.ms
                                                flush.messages
                                                flush.ms
                                                follower.replication.throttled.
                                           replicas
                                                index.interval.bytes
                                                leader.replication.throttled.replicas
                                                local.retention.bytes
                                                local.retention.ms
                                                max.compaction.lag.ms
                                                max.message.bytes
                                                message.downconversion.enable
                                                message.format.version
                                                message.timestamp.difference.max.ms
                                                message.timestamp.type
                                                min.cleanable.dirty.ratio
                                                min.compaction.lag.ms
                                                min.insync.replicas
                                                preallocate
                                                remote.storage.enable
                                                retention.bytes
                                                retention.ms
                                                segment.bytes
                                                segment.index.bytes
                                                segment.jitter.ms
                                                segment.ms
                                                unclean.leader.election.enable
                                         See the Kafka documentation for full
                                           details on the topic configs. It is
                                           supported only in combination with --
                                           create if --bootstrap-server option
                                           is used (the kafka-configs CLI
                                           supports altering topic configs with
                                           a --bootstrap-server option).
--create                                 Create a new topic.
--delete                                 Delete a topic
--delete-config <String: name>           A topic configuration override to be
                                           removed for an existing topic (see
                                           the list of configurations under the
                                           --config option). Not supported with
                                           the --bootstrap-server option.
--describe                               List details for the given topics.
--disable-rack-aware                     Disable rack aware replica assignment
--exclude-internal                       exclude internal topics when running
                                           list or describe command. The
                                           internal topics will be listed by
                                           default
--help                                   Print usage information.
--if-exists                              if set when altering or deleting or
                                           describing topics, the action will
                                           only execute if the topic exists.
--if-not-exists                          if set when creating topics, the
                                           action will only execute if the
                                           topic does not already exist.
--list                                   List all available topics.
--partitions <Integer: # of partitions>  The number of partitions for the topic
                                           being created or altered (WARNING:
                                           If partitions are increased for a
                                           topic that has a key, the partition
                                           logic or ordering of the messages
                                           will be affected). If not supplied
                                           for create, defaults to the cluster
                                           default.
--replica-assignment <String:            A list of manual partition-to-broker
  broker_id_for_part1_replica1 :           assignments for the topic being
  broker_id_for_part1_replica2 ,           created or altered.
  broker_id_for_part2_replica1 :
  broker_id_for_part2_replica2 , ...>
--replication-factor <Integer:           The replication factor for each
  replication factor>                      partition in the topic being
                                           created. If not supplied, defaults
                                           to the cluster default.
--topic <String: topic>                  The topic to create, alter, describe
                                           or delete. It also accepts a regular
                                           expression, except for --create
                                           option. Put topic name in double
                                           quotes and use the '\' prefix to
                                           escape regular expression symbols; e.
                                           g. "test\.topic".
--topic-id <String: topic-id>            The topic-id to describe.This is used
                                           only with --bootstrap-server option
                                           for describing topics.
--topics-with-overrides                  if set when describing topics, only
                                           show topics that have overridden
                                           configs
--unavailable-partitions                 if set when describing topics, only
                                           show partitions whose leader is not
                                           available
--under-min-isr-partitions               if set when describing topics, only
                                           show partitions whose isr count is
                                           less than the configured minimum.
--under-replicated-partitions            if set when describing topics, only
                                           show under replicated partitions
--version                                Display Kafka version.

操作topic

#创建主题,名字叫做firstTopic ,分区1,副本3
[root@kafka1 bin]# ./kafka-topics.sh --bootstrap-server node1:9092 --topic firstTopic --create --partitions 1 --replication-factor 3
Created topic firstTopic.

#查看全部主题,此处两个是因为多创建了一个
[root@kafka1 bin]# ./kafka-topics.sh --bootstrap-server node1:9092 --list
firstTopic
first_topic

#查看主题详细信息
#topic名字 firstTopic
#PartitionCount: 1  分区数 1
#ReplicationFactor: 3 副本数 3
#Partition: 0    数据在0号分区,分区即把大文件切分成多份
#Leader: 0       数据主备,leader为主,分别存放集群不同服务器上。此处leader存放在broker.id=0的服务器上
#Replicas: 0,2,1 Isr: 0,2,1 副本存放在集群broker.id=0,1,2的主机,Isr显示集群运行正常的节点信息,节点运行不正常时会被从Isr中踢出
[root@kafka1 bin]# ./kafka-topics.sh --bootstrap-server node1:9092 --topic firstTopic --describe
Topic: firstTopic       TopicId: HqizKYTWQ06Ed1_IF4B3Jw PartitionCount: 1       ReplicationFactor: 3    Configs:
        Topic: firstTopic       Partition: 0    Leader: 0       Replicas: 0,2,1 Isr: 0,2,1

#修改主题分区为3,分区数只能增加不能减少
[root@kafka1 bin]# ./kafka-topics.sh --bootstrap-server node1:9092 --topic firstTopic --alter --partitions 3

#再次查看主题详细信息
[root@kafka1 bin]# ./kafka-topics.sh --bootstrap-server node1:9092 --topic firstTopic --describe
Topic: firstTopic       TopicId: HqizKYTWQ06Ed1_IF4B3Jw PartitionCount: 3       ReplicationFactor: 3    Configs:
        Topic: firstTopic       Partition: 0    Leader: 0       Replicas: 0,2,1 Isr: 0,2,1
        Topic: firstTopic       Partition: 1    Leader: 1       Replicas: 1,2,0 Isr: 1,2,0
        Topic: firstTopic       Partition: 2    Leader: 2       Replicas: 2,0,1 Isr: 2,0,1

6. 操作生产者

#生产者使用的脚本
kafka-console-producer.sh

[root@kafka1 bin]# ./kafka-console-producer.sh
Missing required option(s) [bootstrap-server]
Option                                   Description
------                                   -----------
--batch-size <Integer: size>             Number of messages to send in a single
                                           batch if they are not being sent
                                           synchronously. please note that this
                                           option will be replaced if max-
                                           partition-memory-bytes is also set
                                           (default: 16384)
--bootstrap-server <String: server to    REQUIRED unless --broker-list
  connect to>                              (deprecated) is specified. The server
                                           (s) to connect to. The broker list
                                           string in the form HOST1:PORT1,HOST2:
                                           PORT2.
--broker-list <String: broker-list>      DEPRECATED, use --bootstrap-server
                                           instead; ignored if --bootstrap-
                                           server is specified.  The broker
                                           list string in the form HOST1:PORT1,
                                           HOST2:PORT2.
--compression-codec [String:             The compression codec: either 'none',
  compression-codec]                       'gzip', 'snappy', 'lz4', or 'zstd'.
                                           If specified without value, then it
                                           defaults to 'gzip'
--help                                   Print usage information.
--line-reader <String: reader_class>     The class name of the class to use for
                                           reading lines from standard in. By
                                           default each line is read as a
                                           separate message. (default: kafka.
                                           tools.
                                           ConsoleProducer$LineMessageReader)
--max-block-ms <Long: max block on       The max time that the producer will
  send>                                    block for during a send request.
                                           (default: 60000)
--max-memory-bytes <Long: total memory   The total memory used by the producer
  in bytes>                                to buffer records waiting to be sent
                                           to the server. This is the option to
                                           control `buffer.memory` in producer
                                           configs. (default: 33554432)
--max-partition-memory-bytes <Integer:   The buffer size allocated for a
  memory in bytes per partition>           partition. When records are received
                                           which are smaller than this size the
                                           producer will attempt to
                                           optimistically group them together
                                           until this size is reached. This is
                                           the option to control `batch.size`
                                           in producer configs. (default: 16384)
--message-send-max-retries <Integer>     Brokers can fail receiving the message
                                           for multiple reasons, and being
                                           unavailable transiently is just one
                                           of them. This property specifies the
                                           number of retries before the
                                           producer give up and drop this
                                           message. This is the option to
                                           control `retries` in producer
                                           configs. (default: 3)
--metadata-expiry-ms <Long: metadata     The period of time in milliseconds
  expiration interval>                     after which we force a refresh of
                                           metadata even if we haven't seen any
                                           leadership changes. This is the
                                           option to control `metadata.max.age.
                                           ms` in producer configs. (default:
                                           300000)
--producer-property <String:             A mechanism to pass user-defined
  producer_prop>                           properties in the form key=value to
                                           the producer.
--producer.config <String: config file>  Producer config properties file. Note
                                           that [producer-property] takes
                                           precedence over this config.
--property <String: prop>                A mechanism to pass user-defined
                                           properties in the form key=value to
                                           the message reader. This allows
                                           custom configuration for a user-
                                           defined message reader.
                                         Default properties include:
                                          parse.key=false
                                          parse.headers=false
                                          ignore.error=false
                                          key.separator=\t
                                          headers.delimiter=\t
                                          headers.separator=,
                                          headers.key.separator=:
                                          null.marker=   When set, any fields
                                           (key, value and headers) equal to
                                           this will be replaced by null
                                         Default parsing pattern when:
                                          parse.headers=true and parse.key=true:
                                           "h1:v1,h2:v2...\tkey\tvalue"
                                          parse.key=true:
                                           "key\tvalue"
                                          parse.headers=true:
                                           "h1:v1,h2:v2...\tvalue"
--reader-config <String: config file>    Config properties file for the message
                                           reader. Note that [property] takes
                                           precedence over this config.
--request-required-acks <String:         The required `acks` of the producer
  request required acks>                   requests (default: -1)
--request-timeout-ms <Integer: request   The ack timeout of the producer
  timeout ms>                              requests. Value must be non-negative
                                           and non-zero. (default: 1500)
--retry-backoff-ms <Long>                Before each retry, the producer
                                           refreshes the metadata of relevant
                                           topics. Since leader election takes
                                           a bit of time, this property
                                           specifies the amount of time that
                                           the producer waits before refreshing
                                           the metadata. This is the option to
                                           control `retry.backoff.ms` in
                                           producer configs. (default: 100)
--socket-buffer-size <Integer: size>     The size of the tcp RECV size. This is
                                           the option to control `send.buffer.
                                           bytes` in producer configs.
                                           (default: 102400)
--sync                                   If set message send requests to the
                                           brokers are synchronously, one at a
                                           time as they arrive.
--timeout <Long: timeout_ms>             If set and the producer is running in
                                           asynchronous mode, this gives the
                                           maximum amount of time a message
                                           will queue awaiting sufficient batch
                                           size. The value is given in ms. This
                                           is the option to control `linger.ms`
                                           in producer configs. (default: 1000)
--topic <String: topic>                  REQUIRED: The topic id to produce
                                           messages to.
--version                                Display Kafka version.

创建生产者

bin/kafka-console-producer.sh --bootstrap-server node1:9092 --topic firstTopic

7. 操作消费者

创建消费者文章来源地址https://www.toymoban.com/news/detail-435387.html

bin/kafka-console-consumer.sh --bootstrap-server node1:9092 --topic firstTopic

#消费者默认从最新的数据开始消费,可以指定从头开始消费
bin/kafka-console-consumer.sh --bootstrap-server node1:9092 --topic firstTopic --from-beginning

到了这里,关于三节点kafka集群配置的文章就介绍完了。如果您还想了解更多内容,请在右上角搜索TOY模板网以前的文章或继续浏览下面的相关文章,希望大家以后多多支持TOY模板网!

本文来自互联网用户投稿,该文观点仅代表作者本人,不代表本站立场。本站仅提供信息存储空间服务,不拥有所有权,不承担相关法律责任。如若转载,请注明出处: 如若内容造成侵权/违法违规/事实不符,请点击违法举报进行投诉反馈,一经查实,立即删除!

领支付宝红包 赞助服务器费用

相关文章

  • zookeeper集群和kafka的相关概念就部署

    目录 一、Zookeeper概述 1、Zookeeper 定义  2、Zookeeper 工作机制 3、Zookeeper 特点  4、Zookeeper 数据结构  5、Zookeeper 应用场景  (1)统一命名服务  (2)统一配置管理  (3)统一集群管理  (4)服务器动态上下线  (5)软负载均衡  6、Zookeeper选举机制 (1)第一次启动选举机制

    2024年02月14日
    浏览(33)
  • Zookeeper、Kafka集群与Filebeat+Kafka+ELK架构、部署实例

    Zookeeper是一个开源的分布式的,为分布式框架提供协调服务的Apache项目。 Zookeeper:一个领导者(Leader),多个跟随者(Follower)组成的集群。 Zookeeper集群中只要有半数以上节点存活,Zookeeper集群就能正常服务。所以Zookeeper适合安装奇数台服务器。 全局数据一致:每个Server保

    2024年02月08日
    浏览(50)
  • ZooKeeper完美安装指南:在CentOS中从单节点到集群部署全攻略

      ZooKeeper是一个集中式服务,用于维护配置信息、命名、提供分布式同步和提供群组服务。其高效且稳定的特性使得它成为分布式环境中的不可或缺的组件。本文将深入探讨如何在CentOS 系统上安装和配置最新版本的ZooKeeper,涵盖从单节点到集群的全面部署。 系统版本:

    2024年02月06日
    浏览(41)
  • zookeeper+kafka分布式消息队列集群的部署

    目录 一、zookeeper 1.Zookeeper 定义 2.Zookeeper 工作机制 3.Zookeeper 特点 4.Zookeeper 数据结构 5.Zookeeper 应用场景 (1)统一命名服务 (2)统一配置管理 (3)统一集群管理 (4)服务器动态上下线 6.Zookeeper 选举机制 (1)第一次启动选举机制 (2)非第一次启动选举机制 7.部署zookeepe

    2024年02月14日
    浏览(52)
  • Helm方式部署 zookeeper+kafka 集群 ——2023.05

    服务 版本 centos 7.9 kubernetes v1.20.15 helm v3.10.1 zookeeper 3.8.1 kafka 3.4.0 安装方式有两种,在线安装和离线安装,在线安装方便快捷,但是无法修改参数。由于需要修改配置,故本文采用离线安装方式。 2.1 在线安装zookeeper+kafka集群 1. 部署zookeeper集群 提示: 由于这个ApacheZookeeper集群

    2024年02月09日
    浏览(52)
  • 三节点kafka集群配置

    1. 创建数据目录 2. 三台服务器创建myid文件,分别写入1 2 3 3. 配置hosts 4. 配置zookeeper配置文件 5. 启动zookeeper 查看日志kafka1 kafka2 kafka3 是服务器的主机名 6. 查看zookeeper状态 查看第一台 查看第二台 查看第三台 1. 创建数据目录 2. hosts文件配置zookeeper时已配置 3.配置kafka配置文件

    2024年02月03日
    浏览(35)
  • k8s部署zookeeper集群(3节点,1个leader,2个follower)

    环境: centos 7.9 k8s集群 在k8s上面安装zookeeper集群,我们还是按照k8s的官方文档来安装吧,这样比较好,网上有各种各样的安装方式,这里使用 https://kubernetes.io/docs/tutorials/stateful-application/zookeeper/ k8s的官方文档来安装。

    2024年02月13日
    浏览(46)
  • kafka 3.5.0集群环境部署(使用自带的zookeeper)

    管理界面包:kafka-manager-1.3.3.7.zip (链接) kafka二进制包:kafka_2.13-3.5.0.tgz (链接) java包:jdk-8u291-linux-x64.tar.gz (链接) 节点IP broker id 10.10.10.10 0 10.10.10.11 1 10.10.10.12 2 解压:tar -zxf jdk-8u291-linux-x64.tar.gz 编辑/etc/profile配置文件,文件末尾添加如下配置: 让配置生效source /etc/profile 查看

    2024年02月07日
    浏览(42)
  • 【Kafka】Zookeeper和Kafka集群的安装和配置

    一、集群环境说明 1. 虚拟机:192.168.223.101/103/105 2. 系统版本:CentOS 7.9 3. JDK版本:11.0.18.0.1 4. Zookeeper版本:3.7.1 5. Kafka版本:2.13-2.8.2 备注:无论是ZK,还是Kafka的安装,都需要用到JDK,上面给出的ZK和Kafka版本,都已经支持JDK11(JDK 11 Supported)。这三者之间的兼容关系,感兴趣

    2024年02月12日
    浏览(56)
  • 【简单认识zookeeper+kafka分布式消息队列集群的部署】

    Zookeeper是一个开源的分布式的,为分布式框架提供协调服务的Apache项目。 Zookeeper从设计模式角度来理解:是一个基于观察者模式设计的分布式服务管理框架,它负责存储和管理大家都关心的数据,然后接受观察者的注册,一旦这些数据的状态发生变化,Zookeeper就将负责通知已

    2024年02月13日
    浏览(42)

觉得文章有用就打赏一下文章作者

支付宝扫一扫打赏

博客赞助

微信扫一扫打赏

请作者喝杯咖啡吧~博客赞助

支付宝扫一扫领取红包,优惠每天领

二维码1

领取红包

二维码2

领红包