当前位置: 首页 > news >正文

服装网站建设费用预算网站建设 开题报告

服装网站建设费用预算,网站建设 开题报告,广州注册个体户流程及费用,那个网站百度收录好简介 Kafka 是一种分布式、支持分区、多副本的消息中间件#xff0c;支持发布-订阅模式#xff0c;多用于实时处理大量数据缓存的场景#xff0c;类似于一个“缓存池”。 架构 Producer#xff1a;消息生产者#xff1b;Consumer#xff1a;消息消费者#xff1b;Brok…简介 Kafka 是一种分布式、支持分区、多副本的消息中间件支持发布-订阅模式多用于实时处理大量数据缓存的场景类似于一个“缓存池”。 架构 Producer消息生产者Consumer消息消费者Broker一台kafka服务器也称作一个brokerkafka集群包含多个brokerTopic一个topic为一个消息队列生产者、消费者基于topic进行发布-订阅Partition消息分区一个topic可以分为多个partition每个partition是一个消息队列Replica副本。一个topic的每个分区都有若干个副本一个Leader和若干个FollowerLeader每个分区多个副本的“主”生产者发送数据的对象以及消费者消费数据的对象都是LeaderFollower每个分区多个副本中的“从”实时从Leader中同步数据保持和Leader数据的同步。Leader发生故障时某个Follower会成为新的Leader 安装部署 下载kafka安装包 # 下载地址,取最新包即可,最新版本为kafka_2.13-3.6.1.tgz https://kafka.apache.org/downloads解压 tar -zxvf kafka_2.13-3.6.1.tgzmv kafka_2.13-3.6.1 kafkacd kafka/修改配置文件 1本次集群配置采用kraft模式集群节点为192.168.1.12~14 2在3个节点解压后进入kafka/config/kraft目录修改server.properties文件 192.168.1.12节点配置文件修改如下 node.id1 controller.quorum.voters1192.168.1.12:9093,1192.168.1.13:9093,3192.168.1.14:9093 advertised.listenersPLAINTEXT://192.168.1.12:9092192.168.1.13节点配置文件修改如下 node.id2 controller.quorum.voters1192.168.1.12:9093,1192.168.1.13:9093,3192.168.1.14:9093 advertised.listenersPLAINTEXT://192.168.1.13:9092192.168.1.14节点配置文件修改如下 node.id3 controller.quorum.voters1192.168.1.12:9093,1192.168.1.13:9093,3192.168.1.14:9093 advertised.listenersPLAINTEXT://192.168.1.14:9092节点说明 # Broker节点 Kafka集群中的数据节点消息队列它们负责接收客户端的消息和传递消息给客户端默认情况下每个Broker节点会监听9092端口该端口用于与客户端进行通信客户端可以将消息发送到这个端口或者从这个端口接收消息这个端口可以称作客户端通信端口。# Controller节点 Kafka集群中的控制器节点负责管理集群的状态和元数据Controller节点监听的端口通常是9093该端口用于集群中其他节点获取元数据或在混合节点选举新的Controller时进行通信通过该端口其他节点可以与Controller节点交互获取集群的元数据信息或参与控制器的选举过程这个端口可以称作控制器端口。# 混合节点 同时担任Broker和Controller角色的节点这两个端口都会被使用默认情况下混合节点将监听9092端口接收和传递消息给客户端并监听9093端口用于与其他节点进行元数据交换和控制器选举通信可见混合节点会同时使用两个端口分别作为客户端通信端口与控制器端口。生成集群id # KRaft模式的kafka集群需要设定一个id,在任意节点执行如下命令: sh /usr/local/incvs-kafka/bin/kafka-storage.sh random-uuid格式化数据目录 # 在集群所有节点执行如下命令: sh /home/kafka/bin/kafka-storage.sh format -t 步骤4)生成的id -c /home/kafka/config/kraft/server.properties启动Kakfa # 在3个节点依次执行如下命令: sh /home/kafka/bin/kafka-server-start.sh /home/kafka/config/kraft/server.properties测试 在192.168.1.13节点启动消费者脚本订阅“MykafkaTopic” sh /home/kafka/bin/kafka-console-consumer.sh --bootstrap-server 192.168.1.13:9092 --topic MykafkaTopic hello kafka在192.168.1.12节点启动生产者脚本往topic生产消息 sh /home/kafka/bin/kafka-console-producer.sh --bootstrap-server 192.168.1.12:9092 --topic MykafkaTopic hello kafka配置文件说明 server.properties配置示例 ############################# Server Basics ############################## The role of this server. Setting this puts us in KRaft mode process.rolesbroker,controller# The node id associated with this instances roles node.id1# The connect string for the controller quorum controller.quorum.voters1localhost:9093############################# Socket Server Settings ############################## The address the socket server listens on. # Combined nodes (i.e. those with process.rolesbroker,controller) must list the controller listener here at a minimum. # If the broker listener is not defined, the default listener will use a host name that is equal to the value of java.net.InetAddress.getCanonicalHostName(), # with PLAINTEXT listener name, and port 9092. # FORMAT: # listeners listener_name://host_name:port # EXAMPLE: # listeners PLAINTEXT://your.host.name:9092 listenersPLAINTEXT://:9092,CONTROLLER://:9093# Name of listener used for communication between brokers. inter.broker.listener.namePLAINTEXT# Listener name, hostname and port the broker will advertise to clients. # If not set, it uses the value for listeners. advertised.listenersPLAINTEXT://localhost:9092# A comma-separated list of the names of the listeners used by the controller. # If no explicit mapping set in listener.security.protocol.map, default will be using PLAINTEXT protocol # This is required if running in KRaft mode. controller.listener.namesCONTROLLER# Maps listener names to security protocols, the default is for them to be the same. See the config documentation for more details listener.security.protocol.mapCONTROLLER:PLAINTEXT,PLAINTEXT:PLAINTEXT,SSL:SSL,SASL_PLAINTEXT:SASL_PLAINTEXT,SASL_SSL:SASL_SSL# The number of threads that the server uses for receiving requests from the network and sending responses to the network num.network.threads3# The number of threads that the server uses for processing requests, which may include disk I/O num.io.threads8# The send buffer (SO_SNDBUF) used by the socket server socket.send.buffer.bytes102400# The receive buffer (SO_RCVBUF) used by the socket server socket.receive.buffer.bytes102400# The maximum size of a request that the socket server will accept (protection against OOM) socket.request.max.bytes104857600############################# Log Basics ############################## A comma separated list of directories under which to store log files log.dirs/tmp/kraft-combined-logs# The default number of log partitions per topic. More partitions allow greater # parallelism for consumption, but this will also result in more files across # the brokers. num.partitions1# The number of threads per data directory to be used for log recovery at startup and flushing at shutdown. # This value is recommended to be increased for installations with data dirs located in RAID array. num.recovery.threads.per.data.dir1############################# Internal Topic Settings ############################# # The replication factor for the group metadata internal topics __consumer_offsets and __transaction_state # For anything other than development testing, a value greater than 1 is recommended to ensure availability such as 3. offsets.topic.replication.factor1 transaction.state.log.replication.factor1 transaction.state.log.min.isr1############################# Log Flush Policy ############################## Messages are immediately written to the filesystem but by default we only fsync() to sync # the OS cache lazily. The following configurations control the flush of data to disk. # There are a few important trade-offs here: # 1. Durability: Unflushed data may be lost if you are not using replication. # 2. Latency: Very large flush intervals may lead to latency spikes when the flush does occur as there will be a lot of data to flush. # 3. Throughput: The flush is generally the most expensive operation, and a small flush interval may lead to excessive seeks. # The settings below allow one to configure the flush policy to flush data after a period of time or # every N messages (or both). This can be done globally and overridden on a per-topic basis.# The number of messages to accept before forcing a flush of data to disk #log.flush.interval.messages10000# The maximum amount of time a message can sit in a log before we force a flush #log.flush.interval.ms1000############################# Log Retention Policy ############################## The following configurations control the disposal of log segments. The policy can # be set to delete segments after a period of time, or after a given size has accumulated. # A segment will be deleted whenever *either* of these criteria are met. Deletion always happens # from the end of the log.# The minimum age of a log file to be eligible for deletion due to age log.retention.hours168# A size-based retention policy for logs. Segments are pruned from the log unless the remaining # segments drop below log.retention.bytes. Functions independently of log.retention.hours. #log.retention.bytes1073741824# The maximum size of a log segment file. When this size is reached a new log segment will be created. log.segment.bytes1073741824# The interval at which log segments are checked to see if they can be deleted according # to the retention policies log.retention.check.interval.ms300000process.roles服务器角色包含broker和controller。如示例中使用kraft模式使用的混合模式即“process.rolesbroker,controller”node.id节点ID一个集群中的node.id不能重复controller.quorum.voters控制选举的投票节点格式为“node-idhost:port”listeners服务器监听的地址格式为“listener_name://host_name:port”inter.broker.listener.name用于broker之间通信的监听器名称advertised.listeners服务器向客户端宣告的监听器名称、主机名和端口controller.listener.names控制器使用的监听器名称列表listener.security.protocol.map监听器名称到安全协议的映射。默认情况下它们是相同的num.network.threads服务器用于从网络接收请求和向网络发送响应的线程数num.io.threads服务器用于处理请求可能包括磁盘 I/O的线程数socket.send.buffer.bytes服务器用于发送数据的缓冲区大小socket.receive.buffer.bytes服务器用于接收数据的缓冲区大小socket.request.max.bytes服务器接受的请求的最大大小用于防止内存溢出log.dirs用于存储日志文件的目录列表重要需清楚日志存储的原理num.partitions每个主题的默认日志分区数num.recovery.threads.per.data.dir每个数据目录在启动时用于日志恢复和关闭时用于刷新的线程数offsets.topic.replication.factor内部主题 “__consumer_offsets” 和 “__transaction_state” 的复制因子transaction.state.log.replication.factor事务状态日志的复制因子transaction.state.log.min.isr事务状态日志的最小同步副本数log.flush.interval.messages强制将数据刷新到磁盘之前接受的消息数log.flush.interval.ms消息在日志中停留的最大时间超过这个时间就会强制刷新到磁盘log.retention.hours由于年龄而使日志文件有资格被删除的最小年龄log.retention.bytes基于大小的日志保留策略默认1Glog.segment.bytes日志段文件的最大大小log.retention.check.interval.ms检查日志段是否可以根据保留策略被删除的间隔
http://www.zqtcl.cn/news/720400/

相关文章:

  • 阿里云备案网站 网站名称怎么写京icp备案查询
  • 网站开发岗位思维导图alexa排名
  • 自适应网站建设济南济南网站建设公司
  • 巴州网站建设库尔勒网站建设钟爱网络杭州微信网站制作
  • 52做网站南京市住房城乡建设门户网站
  • 网站开发精品课程贵阳市白云区官方网站
  • seo整站优化服务会计培训班一般收费多少
  • 批量网站访问检测怎么做好手机网站开发
  • 深圳网站建设公司哪家比较好shortcodes wordpress
  • 网站内链越多越好嘛可以做3d电影网站
  • 企业网站需求文档微商引流客源最快的方法
  • 交互式网站备案业务网站在线生成
  • 自建网站百度个人网站如何在百度上做推广
  • 如何安装wordpress模板竞价网站做seo
  • 做论坛网站如何赚钱电子商务营销推广
  • 想要自己做一个网站怎么做济宁百度网站建设
  • 海会网络建设网站wordpress刷不出图片
  • 一个人做商城网站网站推广的几个阶段
  • 做国学类网站合法吗html5教程pdf下载
  • 云南省文化馆网站建设二级域名分发平台
  • 网站版面布局结构图网站收录批量查询
  • 网站开发手机模拟器常州到丹阳
  • 淮南医院网站建设班级网站开发报告
  • 东莞营销网站建设哪家好微信api接口
  • 凡科建站怎么导出网页wordpress视频采集插件
  • 个人介绍网站源码云主机上传网站
  • app推广平台网站系统登录入口
  • 做公司宣传册的网站成crm网
  • 新乡公司做网站军事新闻内容摘抄
  • 讯美智能网站建设泰安网络科技有限公司电话