我们都知道, Kafka的每个Topic的存储在逻辑上分成若干个Partition,每个Partition又可以设置自己的副本Replica;
这样的设计就引出了几个概念:
Partition: 消息在Kafka上存储的最小逻辑单元, 在物理上对应在不同的Broker机器上;
Replica: 每个Partition可以设置自己的副本Partition, 这样主Partition叫作
Leader
, 副本叫作Replica
;从灾备的角度考虑, 在物理上Replica尽量不要与Leader在同一台Broker物理机上;Ack: 客户端produce消息时, 可以设置Kafka服务端回应ack的策略:
3.1 不用回Ack, 客户端发送效率最高, 但无法确认是否真的发送成功;
3.2 仅Partition leader回ack, 发送效率次之, 可以确认Leader已经接收到消息;在这种情况下,如果leader挂了, 客户端将无法消费到这个消息;
3.3 所有Replica(实际上这不是真的)都需要回ack, 发送效率最差, Replica需要从Leader拉取消息;ISR:
In Sync Replica
, 是所有Replica的一个子集. Partition的replica可能很多, 针对上面的3.3,如果需要所有replicat都拉取到消息后再回ack,发送效率会很差,因此Kafka用了折衷的办法, 仅需要ISR中的replica接收了消息即可.ISR中的replica的消息应一直与leader同步;
既然有Leader
的角色,又有多个replica, 就存在一个在选主的问题, 我们就来讲下多种情况下的选主策略;
PartitionLeaderSelector
所在文件: core/src/main/scala/kafka/controller/PartitionLeaderSelector.scala
这个
trait
, 各种选主策略类都实现了它.声明了如下的方法, 返回LeaderAndIsr
类型的request
/** * @param topicAndPartition The topic and partition whose leader needs to be elected * @param currentLeaderAndIsr The current leader and isr of input partition read from zookeeper * @throws NoReplicaOnlineException If no replica in the assigned replicas list is alive * @return The leader and isr request, with the newly selected leader and isr, and the set of replicas to receive * the LeaderAndIsrRequest. */ def selectLeader(topicAndPartition: TopicAndPartition, currentLeaderAndIsr: LeaderAndIsr): (LeaderAndIsr, Seq[Int])
OfflinePartitionLeaderSelector
所在 core/src/main/scala/kafka/controller/PartitionLeaderSelector.scala
可用于Offline状态Partitions的选主,比如Topic刚刚创建后;
规则, 源码中的注释
Select the new leader, new isr and receiving replicas (for the LeaderAndIsrRequest):
If at least one broker from the isr is alive, it picks a broker from the live isr as the new leader and the live isr as the new isr.
Else, if unclean leader election for the topic is disabled, it throws a NoReplicaOnlineException.
Else, it picks some alive broker from the assigned replica list as the new leader and the new isr.
If no broker in the assigned replica list is alive, it throws a NoReplicaOnlineException
Replicas to receive LeaderAndIsr request = live assigned replicasOnce the leader is successfully registered in zookeeper, it updates the allLeaders cache
翻译成图:
PartitionLeaderSelector.png
ReassignedPartitionLeaderSelector
所在 core/src/main/scala/kafka/controller/PartitionLeaderSelector.scala
用于Partitions再分配后的LeaderSelect;
规则:
New leader = a live in-sync reassigned replica
New isr = current isr
Replicas to receive LeaderAndIsr request = reassigned replicas
ControlledShutdownLeaderSelector
所在 core/src/main/scala/kafka/controller/PartitionLeaderSelector.scala
用于ControllerShutdown时的leader select
规则:
New leader = replica in isr that's not being shutdown;
New isr = current isr - shutdown replica;
Replicas to receive LeaderAndIsr request = live assigned replicas
NoOpLeaderSelector
所在 core/src/main/scala/kafka/controller/PartitionLeaderSelector.scala
其实什么都不作,返回当前的Leader, ISR和Replicas;
def selectLeader(topicAndPartition: TopicAndPartition, currentLeaderAndIsr: LeaderAndIsr): (LeaderAndIsr, Seq[Int]) = { warn("I should never have been asked to perform leader election, returning the current LeaderAndIsr and replica assignment.") (currentLeaderAndIsr, controllerContext.partitionReplicaAssignment(topicAndPartition)) }
作者:扫帚的影子
链接:https://www.jianshu.com/p/505fa1f9b61a
共同学习,写下你的评论
评论加载中...
作者其他优质文章