`

hadoop 2.x-HDFS HA --Part I: abstraction

 
阅读更多

  below are the outlines of this article ,using by hadoop-2.5.1 :

1.abstraction
 1.1 what is HDFS HA
 1.2 how to 
 1.3 other HAs
2.installation
 2.1 manual failover
 2.2 auto failover
3.conclusion

 

1.abstraction

 1.1 what is HDFS HA

  in hadoop distributed file system ,there are some processes distinguished by roles,e.g. namenode,secondarynode,datanode,backupnode etc.and the most important role is namenode,which mantains the name space service,resouce assignment,heartbeat detection etc.

  before hadoop-2.x,the secondary node only do fs image merge with edits from namenode(nn),to speedup startup of nn.and the backup node is same as it but do in memory.so in fact there is only 'a brain' in the system ,this is very different from some design mode of related db system:master and slave(hot spare or named standby). so if something wrong occurs (eg. too many file causes OOME of nn) the nn will be in deep compressure and slower reponses or even 'play dead' to not reponse any requests.

  so a HA(high availablity) is ,of course,come here:keeps the system running as healthy as possible for any failures(hard ware faults,soft ware bugs etc),when a namenode is failed to reponse or down,the other one will updertake it immediately,yes,this switch is transparent to all clients.so its simple and cost lowly.

  1.2 how to implement

   in distributed coordination system,there is a term "most part" whcih means that all the single-ones are over the average of total.that is if N is an odd number ,then formular math.int((N+1)/2) is the so called 'most part'.so if N is an even number, that means prior formular is not of it.therefor,they always use odd number to construct the coordination system base number.

   of course,hadoop's HA(hdfs/namenode HA) uses this term to generate a coordinated service:journal node(ie edit logs),see figure 1 below:



             figure 1

  u can see ,the # of  nn is not necessary to construct to an odd number BUT journal nodes.when the active nn write to journal nodes if 'most part' nodes are success then this write oper is success ;esle failure.

  so the journal nodes consists of 'coordination system' for namenodes.also named as "quorum journal manager".

  then the standby nn will read the edits immediately and maybe combine it with existed fs image for reducing the amount o f edits.so both the active nn and standby one are thinked as 'the same'

  1.3 other HAs

   a.Facebook AvatarNode

   b.Linux HA

   c.ip failover 

 

 

ref:

jira:High Availability Framework for HDFS NN

HDFS High Availability Using the Quorum Journal Manager

hadoop 2.x-HDFS HA --Part II: installation 

  • 大小: 161.5 KB
分享到:
评论

相关推荐

    Hadoop 3.x(HDFS)----【HDFS 的 API 操作】---- 代码

    Hadoop 3.x(HDFS)----【HDFS 的 API 操作】---- 代码 Hadoop 3.x(HDFS)----【HDFS 的 API 操作】---- 代码 Hadoop 3.x(HDFS)----【HDFS 的 API 操作】---- 代码 Hadoop 3.x(HDFS)----【HDFS 的 API 操作】--...

    hadoop最新版本3.1.1全量jar包

    hadoop-auth-3.1.1.jar hadoop-hdfs-3.1.1.jar hadoop-mapreduce-client-hs-3.1.1.jar hadoop-yarn-client-3.1.1.jar hadoop-client-api-3.1.1.jar hadoop-hdfs-client-3.1.1.jar hadoop-mapreduce-client-jobclient...

    hadoop-hdfs-client-2.9.1-API文档-中文版.zip

    赠送jar包:hadoop-hdfs-client-2.9.1.jar 赠送原API文档:hadoop-hdfs-client-2.9.1-javadoc.jar 赠送源代码:hadoop-hdfs-client-2.9.1-sources.jar 包含翻译后的API文档:hadoop-hdfs-client-2.9.1-javadoc-...

    flink-shaded-hadoop-3-uber-3.1.1.7.1.1.0-565-9.0.jar.tar.gz

    本文将针对"flink-shaded-hadoop-3-uber-3.1.1.7.1.1.0-565-9.0.jar.tar.gz"这一特定压缩包,探讨Flink 1.14.0如何与Hadoop 3.x实现兼容,并深入解析其背后的原理。 Flink 1.14.0是一个强大的流处理引擎,它提供了...

    hadoop-hdfs-2.7.3-API文档-中英对照版.zip

    赠送jar包:hadoop-hdfs-2.7.3.jar; 赠送原API文档:hadoop-hdfs-2.7.3-javadoc.jar; 赠送源代码:hadoop-hdfs-2.7.3-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.7.3.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-client-2.9.1-API文档-中英对照版.zip

    赠送jar包:hadoop-hdfs-client-2.9.1.jar; 赠送原API文档:hadoop-hdfs-client-2.9.1-javadoc.jar; 赠送源代码:hadoop-hdfs-client-2.9.1-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-client-2.9.1.pom;...

    Hadoop 2.X HDFS源码剖析-高清-完整目录-2016年3月

    Hadoop 2.X HDFS源码剖析-高清-完整目录-2016年3月,分享给所有需要的人!

    hadoop-hdfs-2.6.5-API文档-中文版.zip

    赠送jar包:hadoop-hdfs-2.6.5.jar; 赠送原API文档:hadoop-hdfs-2.6.5-javadoc.jar; 赠送源代码:hadoop-hdfs-2.6.5-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.6.5.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-2.5.1-API文档-中文版.zip

    赠送jar包:hadoop-hdfs-2.5.1.jar; 赠送原API文档:hadoop-hdfs-2.5.1-javadoc.jar; 赠送源代码:hadoop-hdfs-2.5.1-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.5.1.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-2.7.3-API文档-中文版.zip

    赠送jar包:hadoop-hdfs-2.7.3.jar; 赠送原API文档:hadoop-hdfs-2.7.3-javadoc.jar; 赠送源代码:hadoop-hdfs-2.7.3-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.7.3.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-2.5.1-API文档-中英对照版.zip

    赠送jar包:hadoop-hdfs-2.5.1.jar; 赠送原API文档:hadoop-hdfs-2.5.1-javadoc.jar; 赠送源代码:hadoop-hdfs-2.5.1-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.5.1.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-2.6.5-API文档-中英对照版.zip

    赠送jar包:hadoop-hdfs-2.6.5.jar; 赠送原API文档:hadoop-hdfs-2.6.5-javadoc.jar; 赠送源代码:hadoop-hdfs-2.6.5-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.6.5.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-2.9.1-API文档-中英对照版.zip

    赠送jar包:hadoop-hdfs-2.9.1.jar; 赠送原API文档:hadoop-hdfs-2.9.1-javadoc.jar; 赠送源代码:hadoop-hdfs-2.9.1-sources.jar; 赠送Maven依赖信息文件:hadoop-hdfs-2.9.1.pom; 包含翻译后的API文档:hadoop...

    hadoop-hdfs-2.9.1-API文档-中文版.zip

    赠送jar包:hadoop-hdfs-2.9.1.jar 赠送原API文档:hadoop-hdfs-2.9.1-javadoc.jar 赠送源代码:hadoop-hdfs-2.9.1-sources.jar 包含翻译后的API文档:hadoop-hdfs-2.9.1-javadoc-API文档-中文(简体)版.zip 对应...

    Hadoop3.2.2资源包+安装文档

    Hadoop 2.x - 对于数据平衡使用HDFS平衡器。 Hadoop 3.x - 对于数据平衡使用Intra-data节点平衡器,该平衡器通过HDFS磁盘平衡器CLI调用。 存储Scheme Hadoop 2.x - 使用3X副本Scheme Hadoop 3.x - 支持HDFS中的擦除...

    Hadoop_2.X_HDFS源码剖析_带索引书签目录_徐鹏

    《Hadoop_2.X_HDFS源码剖析》是由徐鹏编著的一本深入解析Hadoop 2.x版本中HDFS(Hadoop Distributed File System)源码的专业书籍。这本书旨在帮助读者理解HDFS的核心机制,提升在分布式存储系统方面的专业技能。 ...

    hadoop.dll-and-winutils.exe-for-hadoop2.9.0-on-windows_X64

    4. 配置hadoop-env.cmd:打开Hadoop安装目录下的conf子目录,找到hadoop-env.cmd文件,编辑该文件,将`%JAVA_HOME%`替换为你本机Java JDK的安装路径。 5. 初始化HDFS:在命令行中,使用`winutils.exe fs -mkdir /...

    hadoop.dll-winutils.exe-hadoop2.7.x

    标题提到的"hadop.dll-winutils.exe-hadoop2.7.x"指的是针对Hadoop 2.7.2版本的特定解决方法,描述表明了在该环境中使用这两个文件可以消除错误。 `hadoop.dll` 是一个动态链接库文件,主要在Windows环境下为Hadoop...

    hadoop1.x与hadoop2.x配置异同

    在探讨Hadoop1.x与Hadoop2.x配置的异同之前,我们首先简要回顾一下GridGain In-Memory HDFS的特性,这是基于行业首个高性能双模式内存文件系统,完全兼容HDFS。GridGain FileSystem(GGFS)作为Hadoop HDFS的即插即...

    hadoop.dll-and-winutils.exe-for-hadoop2.7.7-on-windows_X64-master

    标题 "hadoop.dll-and-winutils.exe-for-hadoop2.7.7-on-windows_X64-master" 暗示了这是一个针对64位Windows系统优化的Hadoop 2.7.7版本的特定组件集合,主要包含`hadoop.dll`和`winutils.exe`两个关键文件。...

Global site tag (gtag.js) - Google Analytics