Hadoop(HA)分布式集群部署和单节点namenode部署其实一样,只是配置文件的不同罢了。

这篇就讲解hadoop双namenode的部署,实现高可用.

系统环境:

OS: CentOS 6.8
内存:2G
CPU:1核
Software:jdk-8u151-linux-x64.rpm
hadoop-2.7.4.tar.gz
zookeeper-3.3.6.tar.gz

下载地址:

wget http://mirrors.hust.edu.cn/apache/hadoop/common/hadoop-2.7.4/hadoop-2.7.4.tar.gz
wget http://mirror.bit.edu.cn/apache/zookeeper/zookeeper-3.3.6/zookeeper-3.3.6.tar.gz

主机列表信息:

主机名

IP 地址

安装软件

角色

运行服务

hadoop-01

192.168.153.128

Jdk,hadoop

NameNode

NameNode,ResourceManager,DFSZKFailoverController,QuorumPeerMain,JournalNode,JobHistoryServer

hadoop-02

192.168.153.129

Jdk,hadoop

NameNode、DataNode

NameNode,ResourceManager,DFSZKFailoverController,QuorumPeerMain,JournalNode

hadoop-03

192.168.153.130

Jdk,Hadoop

zookeep

DataNode

DataNode,NodeManager,QuorumPeerMain, mapred,QuorumPeerMain

1.Hosts文件设置(三台主机的host文件需要保持一致)

[root@hadoop- ~]# cat /etc/hosts
127.0.0.1 localhost localhost.localdomain localhost4 localhost4.localdomain4
:: localhost localhost.localdomain localhost6 localhost6.localdomain6
#######################################
192.168.153.128 hadoop-
192.168.153.129 hadoop-
192.168.153.130 hadoop-

2.创建hadoop账户,使用该账户运行hadoop服务(三台主机都创建hadoop用户).

创建hadoop用户,设置拥有sudo权限
[root@hadoop-01 ~]# useradd hadoop && echo hadoop | passwd --stdin hadoop
Changing password for user hadoop.
passwd: all authentication tokens updated successfully.
[root@hadoop-01 ~]# echo "hadoop ALL=(ALL) NOPASSWD:ALL" >> /etc/sudoers

3.生成SSH免秘钥认证文件(在三台主机都执行)

[root@hadoop-01 ~]# su - hadoop
[hadoop@hadoop-01 ~]$ ssh-keygen -t rsa

4.SSH密钥认证文件分发

1.>将hadoop-01的公钥信息复制到另外两台机器上面
[hadoop@hadoop-01 ~]$ ssh-copy-id 192.168.153.128
[hadoop@hadoop-01 ~]$ ssh-copy-id 192.168.153.129
[hadoop@hadoop-01 ~]$ ssh-copy-id 192.168.153.130 2.>分发hadoop@hadoop-02的公钥到其他两台主机
hadoop@hadoop-02 ~]$ ssh-copy-id 192.168.153.128
[hadoop@hadoop-02 ~]$ ssh-copy-id 192.168.153.129
[hadoop@hadoop-02 ~]$ ssh-copy-id 192.168.153.130 3.>分发hadoop@hadoop-03的公钥到其他两台主机
[hadoop@hadoop-03 ~]$ ssh-copy-id 192.168.153.128
[hadoop@hadoop-03 ~]$ ssh-copy-id 192.168.153.129
[hadoop@hadoop-03 ~]$ ssh-copy-id 192.168.153.130
这样三台机器都可以互相免密钥访问, ssh-copy-id会以追加的方式进行密钥的分发

5.安装jdk(三台主机jdk安装)

[root@hadoop-01 ~]# rpm -ivh jdk-8u151-linux-x64.rpm
Preparing... ########################################### [100%]
1:jdk1.8 ########################################### [100%]
[root@hadoop-01 ~]# export JAVA_HOME=/usr/java/jdk1.8.0_151/
[root@hadoop-01 ~]# export PATH=$JAVA_HOME/bin:$PATH
[root@hadoop-01 ~]# java -version
java version "1.8.0_151"
Java(TM) SE Runtime Environment (build 1.8.0_151-b12)
Java HotSpot(TM) 64-Bit Server VM (build 25.151-b12, mixed mode)

6、安装hadoop:

[hadoop@hadoop-01~]$ wget http://mirrors.hust.edu.cn/apache/hadoop/common/hadoop-2.7.4/hadoop-2.7.4-src.tar.gz
[hadoop@hadoop-01 ~]$ sudo tar zxvf hadoop-2.7.4.tar.gz -C /home/hadoop/ && cd /home/hadoop
[hadoop@hadoop-01 ~]$ sudo mv hadoop-2.7.4/ hadoop
[hadoop@hadoop-01 ~]$ sudo chown -R hadoop:hadoop hadoop/
#将hadoop的二进制目录添加到PATH变量,并设置HADOOP_HOME环境变量
[hadoop@hadoop-01 ~]$ export HADOOP_HOME=/home/hadoop/hadoop/
[hadoop@hadoop-01 ~]$ export PATH=$HADOOP_HOME/bin:$PATH
#在安装hadoop中,只需在hadoop-01上面下载和解压即可,配置完成后再同步到其他节点即可.

7、修改hadoop的配置文件

配置文件位置:/home/hadoop/hadoop/etc/hadoop

1.>修改hadoop-env.sh配置文件,指定JAVA_HOME为JAVA的安装路径.

export JAVA_HOME=${JAVA_HOME}
修改为:
export JAVA_HOME=/usr/java/jdk1.8.0_151/

2.> 修改yarn-env.sh文件

指定yran框架的java运行环境,该文件是yarn框架运行环境的配置文件,需要修改JAVA_HOME的位置

# export JAVA_HOME=/home/y/libexec/jdk1.6.0/
修改为:
export JAVA_HOME=/usr/java/jdk1..0_151/

3.> 修改slaves文件

指定DataNode数据存储服务器,将所有的DataNode的机器的主机名写入到此文件中,如下:

[hadoop@hadoop- hadoop]$ cat slaves
hadoop-
hadoop-

4.>修改core-site.xml文件.在<configuration>……</configuration>中添加如下:

<configuration>
<!-- 指定hdfs的nameservice为ns -->
<property>
<name>fs.defaultFS</name>
<value>hdfs://ns</value>
</property>
<!--指定hadoop数据临时存放目录-->
<property>
<name>hadoop.tmp.dir</name>
<value>/home/hadoop/tmp</value>
</property>
<property>
<name>io.file.buffer.size</name>
<value></value>
</property>
<property>
<name>fs.trash.interval</name>
<value></value>
</property>
<!--指定zookeeper地址-->
<property>
<name>ha.zookeeper.quorum</name>
<value>hadoop-:,hadoop-:,hadoop-:</value>
</property>
</configuration>

5.> 修改hdfs-site.xml配置文件,在<configuration>……</configuration>中添加如下:

<configuration>
<!--指定hdfs的nameservice为ns,需要和core-site.xml中的保持一致 -->
<property>
<name>dfs.nameservices</name>
<value>ns</value>
</property>
<!-- ns下面有两个NameNode,分别是nn1,nn2 -->
<property>
<name>dfs.ha.namenodes.ns</name>
<value>nn1,nn2</value>
</property>
<!-- nn1的RPC通信地址 -->
<property>
<name>dfs.namenode.rpc-address.ns.nn1</name>
<value>hadoop-:</value>
</property>
<!-- nn1的http通信地址 -->
<property>
<name>dfs.namenode.http-address.ns.nn1</name>
<value>hadoop-:</value>
</property>
<!-- nn2的RPC通信地址 -->
<property>
<name>dfs.namenode.rpc-address.ns.nn2</name>
<value>hadoop-:</value>
</property>
<!-- nn2的http通信地址 -->
<property>
<name>dfs.namenode.http-address.ns.nn2</name>
<value>hadoop-:</value>
</property>
<!-- 指定NameNode的元数据在JournalNode上的存放位置 -->
<property>
<name>dfs.namenode.shared.edits.dir</name>
<value>qjournal://hadoop-01:8485;hadoop-02:8485;hadoop-03:8485/ns</value>
</property>
<!-- 指定JournalNode在本地磁盘存放数据的位置 -->
<property>
<name>dfs.journalnode.edits.dir</name>
<value>/home/hadoop/journal</value>
</property>
<!-- 开启NameNode故障时自动切换 -->
<property>
<name>dfs.ha.automatic-failover.enabled</name>
<value>true</value>
</property>
<!-- 配置失败自动切换实现方式 -->
<property>
<name>dfs.client.failover.proxy.provider.ns</name>
<value>org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider</value>
</property>
<!--配置:Fencing实现自动切换-->
<property>
<name>dfs.ha.fencing.methods</name>
<value>sshfence</value>
</property>
<!-- 使用隔离机制时需要ssh免登陆 -->
<property>
<name>dfs.ha.fencing.ssh.private-key-files</name>
<value>/home/hadoop/.ssh/id_rsa</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/home/hadoop/dfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/home/hadoop/dfs/data</value>
</property>
<property>
<name>dfs.replication</name>
<value></value>
</property>
<!-- 在NN和DN上开启WebHDFS (REST API)功能,不是必须 -->
<property>
<name>dfs.webhdfs.enabled</name>
<value>true</value>
</property>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>hadoop-:</value>
<description># 通过web界面来查看HDFS状态 </description>
</property>
</configuration>

6.> 修改mapred-site.xml

指定Hadoop的MapReduce运行在YARN环境

[hadoop@hadoop- hadoop]$ cp mapred-site.xml.template mapred-site.xml
[hadoop@hadoop- hadoop]$ vim mapred-site.xml
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>hadoop-:</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>hadoop-:</value>
</property>
</configuration>

7.> 修改yarn-site.xml

#该文件为yarn架构的相关配置

<configuration>
<!-- Site specific YARN configuration properties -->
<!--HA配置-->
<property>
<name>yarn.resourcemanager.ha.enabled</name>
<value>true</value>
</property>
<!--集群名字-->
<property>
<name>yarn.resourcemanager.cluster-id</name>
<value>ns</value>
</property>
<property>
<name>yarn.resourcemanager.ha.rm-ids</name>
<value>rm1,rm2</value>
</property>
<property>
<name>yarn.resourcemanager.ha.id</name>
<value>rm1</value>
</property>
<property>
<name>yarn.resourcemanager.hostname.rm1</name>
<value>hadoop-</value>
</property>
<property>
<name>yarn.resourcemanager.hostname.rm2</name>
<value>hadoop-</value>
</property>
<property>
<name>yarn.resourcemanager.address.rm1</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.address.rm2</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address.rm1</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address.rm2</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address.rm1</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address.rm2</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address.rm1</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address.rm2</name>
<value>hadoop-:</value>
</property>
<property>
<name>yarn.resourcemanager.zk-address</name>
<value>hadoop-:,hadoop-:,hadoop-:</value>
</property>
<!--故障切换-->
<property>
<name>yarn.resourcemanager.recovery.enabled</name>
<value>true</value>
</property>
<!--HA配置结束--> <!--shuffle配置-->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property> <!--资源配置-->
<property>
<name>yarn.nodemanager.resource.memory-mb</name>
<value></value>
</property>
<property>
<name>yarn.nodemanager.vmem-pmem-ratio</name>
<value></value>
</property>
<property>
<name>yarn.scheduler.minimum-allocation-mb</name>
<value></value>
</property>
<property>
<name>yarn.scheduler.maximum-allocation-mb</name>
<value></value>
</property>
<property>
<name>yarn.nodemanager.resource.cpu-vcores</name>
<value></value>
</property>
<property>
<name>yarn.scheduler.minimum-allocation-vcores</name>
<value></value>
</property>
<property>
<name>yarn.scheduler.maximum-allocation-vcores</name>
<value></value>
</property>
</configuration>

8.复制hadoop到其他节点.

[hadoop@hadoop-01 ~]$ scp -r /home/hadoop/hadoop/ 192.168.153.129:/home/hadoop/
[hadoop@hadoop-01 ~]$ scp -r /home/hadoop/hadoop/ 192.168.153.130:/home/hadoop/

zookeeper集群配置:

由于我的hosts和jdk在安装hadoop时已经配置好,此处就不在重复配置

下载zookeeper及修改配置:

[hadoop@hadoop- ~]$cd /opt/
#sudo wget http://mirror.bit.edu.cn/apache/zookeeper/zookeeper-3.3.6/zookeeper-3.3.6.tar.gz
[hadoop@hadoop- ~]$tar -zxvf zookeeper-3.3..tar.gz -C /home/hadoop/
[hadoop@hadoop- ~]$mv zookeeper-3.3./ zookeeper
[hadoop@hadoop- ~]$cd zookeeper/conf
[hadoop@hadoop- ~]$cp zoo_sample.cfg zoo.cfg
[hadoop@hadoop- conf]$ vim zoo.cfg
ickTime=
dataDir=/var/lib/zookeeper/
initLimit=
syncLimit=
clientPort=
server.=hadoop_01::
server.=hadoop_02::
server.=hadoop_03::

远程复制分发安装文件:

上面已经在hadoop-01上配置完成ZooKeeper,现在可以将该配置好的安装文件远程拷贝到集群中的各个结点对应的目录下

[hadoop@hadoop- ~]$cd /home/hadoop/
[hadoop@hadoop- ~]$scp -r zookeeper/ 192.168.153.129:/home/hadoop/
[hadoop@hadoop- ~]$scp -r zookeeper/ 192.168.153.130:/home/hadoop/
[hadoop@hadoop- ~]$cd /home/hadoop/zookeeper/

设置myid文件要在每台机器的dataDir下,新建一个myid文件,里面存放一个数字,用来标识当前主机。

[hadoop@hadoop- ~]$ sudo -p mkdir /var/lib/zookeeper/
[hadoop@hadoop- ~]$ sudo chown -R hadoop.hadoop /var/lib/zookeeper/
[hadoop@hadoop- ~]$ sudo -p mkdir /var/lib/zookeeper/
[hadoop@hadoop- ~]$ sudo chown -R hadoop.hadoop /var/lib/zookeeper/
[hadoop@hadoop- ~]$ sudo -p mkdir /var/lib/zookeeper/
[hadoop@hadoop- ~]$ sudo chown -R hadoop.hadoop /var/lib/zookeeper/
[root@hadoop_01 ~]# echo "" >> /var/lib/zookeeper/myid
[root@hadoop_02 ~]# echo "" >> /var/lib/zookeeper/myid
[root@hadoop_03 ~]# echo "" >> /var/lib/zookeeper/myid

启动ZooKeeper集群:

[hadoop@hadoop- zookeeper]$./zkServer.sh start
[hadoop@hadoop- zookeeper]$./zkServer.sh start
[hadoop@hadoop- zookeeper]$./zkServer.sh start
[hadoop@hadoop- zookeeper]$ jps -ml
org.apache.zookeeper.server.quorum.QuorumPeerMain /home/hadoop/zookeeper/bin/../conf/zoo.cfg
sun.tools.jps.Jps –ml

启动hadoop集群服务:

首次初始化启动命令和之后启动的命令是不同的,首次启动比较复杂,步骤不对的话就会报错,不过之后就好了

首次启动命令:

1.首先启动各个节点的Zookeeper,在各个节点上执行以下命令(我的zookeeper集群在上面已经启动,此步骤可以省略.):

[hadoop@hadoop-01 zookeeper]$./zkServer.sh start

2.格式化formatZK,在某一个namenode节点执行如下命令,创建命名空间(我是在hadoop-01上面操作的).

[hadoop@hadoop- ~]$ /home/hadoop/hadoop/bin/hdfs zkfc –formatZK
[hadoop@hadoop- ~]$ echo $?

验证:

连接hadoop客户端,查看是否存在hadoop-ha节点.

查看节点命令:ls  /
[hadoop@hadoop- ~]$ /home/hadoop/zookeeper/bin/zkCli.sh

3.启动journalnode (负责两个namenode之间的通信和数据同步):

我配置文件中配置了三个节点都是journalnode,所以三个节点都得启动journalnode

[hadoop@hadoop-01 ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start journalnode
starting journalnode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-journalnode-hadoop-01.out
[hadoop@hadoop-01 ~]$ jps
95779 QuorumPeerMain
114294 JournalNode
114588 Jps
[hadoop@hadoop-02 ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start journalnode
starting journalnode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-journalnode-hadoop-02.out
[hadoop@hadoop-02 ~]$ jps
29911 JournalNode
29592 QuorumPeerMain
29962 Jps
[hadoop@hadoop-02 ~]$
[hadoop@hadoop-03 ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start journalnode
starting journalnode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-journalnode-hadoop-03.out
[hadoop@hadoop-03 ~]$ jps
43591 QuorumPeerMain
44827 JournalNode
44877 Jps
[hadoop@hadoop-03 ~]$

4.在主namenode节点格式化namenode和journalnode目录.

[hadoop@hadoop- ~]$ /home/hadoop/hadoop/bin/hdfs namenode -format ns
// :: INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG: host = hadoop-01/192.168.153.128
STARTUP_MSG: args = [-format, ns]
STARTUP_MSG: version = 2.7.4
17/11/24 03:58:50 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at hadoop-01/192.168.153.128
************************************************************/
[hadoop@hadoop- ~]$ echo $? [hadoop@hadoop- ~]$

5.在主namenode节点启动namenode进程

[hadoop@hadoop-01 ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start namenode
starting namenode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-namenode-hadoop-01.out

6.在备namenode节点执行第一行命令,这个是把备namenode节点的目录格式化并把元数据从主namenode节点copy过来,并且这个命令不会把journalnode目录再格式化了!然后用第二个命令启动备namenode进程!

[hadoop@hadoop- ~]$ /home/hadoop/hadoop/bin/hdfs namenode –bootstrapStandby
[hadoop@hadoop- ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start namenode
starting namenode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-namenode-hadoop-.out

7.在两个namenode(Hadoop-01、hadoop-02)节点都执行以下命令.

[hadoop@hadoop- ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start zkfc
[hadoop@hadoop- ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start zkfc
starting zkfc, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-zkfc-hadoop-.out

8.在所有datanode节点都执行以下命令启动datanode.

[hadoop@hadoop- ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start datanode
starting datanode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-datanode-hadoop-.out
[hadoop@hadoop- ~]$ /home/hadoop/hadoop/sbin/hadoop-daemon.sh start datanode
starting datanode, logging to /home/hadoop/hadoop/logs/hadoop-hadoop-datanode-hadoop-.out

9.启动yarn分布式计算框架(ResourceManager:接收客户端任务请求,接收和监控NodeManager(NM)的资源情况汇报,负责资源的分配与调度,启动和监控ApplicationMaster(AM))

[hadoop@hadoop- ~]$ /home/hadoop/hadoop/sbin/start-yarn.sh

10.启动historyserver服务:JobHistoryServer

[hadoop@hadoop- hadoop]$ sbin/mr-jobhistory-daemon.sh start historyserver
starting historyserver, logging to /home/hadoop/hadoop/logs/mapred-hadoop-historyserver-hadoop-.out

三台主机查看分别运行的服务:

Hadoop-01:

Hadoop-02:

Hadoop-03:

13.查看HDFS分布式文件系统状态

[hadoop@hadoop-01 ~]$ /home/hadoop/hadoop/bin/hdfs dfsadmin -report

#查看文件块组成,一个文件由那些块组成

[hadoop@hadoop-01 ~]$ /home/hadoop/hadoop/bin/hdfs fsck / -files -blocks

14. 日常启停命令

sbin/start-dfs.sh
sbin/stop-dfs.sh
或者/sbin/start-all.sh
/sbin/stop-all.sh 这两个包含(start-dfs.sh and start-yarn.sh)的启动和停止.

15. web页面查看hadoop集群状态

查看HDFS状态:http://192.168.153.128:50070

查看Hadoop集群状态: http://192.168.153.128:8088/cluster

验证HDFS - HA主备切换:

.验证Hadoop-01服务器NameNode和Hadoop-02服务器NameNode数据文件是否可以共享:
执行命令:./hdfs dfs -put /etc/hosts /
[hadoop@hadoop- bin]$ ./hdfs dfs -put /etc/hosts /
[hadoop@hadoop- bin]$ ./hdfs dfs -ls -h /
Found items
drwxr-xr-x - hadoop supergroup -- : /home
-rw-r--r-- hadoop supergroup -- : /hosts
[hadoop@hadoop- bin]$
在hadoop-02上面执行查看/etc/hosts文件是否同步过来.
[hadoop@hadoop- bin]$
[hadoop@hadoop- bin]$ ./hdfs dfs -ls -h /
Found items
drwxr-xr-x - hadoop supergroup -- : /home
-rw-r--r-- hadoop supergroup -- : /hosts
[hadoop@hadoop- bin]$
Hadoop-02上面文件存在,说明同步成功.

2.验证nameserver 主备切换.

首先在浏览器分别打开两个节点的namenode状态,其中一个显示active,另一个显示standby:

Hadoop-01访问地址:http://192.168.153.128:50070

Hadoop-02访问地址:http://192.168.153.129:50070

然后在active所在的namenode(Hadoop-01)节点执行jps,杀掉相应的namenode进程

前面standby所对应的namenode(Hadoop-02)变成active.

至此hadoop双namserver节点服务已经部署和测试成功.

参考文档:

    http://blog.csdn.net/trigl/article/details/55101826

    http://blog.csdn.net/yuan_xw/article/details/51175171

Hadoop(HA)分布式集群部署的更多相关文章

  1. hadoop HA分布式集群搭建

    概述 hadoop2中NameNode可以有多个(目前只支持2个).每一个都有相同的职能.一个是active状态的,一个是standby状态的.当集群运行时,只有active状态的NameNode是正 ...

  2. 1、hadoop HA分布式集群搭建

    概述 hadoop2中NameNode可以有多个(目前只支持2个).每一个都有相同的职能.一个是active状态的,一个是standby状态的.当集群运行时,只有active状态的NameNode是正 ...

  3. Ambari安装之部署3个节点的HA分布式集群

    前期博客 Ambari安装之部署单节点集群 其实,按照这个步骤是一样的.只是按照好3个节点后,再做下HA即可. 部署3个节点的HA分布式集群 (1)添加机器 和添加服务的操作类似,如下图 之后的添加a ...

  4. Hadoop教程(五)Hadoop分布式集群部署安装

    Hadoop教程(五)Hadoop分布式集群部署安装 1 Hadoop分布式集群部署安装 在hadoop2.0中通常由两个NameNode组成,一个处于active状态,还有一个处于standby状态 ...

  5. Hadoop分布式集群部署(单namenode节点)

    Hadoop分布式集群部署 系统系统环境: OS: CentOS 6.8 内存:2G CPU:1核 Software:jdk-8u151-linux-x64.rpm hadoop-2.7.4.tar. ...

  6. HBase HA的分布式集群部署(适合3、5节点)

    本博文的主要内容有: .HBase的分布模式(3.5节点)安装    .HBase的分布模式(3.5节点)的启动 .HBase HA的分布式集群的安装 .HBase HA的分布式集群的启动    .H ...

  7. Hadoop及Zookeeper+HBase完全分布式集群部署

    Hadoop及HBase集群部署 一. 集群环境 系统版本 虚拟机:内存 16G CPU 双核心 系统: CentOS-7 64位 系统下载地址: http://124.202.164.6/files ...

  8. hadoop分布式集群部署①

     Linux系统的安装和配置.(在VM虚拟机上) 一:安装虚拟机VMware Workstation 14 Pro 以上,虚拟机软件安装完成. 二:创建虚拟机. 三:安装CentOS系统 (1)上面步 ...

  9. HBase HA分布式集群搭建

    HBase HA分布式集群搭建部署———集群架构 搭建之前建议先学习好HBase基本构架原理:https://www.cnblogs.com/lyywj170403/p/9203012.html 集群 ...

随机推荐

  1. 【SimpleMsgPack.NET】发布一个msgpack协议C#版本的解析开源库

    这两年一直都关注这IOCP在网络通信这方面的应用,当然数据的传递是经常需要的.今年接触了MsgPack格式,发现他用来做传输时数据打包真是太爽了.因为他可以直接打包二进制数据,不需要任何的转换.有人会 ...

  2. js 将数字转换成人民币大写的方法

    //将数字转换成人民币大写的方法 var digitUppercase = function (n) { var fraction = ['角', '分']; var digit = [ '零', ' ...

  3. 【linux】dpkg info修复及dpkg: warning: files list file for package

    mv /var/lib/dpkg/info /var/lib/dpkg/info.bak //现将info文件夹更名 sudo mkdir /var/lib/dpkg/info //再新建一个新的in ...

  4. Android实例-获取屏幕的物理分辨率

    相关资料: http://blog.qdac.cc/?p=1161 实例代码: unit Unit1; interface uses System.SysUtils, System.Types, Sy ...

  5. with open

    再考虑一个场景,要读取文件内容,并把年龄和名字的顺序交换存成新文件age_name.txt,这时可以同时打开两个文件:with open('name_age.txt', 'r') as fread, ...

  6. [转]PowerDesigner中name和code取消自动关联

    PowerDesigner中,但修改了某个字段的name,其code也跟着修改,这个问题很讨厌,因为一般来说,name是中文的,code是字段名. 解决方法如下: 1.选择Tools->Gene ...

  7. mysql的骚操作:自增长的字段同时插入到另一个字段

    如题 ' ); select * from information_schema.tables where table_schema ='mytest' and table_name='users'; ...

  8. am335x Watchdog 生效导致 LAN8710A action 异常

    原因是因为 watchdog 生效的时候,产生的 reset 信号,只有 2.5us 左右,太短了,导致 LAN8710a 未完全重启 在 kernel watchdog 里面增加下面的代码即可 86 ...

  9. Go Revel - Validation(验证)

    Revel提供了内建函数用于验证参数.它提供了: 一个`Validation`上下文集合来管理验证错误信息(键与消息内容) 辅助函数用于检查数据并将错误信息放入上下文 一个模板函数用于从`Valida ...

  10. Java编程的逻辑 (44) - 剖析TreeSet

    本系列文章经补充和完善,已修订整理成书<Java编程的逻辑>,由机械工业出版社华章分社出版,于2018年1月上市热销,读者好评如潮!各大网店和书店有售,欢迎购买,京东自营链接:http:/ ...