Hadoop 2.2.0部署安装(笔记,单机安装)
SSH无密安装与配置
具体配置步骤:
◎ 在root根目录下创建.ssh目录 (必须root用户登录)
cd /root & mkdir .ssh
chmod 700 .ssh & cd .ssh
◎ 创建密码为空的 RSA 密钥对:
ssh-keygen -t rsa -P ""
◎ 在提示的对称密钥名称中输入 id_rsa将公钥添加至 authorized_keys 中:
cat id_rsa.pub >> authorized_keys
chmod 644 authorized_keys # 重要
◎ 编辑 sshd 配置文件 /etc/ssh/sshd_config ,把 #AuthorizedKeysFile .ssh/authorized_keys 前面的注释取消掉。
◎ 重启 sshd 服务:
service sshd restart
◎ 测试 SSH 连接。连接时会提示是否连接,按回车后会将此公钥加入至 knows_hosts 中:
ssh localhost# 输入用户名密码
Hadoop 2.2.0部署安装
具体步骤如下:
◎ 下载文件。
◎ 解压hadoop 配置环境。
#在root根目录下创建hadoop文件夹
mkdir hadoop;
cd hadoop;
#将hadoop 2.2.0 安装文件放置到hadoop目录文件夹下。
#解压hadoop 2.2.0 文件
tar -zxvf hadoop-2.2.0.tar.gz
#进入hadoop -2.2.0 文件夹
cd hadoop-2.2.0
#进入hadoop配置文件夹
cd etc/hadoop
#修改core-site.xml
vi core-site.xml 添加以下信息(hadoop.tmp.dir、fs.default.name):
- <?xml version="1.0" encoding="UTF-8"?>
- <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
- <!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
- -->
- <!-- Put site-specific property overrides in this file. -->
- <configuration>
- <property>
- <name>hadoop.tmp.dir</name>
- <value>/root/hadoop/hadoop-${user.name}</value>
- <description>A base for other temporaydirectories</description>
- </property>
- <property>
- <name>fs.default.name</name>
- <value>hdfs://localhost:8010</value>
- <description></description>
- </property>
- </configuration>
- #修改hdfs-site.xml配置文件, namenode和datanode存储路径的设置
- <?xml version="1.0" encoding="UTF-8"?>
- <?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
- <!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
- -->
- <!-- Put site-specific property overrides in this file. -->
- <configuration>
- <property>
- <name>dfs.namenode.name.dir</name>
- <value>/root/hadoop/hdfs/namenode</value>
- <description>Determineswhere on the local filesystem the DFS name node should store the name table. Ifthis is a comma-delimited list of directories then the name table is replicatedin all of the directories, for redundancy. </description>
- <final>true</final>
- </property>
- <property>
- <name>dfs.datanode.data.dir</name>
- <value>/root/hadoop/hdfs/datanode</value>
- <description>Determineswhere on the local filesystem an DFS data node should store its blocks. If thisis a comma-delimited list of directories, then data will be stored in all nameddirectories, typically on different devices.Directories that do not exist areignored.
- </description>
- <final>true</final>
- </property>
- <property>
- <!-- 副本个数-->
- <name>dfs.replication</name>
- <value>1</value>
- </property>
- <property>
- <name>dfs.permissions</name>
- <value>false</value>
- </property>
- </configuration>
#修改mapred-site.xml
添加 dfs.namenode.name.dir、dfs.datanode.data.dir、dfs.replication、dfs.permissions等参数信息
- <?xml version="1.0"?>
- <!--
- Licensed under the Apache License, Version 2.0 (the "License");
- you may not use this file except in compliance with the License.
- You may obtain a copy of the License at
- http://www.apache.org/licenses/LICENSE-2.0
- Unless required by applicable law or agreed to in writing, software
- distributed under the License is distributed on an "AS IS" BASIS,
- WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- See the License for the specific language governing permissions and
- limitations under the License. See accompanying LICENSE file.
- -->
- <!-- Put site-specific property overrides in this file. -->
- <configuration>
- <property>
- <name>mapred.job.tracker</name>
- <value>localhost:54311</value>
- <description>The host and port that the MapReduce job tracker runs
- at. If "local", thenjobs are run in-process as a single map
- and reduce task.
- </description>
- </property>
- <property>
- <name>mapred.map.tasks</name>
- <value>10</value>
- <description>As a rule of thumb, use 10x the number of slaves(i.e., number of tasktrackers).</description>
- </property>
- <property>
- <name>mapred.reduce.tasks</name>
- <value>2</value>
- <description>As a rule of thumb, use 2x the number of slaveprocessors (i.e., number of tasktrackers).</description>
- </property>
- </configuration>
◎ 设置java环境(接上述步骤)。
#修改hadoop-env.sh 设置java路径参数,export JAVA_HOME=/usr/local/jdk1.7。
- # Copyright 2011 The Apache Software Foundation
- #
- # Licensed to the Apache Software Foundation (ASF) under one
- # or more contributor license agreements. See the NOTICE file
- # distributed with this work for additional information
- # regarding copyright ownership. The ASF licenses this file
- # to you under the Apache License, Version 2.0 (the
- # "License"); you may not use this file except in compliance
- # with the License. You may obtain a copy of the License at
- #
- # http://www.apache.org/licenses/LICENSE-2.0
- #
- # Unless required by applicable law or agreed to in writing, software
- # distributed under the License is distributed on an "AS IS" BASIS,
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- # See the License for the specific language governing permissions and
- # limitations under the License.
- # Set Hadoop-specific environment variables here.
- # The only required environment variable is JAVA_HOME. All others are
- # optional. When running a distributed configuration it is best to
- # set JAVA_HOME in this file, so that it is correctly defined on
- # remote nodes.
- # The java implementation to use.
- export JAVA_HOME=/usr/local/jdk1.7
- # The jsvc implementation to use. Jsvc is required to run secure datanodes.
- #export JSVC_HOME=${JSVC_HOME}
- export HADOOP_CONF_DIR=${HADOOP_CONF_DIR:-"/etc/hadoop"}
- # Extra Java CLASSPATH elements. Automatically insert capacity-scheduler.
- for f in $HADOOP_HOME/contrib/capacity-scheduler/*.jar; do
- if [ "$HADOOP_CLASSPATH" ]; then
- export HADOOP_CLASSPATH=$HADOOP_CLASSPATH:$f
- else
- export HADOOP_CLASSPATH=$f
- fi
- done
- # The maximum amount of heap to use, in MB. Default is 1000.
- #export HADOOP_HEAPSIZE=
- #export HADOOP_NAMENODE_INIT_HEAPSIZE=""
- # Extra Java runtime options. Empty by default.
- export HADOOP_OPTS="$HADOOP_OPTS -Djava.net.preferIPv4Stack=true"
- # Command specific options appended to HADOOP_OPTS when specified
- export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS"
- export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS $HADOOP_DATANODE_OPTS"
- export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_SECONDARYNAMENODE_OPTS"
- # The following applies to multiple commands (fs, dfs, fsck, distcp etc)
- export HADOOP_CLIENT_OPTS="-Xmx512m $HADOOP_CLIENT_OPTS"
- #HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS"
- # On secure datanodes, user to run the datanode as after dropping privileges
- export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER}
- # Where log files are stored. $HADOOP_HOME/logs by default.
- #export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER
- # Where log files are stored in the secure data environment.
- export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER}
- # The directory where pid files are stored. /tmp by default.
- # NOTE: this should be set to a directory that can only be written to by
- # the user that will run the hadoop daemons. Otherwise there is the
- # potential for a symlink attack.
- export HADOOP_PID_DIR=${HADOOP_PID_DIR}
- export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR}
- # A string representing this instance of hadoop. $USER by default.
- export HADOOP_IDENT_STRING=$USER
◎ 设置hadoop环境变量[HADOOP_HOME]。
vi /etc/profile 输入 export HADOOP_HOME=/root/hadoop/hadoop-2.2.0
source /etc/profile 让环境变量生效。
测试hadoop环境变量是否生效:
echo $HADOOP_HOME
/root/hadoop/hadoop-2.2.0
◎ 进入hadoop安装目录,进入bin目录,格式化hdfs。
./hadoop namenode –format
◎ 启动hadoop ,进入hadoop安装目录,进入sbin目录。
./start-all.sh
◎ 验证安装,登录 http://localhost:50070/ 。
文章转载请注明出处:http://www.cnblogs.com/likehua/p/3825810.html
相关推荐:
sqoop安装参考:http://www.cnblogs.com/likehua/p/3825489.html
hive安装参考:http://www.cnblogs.com/likehua/p/3825479.html
Hadoop 2.2.0部署安装(笔记,单机安装)的更多相关文章
- Hadoop 2.2.0 4结点集群安装 非HA
总体介绍 虚拟机4台,分布在1个物理机上,配置基于hadoop的集群中包括4个节点: 1个 Master, 3个 Salve,i p分布为: 10.10.96.33 hadoop1 (Master) ...
- Redis单机安装以及集群搭建
今天主要来看一下Redis的安装以及集群搭建(我也是第一次搭建). 环境:CentOS 7.1,redis-5.0.7 一.单机安装 1.将Redis安装包放置服务器并解压 2.进入redis安装目录 ...
- asp.net core2.0 部署centos7/linux系统 --守护进程supervisor(二)
原文:asp.net core2.0 部署centos7/linux系统 --守护进程supervisor(二) 续上一篇文章:asp.net core2.0 部署centos7/linux系统 -- ...
- CentOS7.0分布式安装HADOOP 2.6.0笔记-转载的
三台虚拟机,IP地址通过路由器静态DHCP分配 (这样就无需设置host了). 三台机器信息如下 - 1. hadoop-a: 192.168.0.20 #master 2. ha ...
- (转)ZooKeeper 笔记(1) 安装部署及hello world
ZooKeeper 笔记(1) 安装部署及hello world 先给一堆学习文档,方便以后查看 官网文档地址大全: OverView(概述) http://zookeeper.apache.or ...
- CentOS 7.1静默安装11.2.0.3 64位单机数据库软件
第1章 CentOS 7.1静默安装11.2.0.3 64位单机数据库软件 1.1 安装前的准备工作 1.1.1 软件准备 1.1.2 检查硬件 注意这里的内存应该满足要求,不然 ...
- 单机安装hadoop+hive+presto
系统环境 在个人笔记本上使用virtualbox虚拟机 os:centos -7.x86-64.everything.1611 ,内核 3.10.0-514.el7.x86_64 注:同样可以使用r ...
- Spark Tachyon编译部署(含单机和集群模式安装)
Tachyon编译部署 编译Tachyon 单机部署Tachyon 集群模式部署Tachyon 1.Tachyon编译部署 Tachyon目前的最新发布版为0.7.1,其官方网址为http://tac ...
- Presto单机/集群模式安装笔记
Presto单机/集群模式安装笔记 一.安装环境 二.安装步骤 三.集群模式安装: 3.1 集群模式修改配置部分 3.1.1 coordinator 节点配置. Node172配置 3.1.2 nod ...
随机推荐
- Android 动画之TranslateAnimation应用详解
TranslateAnimation比较常用,比如QQ,网易新闻菜单条的动画,就可以用TranslateAnimation实现, 通过TranslateAnimation(float fromXDel ...
- [CareerCup] 11.3 Search in Rotated Sorted Array 在旋转有序矩阵中搜索
11.3 Given a sorted array of n integers that has been rotated an unknown number of times, write code ...
- 20135202闫佳歆--week 8 课本第4章学习笔记
第四章 进程调度 一.多任务 多任务操作系统就是能同时并发的交互执行多个进程的操作系统. 多任务操作系统使多个进程处于堵塞或者睡眠状态,实际不被投入执行,这些任务尽管位于内存,但是并不处于可运行状态. ...
- Linux10.11-10.18)学习笔记(
1Y86指令集体系结构 ISA:指令集体系结构,一个处理器支持的指令和指令的字节级编码 程序员可见的状态 Y86程序中的每条指令都会读取或修改处理器状态的某些部分,称为程序员可见状态.其中包括: 8个 ...
- POJ2289-Jamie's Contact Groups-二分图多重匹配-ISAP
注意POJ数组越界可能返回TLE!!! 网络流的maxn大小要注意 其他没什么了 裸二分答案+isap乱搞 不过复杂度没搞懂 V=1e3 E = 1e5 那ISAP的O(V^2E)怎么算都不行啊 /* ...
- 学习笔记——Maven实战(六)Gradle,构建工具的未来?
Maven面临的挑战 软件行业新旧交替的速度之快往往令人咂舌,不用多少时间,你就会发现曾经大红大紫的技术已经成为了昨日黄花,当然,Maven也不会例外.虽然目前它基本上是Java构建的事实标准,但我们 ...
- Windows Phone 开发——相机功能开发
相机功能是手机区别于PC的一大功能,在做手机应用时,如果合理的利用了拍照功能,可能会给自己的应用增色很多.使用Windows Phone的相机功能,有两种方法,一种是使用PhotoCamera类来构建 ...
- MVC4/5+jquery+bootstrap样式+dataTables+linq+WCF+EF6后台和前台的框架集合!好蛋疼哦!数据库支持MYSQL 和MSSQL,oracle。集成腾讯企业邮箱收邮件同步用户SSO登陆等功能。
花费了我好多心血,才做出来,下个项目准备用这个框架! 大家有没有做这方面的可以交流一下! 花费了我好多心血,才做出来,下个项目准备用这个框架! 大家有没有做这方面的可以交流一下! 花费了我好多心血,才 ...
- 你应当如何学习C++(以及编程)(转载)
你应当如何学习C++(以及编程)(rev#1) By 刘未鹏(pongba) C++的罗浮宫(http://blog.csdn.net/pongba) Javascript是世界上最受误解的语言,其实 ...
- javascript继承(五)—prototype最优两种继承(空函数和循环拷贝)
一.利用空函数实现继承 参考了文章javascript继承—prototype属性介绍(2) 中叶小钗的评论,对这篇文章中的方案二利用一个空函数进行修改,可以解决创建子类对象时,父类实例化的过程中特权 ...