Hadoop安装


环境配置 - Single Node Cluster

一、JDK配置

Ref: How to install hadoop 2.7.3 single node cluster on ubuntu 16.04

Ubuntu 18 + Hadoop 2.7.3 + Java 8

  1. $ sudo apt-get update
  2. $ sudo apt-get install openjdk--jdk
  3. $ java version

版本若不对, 可以切换.

  1. $ update-alternatives --config java

二、SSH配置

Now we are logined in in ‘hduser’.

  1. $ ssh-keygen -t rsa
  2. NOTE: Leave file name and other things blank.
  3. $ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
  4. $ chmod ~/.ssh/authorized_keys
  5. $ ssh localhost

三、安装 Hadoop

其实就是把内容整体移动到一个新的位置: /usr/local/下面

  1. $ wget http://www-us.apache.org/dist/hadoop/common/hadoop-2.7.3/hadoop-2.7.3.tar.gz
  2. $ tar xvzf hadoop-2.7..tar.gz
  3. $ sudo mkdir -p /usr/local/hadoop
  4. $ cd hadoop-2.7./
  5. $ sudo mv * /usr/local/hadoop
  6. $ sudo chown -R hduser:hadoop /usr/local/hadoop

四、配置 Hadoop

4.1 ~/.bashrc

4.2 hadoop-env.sh

4.3 core-site.xml

4.4 mapred-site.xml

4.5 hdfs-site.xml

4.6 yarn-site.xml

4.1 ~/.bashrc

  1. #HADOOP VARIABLES START
  2. export JAVA_HOME=/usr/lib/jvm/java--openjdk-amd64
  3. export HADOOP_HOME=/usr/local/hadoop
  4. export PATH=$PATH:$HADOOP_HOME/bin
  5. export PATH=$PATH:$HADOOP_HOME/sbin
  6. export HADOOP_MAPRED_HOME=$HADOOP_HOME
  7. export HADOOP_COMMON_HOME=$HADOOP_HOME
  8. export HADOOP_HDFS_HOME=$HADOOP_HOME
  9. export YARN_HOME=$HADOOP_HOME
  10. export HADOOP_COMMON_LIB_NATIVE_DIR=$HADOOP_HOME/lib/native
  11. export HADOOP_OPTS="-Djava.library.path=$HADOOP_HOME/lib"
  12. #HADOOP VARIABLES END

4.2 设置 java env

/usr/local/hadoop/etc/hadoop/hadoop-env.sh 文件中设置。

  1. export JAVA_HOME=/usr/lib/jvm/java-8-openjdk-amd64

4.3 设置 core-site

/usr/local/hadoop/etc/hadoop/core-site.xml 文件中设置。

  1. <configuration>
  2. <property>
  3. <name>hadoop.tmp.dir</name>
  4. <value>/app/hadoop/tmp</value>
  5. <description>A base for other temporary directories.</description>
  6. </property>
  7. <property>
  8. <name>fs.default.name</name>
  9. <value>hdfs://localhost:54310</value>
  10. <description>The name of the default file system. A URI whose scheme and authority determine the FileSystem implementation. The uri's scheme determines the config property (fs.SCHEME.impl) naming the FileSystem implementation class. The uri's authority is used to determine the host, port, etc. for a filesystem.</description>
  11. </property>
  12. </configuration> 

参考: https://blog.csdn.net/Mr_LeeHY/article/details/77049800

  1. <configuration>
  2. <!--指定namenode的地址-->
  3. <property>
  4. <name>fs.defaultFS</name>
  5. <value>hdfs://master:9000</value>
  6. </property>
  7. <!--用来指定使用hadoop时产生文件的存放目录-->
  8. <property>
  9. <name>hadoop.tmp.dir</name>
  10. <value>file:///usr/hadoop/hadoop-2.6.0/tmp</value>
  11. </property>
  12. <!--用来设置检查点备份日志的最长时间-->
  13. <name>fs.checkpoint.period</name>
  14. <value>3600</value>
  15. </configuration>

所以,创建对应的目录存放文件。

  1. $ sudo mkdir -p /app/hadoop/tmp
  2. $ sudo chown hduser:hadoop /app/hadoop/tmp

4.4 设置 mapred-site

/usr/local/hadoop/etc/hadoop/mapred-site.xml 文件中配置。

同级目录下,已提供了一个模板,先拷贝。

  1. $ cp /usr/local/hadoop/etc/hadoop/mapred-site.xml.template /usr/local/hadoop/etc/hadoop/mapred-site.xml

告诉hadoop以后 MR (Map/Reduce) 运行在YARN上。

  1. <configuration>
    <property>
  2. <name>mapred.job.tracker</name>
  3. <value>localhost:54311</value>

  4. <description> The host and port that the MapReduce job tracker runs at. If "local", then jobs are run in-process as a single map and reduce task.
  5. </description>
  6. </property>
  7. <property>
  8. <name>mapreduce.framework.name</name>
  9. <value>yarn</value>
  10. </property>
  11. </configuration>

4.5 设置 hdfs-site

/usr/local/hadoop/etc/hadoop/hdfs-site.xml 文件中配置。

  1. <configuration>
  2. <property>
  3. <name>dfs.replication</name>
  4. <value>1</value>
  5. <description>Default block replication.The actual number of replications can be specified when the file is created. The default is used if replication is not specified in create time.
  6. </description>
  7. </property>
  8. <property>
  9. <name>dfs.namenode.name.dir</name>
  10. <value>file:/usr/local/hadoop_store/hdfs/namenode</value>
  11. </property>
  12. <property>
  13. <name>dfs.datanode.data.dir</name>
  14. <value>file:/usr/local/hadoop_store/hdfs/datanode</value>
  15. </property>
  16. </configuration>

参考:https://blog.csdn.net/Mr_LeeHY/article/details/77049800

  1. <configuration>
  2. <!--指定hdfs保存数据的副本数量-->
  3. <property>
  4. <name>dfs.replication</name>
  5. <value>2</value>
  6. </property>
  7. <!--指定hdfs中namenode的存储位置-->
  8. <property>
  9. <name>dfs.namenode.name.dir</name>
  10. <value>file:/usr/hadoop/hadoop-2.6.0/tmp/dfs/name</value>
  11. </property>
  12. <!--指定hdfs中datanode的存储位置-->
  13. <property>
  14. <name>dfs.datanode.data.dir</name>
  15. <value>file:/usr/hadoop/hadoop-2.6.0/tmp/dfs/data</value>
  16. </property>
  17.  
  18. </configuration>

4.6 配置 yarn-site

/usr/local/hadoop/etc/hadoop/yarn-site.xml 文件中配置。

  1. <configuration>
  2. <property>
  3. <name>yarn.nodemanager.aux-services</name>
  4. <value>mapreduce_shuffle</value>
  5. </property>
  6. </configuration>

参考:https://blog.csdn.net/Mr_LeeHY/article/details/77049800

  1. <configuration>
  2. <!--nomenodeManager获取数据的方式是shuffle-->
  3. <property>
  4. <name>yarn.nodemanager.aux-services</name>
  5. <value>mapreduce_shuffle</value>
  6. </property>
  7. <!--指定Yarn的老大(ResourceManager)的地址-->
  8. <property>
  9. <name>yarn.resourcemanager.hostname</name>
  10. <value>master</value>
  11. </property>
  12. <!--Yarn打印工作日志-->
  13. <property>
  14. <name>yarn.log-aggregation-enable</name>
  15. <value>true</value>
  16. </property>
  17.  
  18. <configuration>

五、初始化 fs,并启动守护进程

守护进程:hadoop daemons。

  1. $ hadoop namenode format
  2.  
  3. $ cd /usr/local/hadoop/sbin
  4. $ start-all.sh

查看启动了哪些守护进程。

  1. hadoop@ThinkPad:~$ jps
  2. ResourceManager
  3. SecondaryNameNode
  4. Jps
  5. NodeManager
  6. NameNode

六、Hadoop 测试样例

需要确保datanode启动,这是一个single node cluster。

删除datanode内容,重启,然后再运行测试。

这里设计到jar包,也就是mapReduce的接口java编程。

  1. $ sudo rm -r /usr/local/hadoop_store/hdfs/datanode/current
  2. $ hadoop namenode -format
  3. $ start-all.sh
  4. $ jps
  5. $ hadoop jar ./share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.3.jar pi 2 5
  1. hadoop@unsw-ThinkPad-T490:/usr/local/hadoop$ hadoop jar ./share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7..jar pi
  2. Number of Maps =
  3. Samples per Map =
  4. // :: WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  5. Wrote input for Map #
  6. Wrote input for Map #
  7. Starting Job
  8. // :: INFO client.RMProxy: Connecting to ResourceManager at /0.0.0.0:
  9. // :: INFO input.FileInputFormat: Total input paths to process :
  10. // :: INFO mapreduce.JobSubmitter: number of splits:
  11. // :: INFO mapreduce.JobSubmitter: Submitting tokens for job: job_1571615960327_0001
  12. // :: INFO impl.YarnClientImpl: Submitted application application_1571615960327_0001
  13. // :: INFO mapreduce.Job: The url to track the job: http://unsw-ThinkPad-T490:8088/proxy/application_1571615960327_0001/
  14. // :: INFO mapreduce.Job: Running job: job_1571615960327_0001
  15. // :: INFO mapreduce.Job: Job job_1571615960327_0001 running in uber mode : false
  16. // :: INFO mapreduce.Job: map % reduce %
  17. // :: INFO mapreduce.Job: map % reduce %
  18. // :: INFO mapreduce.Job: map % reduce %
  19. // :: INFO mapreduce.Job: Job job_1571615960327_0001 completed successfully
  20. // :: INFO mapreduce.Job: Counters:
  21. File System Counters
  22. FILE: Number of bytes read=
  23. FILE: Number of bytes written=
  24. FILE: Number of read operations=
  25. FILE: Number of large read operations=
  26. FILE: Number of write operations=
  27. HDFS: Number of bytes read=
  28. HDFS: Number of bytes written=
  29. HDFS: Number of read operations=
  30. HDFS: Number of large read operations=
  31. HDFS: Number of write operations=
  32. Job Counters
  33. Launched map tasks=
  34. Launched reduce tasks=
  35. Data-local map tasks=
  36. Total time spent by all maps in occupied slots (ms)=
  37. Total time spent by all reduces in occupied slots (ms)=
  38. Total time spent by all map tasks (ms)=
  39. Total time spent by all reduce tasks (ms)=
  40. Total vcore-milliseconds taken by all map tasks=
  41. Total vcore-milliseconds taken by all reduce tasks=
  42. Total megabyte-milliseconds taken by all map tasks=
  43. Total megabyte-milliseconds taken by all reduce tasks=
  44. Map-Reduce Framework
  45. Map input records=
  46. Map output records=
  47. Map output bytes=
  48. Map output materialized bytes=
  49. Input split bytes=
  50. Combine input records=
  51. Combine output records=
  52. Reduce input groups=
  53. Reduce shuffle bytes=
  54. Reduce input records=
  55. Reduce output records=
  56. Spilled Records=
  57. Shuffled Maps =
  58. Failed Shuffles=
  59. Merged Map outputs=
  60. GC time elapsed (ms)=
  61. CPU time spent (ms)=
  62. Physical memory (bytes) snapshot=
  63. Virtual memory (bytes) snapshot=
  64. Total committed heap usage (bytes)=
  65. Shuffle Errors
  66. BAD_ID=
  67. CONNECTION=
  68. IO_ERROR=
  69. WRONG_LENGTH=
  70. WRONG_MAP=
  71. WRONG_REDUCE=
  72. File Input Format Counters
  73. Bytes Read=
  74. File Output Format Counters
  75. Bytes Written=
  76. Job Finished in 15.972 seconds
  77. Estimated value of Pi is 3.60000000000000000000

Log

分布式配置

一、前言

伪分布式配置

安装课程:安装配置【厦大课程视频】

配置手册:Hadoop安装教程_单机/伪分布式配置_Hadoop2.6.0/Ubuntu14.04【厦大课程i笔记】

"手把手"环境配置:玩转大数据分析!Spark2.X+Python 精华实战课程(免费)【其实只是环境搭建】

/* 略,更关心真分布式 */

真分布式配置

本地虚拟机实验:1.3 virtualbox高级应用构建本地大数据集群服务器

三台不同云机器:Hadoop完全分布式安装配置完整过程

二、虚拟机

(1) 配置好virtualbox后(需要关闭security boot),配置IP,最好的固定的。

Goto: 1.3 virtualbox高级应用构建本地大数据集群服务器·【只有服务器配置,未配置hadoop】

只需要修改这里,给不同的slave设置不同的ip地址就好了。

  1. /etc/network/interfaces 文件中配置
  2.  
  3. # (注释的内容忽略)增加的Host-only静态IP设置 (enp0s8 是根据拓扑关系映射的网卡名称(旧规则是eth0,eth1))
  4. # 可以通过 ```ls /sys/class/net```查看,是否为enp0s8
  5.  
  6. auto enp0s8
  7. iface enp0s8 inet static
  8. address 192.168.56.106
  9. netmask 255.255.255.0

(2) 再安装工具:

  1. # 安装网络工具
  2. sudo apt install net-tools
  3. # 查看本地网络情况
  4. ifconfig

(3) 通过ssh登录slave机,进行验证。

  1. $ ssh -p 22 hadoop@192.168.56.101
  2. The authenticity of host '192.168.56.101 (192.168.56.101)' can't be established.
  3. ECDSA key fingerprint is SHA256:IPf76acROSwMC7BQO3hBAThLZCovARuoty765MfTps0.
  4. Are you sure you want to continue connecting (yes/no)? yes
  5. Warning: Permanently added '192.168.56.101' (ECDSA) to the list of known hosts.
  6. hadoop@192.168.56.101's password:

  7. hadoop@node2-VirtualBox:~$ hostname
  8. node2-VirtualBox
  9.  
  10. hadoop@node2-VirtualBox:~$ sudo hostname worker1-VirtualBox
  11. [sudo] password for hadoop:
  12. hadoop@node2-VirtualBox:~$ hostname
  13. worker1-VirtualBox

(4)  永久性修改hostname的方式,注意,要修改两个文件。

  1. Type the following command to edit /etc/hostname using nano or vi text editor:
  2. sudo nano /etc/hostname
  3. Delete the old name and setup new name.
  4. Next Edit the /etc/hosts file:
  5. sudo nano /etc/hosts
  6. Replace any occurrence of the existing computer name with your new one.
  7. Reboot the system to changes take effect:
  8. sudo reboot

(5) 关闭 x-window

Ref: 2.2 Hadoop3.1.0完全分布式集群配置与部署

卸载x-window:Remove packages to transform Desktop to Server?

内存比大概是:140M : 900M

如果不是卸载,而是关掉x-window,则,没啥变化,还在内存里,只是变为了inactive。

  1. 按住ctrl+alt+f1,进入命令行。
  2.  
  3. 输入sudo /etc/init.d/lightdm stop
  4. sudo /etc/init.d/lightdm status
  5.  
  6. 重启xserver?输入sudo /etc/init.d/lightdm restart

(6) ssh无密码登录slave

master免密登录到worker中,以下是示范。

  1. ssh-copy-id -i ~/.ssh/id_rsa.pub master
  2. ssh-copy-id -i ~/.ssh/id_rsa.pub worker1
  3. ssh-copy-id -i ~/.ssh/id_rsa.pub worker2

(7) docker配置集群

厦大手册:使用Docker搭建Hadoop分布式集群

三、分布式slave主机配置

写在前面

可以在本地修改好配置文件后,然后将Hadoop复制到集群服务器中。

Single node cluster的方式有点问题,配置过于繁琐,再次全新地配置一遍。

Ref: Part 1: How To install a 3-Node Hadoop Cluster on Ubuntu 16

Ref: How to Install and Set Up a 3-Node Hadoop Cluster【良心配置,可用】

  1. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  2. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ hdfs namenode -format
  3. 2019-10-24 16:10:48,131 INFO namenode.NameNode: STARTUP_MSG:
  4. /************************************************************
  5. STARTUP_MSG: Starting NameNode
  6. STARTUP_MSG: host = node-master/192.168.56.2
  7. STARTUP_MSG: args = [-format]
  8. STARTUP_MSG: version = 3.1.2
  9. STARTUP_MSG: classpath = /usr/local/hadoop/etc/hadoop:/usr/local/hadoop/share/hadoop/common/lib/asm-5.0.4.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-simplekdc-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jul-to-slf4j-1.7.25.jar:/usr/local/hadoop/share/hadoop/common/lib/jsp-api-2.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/common/lib/curator-recipes-2.13.0.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-webapp-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/httpcore-4.4.4.jar:/usr/local/hadoop/share/hadoop/common/lib/token-provider-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerby-pkix-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/metrics-core-3.2.4.jar:/usr/local/hadoop/share/hadoop/common/lib/jsr311-api-1.1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-annotations-2.7.8.jar:/usr/local/hadoop/share/hadoop/common/lib/nimbus-jose-jwt-4.41.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerby-config-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/audience-annotations-0.5.0.jar:/usr/local/hadoop/share/hadoop/common/lib/woodstox-core-5.0.3.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-json-1.19.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-server-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerby-asn1-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerby-xdr-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-configuration2-2.1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jsr305-3.0.0.jar:/usr/local/hadoop/share/hadoop/common/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-core-2.7.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jcip-annotations-1.0-1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-collections-3.2.2.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-security-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-server-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/common/lib/slf4j-api-1.7.25.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-lang3-3.4.jar:/usr/local/hadoop/share/hadoop/common/lib/jettison-1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/zookeeper-3.4.13.jar:/usr/local/hadoop/share/hadoop/common/lib/gson-2.2.4.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-compress-1.18.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-beanutils-1.9.3.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-codec-1.11.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-util-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/json-smart-2.3.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-databind-2.7.8.jar:/usr/local/hadoop/share/hadoop/common/lib/jackson-jaxrs-1.9.13.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-net-3.6.jar:/usr/local/hadoop/share/hadoop/common/lib/re2j-1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-xml-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-core-1.19.jar:/usr/local/hadoop/share/hadoop/common/lib/hadoop-annotations-3.1.2.jar:/usr/local/hadoop/share/hadoop/common/lib/htrace-core4-4.1.0-incubating.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-io-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.25.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-servlet-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-common-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-core-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/hadoop-auth-3.1.2.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-io-2.5.jar:/usr/local/hadoop/share/hadoop/common/lib/kerby-util-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/stax2-api-3.1.4.jar:/usr/local/hadoop/share/hadoop/common/lib/curator-framework-2.13.0.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-math3-3.1.1.jar:/usr/local/hadoop/share/hadoop/common/lib/javax.servlet-api-3.1.0.jar:/usr/local/hadoop/share/hadoop/common/lib/httpclient-4.5.2.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-util-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/netty-3.10.5.Final.jar:/usr/local/hadoop/share/hadoop/common/lib/snappy-java-1.0.5.jar:/usr/local/hadoop/share/hadoop/common/lib/paranamer-2.3.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-server-1.19.jar:/usr/local/hadoop/share/hadoop/common/lib/curator-client-2.13.0.jar:/usr/local/hadoop/share/hadoop/common/lib/avro-1.7.7.jar:/usr/local/hadoop/share/hadoop/common/lib/jsch-0.1.54.jar:/usr/local/hadoop/share/hadoop/common/lib/accessors-smart-1.2.jar:/usr/local/hadoop/share/hadoop/common/lib/jetty-http-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-crypto-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-identity-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/jersey-servlet-1.19.jar:/usr/local/hadoop/share/hadoop/common/lib/jaxb-api-2.2.11.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-client-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/kerb-admin-1.0.1.jar:/usr/local/hadoop/share/hadoop/common/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/common/hadoop-common-3.1.2.jar:/usr/local/hadoop/share/hadoop/common/hadoop-common-3.1.2-tests.jar:/usr/local/hadoop/share/hadoop/common/hadoop-kms-3.1.2.jar:/usr/local/hadoop/share/hadoop/common/hadoop-nfs-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-util-ajax-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/asm-5.0.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-simplekdc-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/curator-recipes-2.13.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-webapp-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/httpcore-4.4.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/token-provider-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerby-pkix-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsr311-api-1.1.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-xc-1.9.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-annotations-2.7.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/nimbus-jose-jwt-4.41.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerby-config-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/audience-annotations-0.5.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/woodstox-core-5.0.3.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-json-1.19.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-server-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerby-asn1-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerby-xdr-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-configuration2-2.1.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsr305-3.0.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/okhttp-2.7.5.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/netty-all-4.0.52.Final.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-core-2.7.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jcip-annotations-1.0-1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-collections-3.2.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-security-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-server-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-lang3-3.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jettison-1.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/zookeeper-3.4.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/gson-2.2.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-compress-1.18.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-beanutils-1.9.3.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-codec-1.11.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-util-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/json-smart-2.3.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-databind-2.7.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jackson-jaxrs-1.9.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-net-3.6.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/re2j-1.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-xml-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-core-1.19.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/hadoop-annotations-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/htrace-core4-4.1.0-incubating.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-io-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-servlet-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/guava-11.0.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-common-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-core-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/hadoop-auth-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-io-2.5.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerby-util-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/stax2-api-3.1.4.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/curator-framework-2.13.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/json-simple-1.1.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-math3-3.1.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/javax.servlet-api-3.1.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/httpclient-4.5.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-util-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/netty-3.10.5.Final.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/snappy-java-1.0.5.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/paranamer-2.3.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-server-1.19.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/curator-client-2.13.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/okio-1.6.0.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/avro-1.7.7.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jsch-0.1.54.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/accessors-smart-1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jetty-http-9.3.24.v20180605.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jaxb-impl-2.2.3-1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-crypto-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-identity-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jersey-servlet-1.19.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/jaxb-api-2.2.11.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-client-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/kerb-admin-1.0.1.jar:/usr/local/hadoop/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-native-client-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-client-3.1.2-tests.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-client-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-rbf-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-rbf-3.1.2-tests.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-nfs-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-httpfs-3.1.2.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-native-client-3.1.2-tests.jar:/usr/local/hadoop/share/hadoop/hdfs/hadoop-hdfs-3.1.2-tests.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/hamcrest-core-1.3.jar:/usr/local/hadoop/share/hadoop/mapreduce/lib/junit-4.11.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-uploader-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-app-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-nativetask-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-3.1.2-tests.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-core-3.1.2.jar:/usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-client-common-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn:/usr/local/hadoop/share/hadoop/yarn/lib/geronimo-jcache_1.0_spec-1.0-alpha-1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/ehcache-3.3.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-client-1.19.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-base-2.7.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/metrics-core-3.2.4.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jersey-guice-1.19.jar:/usr/local/hadoop/share/hadoop/yarn/lib/java-util-1.9.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/aopalliance-1.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/fst-2.50.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-module-jaxb-annotations-2.7.8.jar:/usr/local/hadoop/share/hadoop/yarn/lib/snakeyaml-1.16.jar:/usr/local/hadoop/share/hadoop/yarn/lib/mssql-jdbc-6.2.1.jre7.jar:/usr/local/hadoop/share/hadoop/yarn/lib/swagger-annotations-1.5.4.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guice-servlet-4.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/dnsjava-2.1.7.jar:/usr/local/hadoop/share/hadoop/yarn/lib/javax.inject-1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/guice-4.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/objenesis-1.0.jar:/usr/local/hadoop/share/hadoop/yarn/lib/json-io-2.5.1.jar:/usr/local/hadoop/share/hadoop/yarn/lib/HikariCP-java7-2.4.12.jar:/usr/local/hadoop/share/hadoop/yarn/lib/jackson-jaxrs-json-provider-2.7.8.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-common-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-services-core-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-tests-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-common-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-web-proxy-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-router-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-services-api-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-client-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-api-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-timeline-pluginstorage-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-server-nodemanager-3.1.2.jar:/usr/local/hadoop/share/hadoop/yarn/hadoop-yarn-registry-3.1.2.jar
  10. STARTUP_MSG: build = https://github.com/apache/hadoop.git -r 1019dde65bcf12e05ef48ac71e84550d589e5d9a; compiled by 'sunilg' on 2019-01-29T01:39Z
  11. STARTUP_MSG: java = 1.8.0_222
  12. ************************************************************/
  13. 2019-10-24 16:10:48,148 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
  14. 2019-10-24 16:10:48,243 INFO namenode.NameNode: createNameNode [-format]
  15. 2019-10-24 16:10:48,354 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  16. 2019-10-24 16:10:48,607 INFO common.Util: Assuming 'file' scheme for path /usr/local/hadoop/data/nameNode in configuration.
  17. 2019-10-24 16:10:48,607 INFO common.Util: Assuming 'file' scheme for path /usr/local/hadoop/data/nameNode in configuration.
  18. Formatting using clusterid: CID-cadb861e-e62d-42e6-b62b-f834bbf05bca
  19. 2019-10-24 16:10:48,637 INFO namenode.FSEditLog: Edit logging is async:true
  20. 2019-10-24 16:10:48,653 INFO namenode.FSNamesystem: KeyProvider: null
  21. 2019-10-24 16:10:48,654 INFO namenode.FSNamesystem: fsLock is fair: true
  22. 2019-10-24 16:10:48,657 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false
  23. 2019-10-24 16:10:48,666 INFO namenode.FSNamesystem: fsOwner = hadoop (auth:SIMPLE)
  24. 2019-10-24 16:10:48,666 INFO namenode.FSNamesystem: supergroup = supergroup
  25. 2019-10-24 16:10:48,667 INFO namenode.FSNamesystem: isPermissionEnabled = true
  26. 2019-10-24 16:10:48,667 INFO namenode.FSNamesystem: HA Enabled: false
  27. 2019-10-24 16:10:48,708 INFO common.Util: dfs.datanode.fileio.profiling.sampling.percentage set to 0. Disabling file IO profiling
  28. 2019-10-24 16:10:48,717 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit: configured=1000, counted=60, effected=1000
  29. 2019-10-24 16:10:48,718 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
  30. 2019-10-24 16:10:48,721 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
  31. 2019-10-24 16:10:48,721 INFO blockmanagement.BlockManager: The block deletion will start around 2019 Oct 24 16:10:48
  32. 2019-10-24 16:10:48,723 INFO util.GSet: Computing capacity for map BlocksMap
  33. 2019-10-24 16:10:48,723 INFO util.GSet: VM type = 64-bit
  34. 2019-10-24 16:10:48,725 INFO util.GSet: 2.0% max memory 443 MB = 8.9 MB
  35. 2019-10-24 16:10:48,725 INFO util.GSet: capacity = 2^20 = 1048576 entries
  36. 2019-10-24 16:10:48,731 INFO blockmanagement.BlockManager: dfs.block.access.token.enable = false
  37. 2019-10-24 16:10:48,736 INFO Configuration.deprecation: No unit for dfs.namenode.safemode.extension(30000) assuming MILLISECONDS
  38. 2019-10-24 16:10:48,737 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
  39. 2019-10-24 16:10:48,737 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.min.datanodes = 0
  40. 2019-10-24 16:10:48,737 INFO blockmanagement.BlockManagerSafeMode: dfs.namenode.safemode.extension = 30000
  41. 2019-10-24 16:10:48,737 INFO blockmanagement.BlockManager: defaultReplication = 1
  42. 2019-10-24 16:10:48,737 INFO blockmanagement.BlockManager: maxReplication = 512
  43. 2019-10-24 16:10:48,738 INFO blockmanagement.BlockManager: minReplication = 1
  44. 2019-10-24 16:10:48,738 INFO blockmanagement.BlockManager: maxReplicationStreams = 2
  45. 2019-10-24 16:10:48,738 INFO blockmanagement.BlockManager: redundancyRecheckInterval = 3000ms
  46. 2019-10-24 16:10:48,738 INFO blockmanagement.BlockManager: encryptDataTransfer = false
  47. 2019-10-24 16:10:48,738 INFO blockmanagement.BlockManager: maxNumBlocksToLog = 1000
  48. 2019-10-24 16:10:48,756 INFO namenode.FSDirectory: GLOBAL serial map: bits=24 maxEntries=16777215
  49. 2019-10-24 16:10:48,767 INFO util.GSet: Computing capacity for map INodeMap
  50. 2019-10-24 16:10:48,767 INFO util.GSet: VM type = 64-bit
  51. 2019-10-24 16:10:48,768 INFO util.GSet: 1.0% max memory 443 MB = 4.4 MB
  52. 2019-10-24 16:10:48,768 INFO util.GSet: capacity = 2^19 = 524288 entries
  53. 2019-10-24 16:10:48,768 INFO namenode.FSDirectory: ACLs enabled? false
  54. 2019-10-24 16:10:48,769 INFO namenode.FSDirectory: POSIX ACL inheritance enabled? true
  55. 2019-10-24 16:10:48,769 INFO namenode.FSDirectory: XAttrs enabled? true
  56. 2019-10-24 16:10:48,769 INFO namenode.NameNode: Caching file names occurring more than 10 times
  57. 2019-10-24 16:10:48,773 INFO snapshot.SnapshotManager: Loaded config captureOpenFiles: false, skipCaptureAccessTimeOnlyChange: false, snapshotDiffAllowSnapRootDescendant: true, maxSnapshotLimit: 65536
  58. 2019-10-24 16:10:48,775 INFO snapshot.SnapshotManager: SkipList is disabled
  59. 2019-10-24 16:10:48,778 INFO util.GSet: Computing capacity for map cachedBlocks
  60. 2019-10-24 16:10:48,778 INFO util.GSet: VM type = 64-bit
  61. 2019-10-24 16:10:48,779 INFO util.GSet: 0.25% max memory 443 MB = 1.1 MB
  62. 2019-10-24 16:10:48,779 INFO util.GSet: capacity = 2^17 = 131072 entries
  63. 2019-10-24 16:10:48,784 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10
  64. 2019-10-24 16:10:48,784 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10
  65. 2019-10-24 16:10:48,784 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
  66. 2019-10-24 16:10:48,787 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
  67. 2019-10-24 16:10:48,787 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
  68. 2019-10-24 16:10:48,789 INFO util.GSet: Computing capacity for map NameNodeRetryCache
  69. 2019-10-24 16:10:48,789 INFO util.GSet: VM type = 64-bit
  70. 2019-10-24 16:10:48,789 INFO util.GSet: 0.029999999329447746% max memory 443 MB = 136.1 KB
  71. 2019-10-24 16:10:48,789 INFO util.GSet: capacity = 2^14 = 16384 entries
  72. 2019-10-24 16:10:48,814 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1068893594-192.168.56.2-1571893848809
  73. 2019-10-24 16:10:48,832 INFO common.Storage: Storage directory /usr/local/hadoop/data/nameNode has been successfully formatted.
  74. 2019-10-24 16:10:48,838 INFO namenode.FSImageFormatProtobuf: Saving image file /usr/local/hadoop/data/nameNode/current/fsimage.ckpt_0000000000000000000 using no compression
  75. 2019-10-24 16:10:48,910 INFO namenode.FSImageFormatProtobuf: Image file /usr/local/hadoop/data/nameNode/current/fsimage.ckpt_0000000000000000000 of size 393 bytes saved in 0 seconds .
  76. 2019-10-24 16:10:48,918 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
  77. 2019-10-24 16:10:48,924 INFO namenode.NameNode: SHUTDOWN_MSG:
  78. /************************************************************
  79. SHUTDOWN_MSG: Shutting down NameNode at node-master/192.168.56.2
  80. ************************************************************/
  81. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  82. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  83. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  84. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ jps
  85. 2128 Jps
  86. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  87. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  88. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  89. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ start-dfs.sh
  90. Starting namenodes on [node-master]
  91. Starting datanodes
  92. Starting secondary namenodes [node-master]
  93. 2019-10-24 16:11:40,842 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  94. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  95. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  96. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  97. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ jps
  98. 2739 Jps
  99. 2342 NameNode
  100. 2617 SecondaryNameNode
  101. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  102. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  103. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ stop-dfs.sh
  104. Stopping namenodes on [node-master]
  105. Stopping datanodes
  106. Stopping secondary namenodes [node-master]
  107. 2019-10-24 16:12:18,740 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  108. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  109. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  110. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  111. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ hdfs dfsadmin -report
  112. 2019-10-24 16:12:41,062 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  113. report: Call From node-master/192.168.56.2 to node-master:9000 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
  114. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  115. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  116. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  117. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  118. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ start-dfs.sh
  119. Starting namenodes on [node-master]
  120. Starting datanodes
  121. Starting secondary namenodes [node-master]
  122. 2019-10-24 16:13:16,921 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  123. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  124. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  125. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  126. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ hdfs dfsadmin -report
  127. 2019-10-24 16:13:21,162 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  128. Configured Capacity: 20014161920 (18.64 GB)
  129. Present Capacity: 2642968576 (2.46 GB)
  130. DFS Remaining: 2642919424 (2.46 GB)
  131. DFS Used: 49152 (48 KB)
  132. DFS Used%: 0.00%
  133. Replicated Blocks:
  134. Under replicated blocks: 0
  135. Blocks with corrupt replicas: 0
  136. Missing blocks: 0
  137. Missing blocks (with replication factor 1): 0
  138. Low redundancy blocks with highest priority to recover: 0
  139. Pending deletion blocks: 0
  140. Erasure Coded Block Groups:
  141. Low redundancy block groups: 0
  142. Block groups with corrupt internal blocks: 0
  143. Missing block groups: 0
  144. Low redundancy blocks with highest priority to recover: 0
  145. Pending deletion blocks: 0
  146.  
  147. -------------------------------------------------
  148. Live datanodes (2):
  149.  
  150. Name: 192.168.56.101:9866 (node1)
  151. Hostname: node1
  152. Decommission Status : Normal
  153. Configured Capacity: 10007080960 (9.32 GB)
  154. DFS Used: 24576 (24 KB)
  155. Non DFS Used: 7705030656 (7.18 GB)
  156. DFS Remaining: 1773494272 (1.65 GB)
  157. DFS Used%: 0.00%
  158. DFS Remaining%: 17.72%
  159. Configured Cache Capacity: 0 (0 B)
  160. Cache Used: 0 (0 B)
  161. Cache Remaining: 0 (0 B)
  162. Cache Used%: 100.00%
  163. Cache Remaining%: 0.00%
  164. Xceivers: 1
  165. Last contact: Thu Oct 24 16:13:18 AEDT 2019
  166. Last Block Report: Thu Oct 24 16:13:12 AEDT 2019
  167. Num of Blocks: 0
  168.  
  169. Name: 192.168.56.102:9866 (node2)
  170. Hostname: node2
  171. Decommission Status : Normal
  172. Configured Capacity: 10007080960 (9.32 GB)
  173. DFS Used: 24576 (24 KB)
  174. Non DFS Used: 8609099776 (8.02 GB)
  175. DFS Remaining: 869425152 (829.15 MB)
  176. DFS Used%: 0.00%
  177. DFS Remaining%: 8.69%
  178. Configured Cache Capacity: 0 (0 B)
  179. Cache Used: 0 (0 B)
  180. Cache Remaining: 0 (0 B)
  181. Cache Used%: 100.00%
  182. Cache Remaining%: 0.00%
  183. Xceivers: 1
  184. Last contact: Thu Oct 24 16:13:18 AEDT 2019
  185. Last Block Report: Thu Oct 24 16:13:12 AEDT 2019
  186. Num of Blocks: 0
  187.  
  188. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  189. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  190. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  191. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ jps
  192. 3795 SecondaryNameNode
  193. 3983 Jps
  194. 3519 NameNode
  195. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  196. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  197. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ hdfs dfs -mkdir -p /user/hadoop
  198. 2019-10-24 16:16:17,898 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  199. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  200. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  201. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ hdfs dfs -mkdir books
  202. 2019-10-24 16:16:24,789 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  203. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  204. hadoop@node-master:/usr/local/hadoop/etc/hadoop$
  205. hadoop@node-master:/usr/local/hadoop/etc/hadoop$ cd /home/hadoop
  206. hadoop@node-master:~$ wget -O alice.txt https://www.gutenberg.org/files/11/11-0.txt
  207. --2019-10-24 16:16:42-- https://www.gutenberg.org/files/11/11-0.txt
  208. Resolving www.gutenberg.org (www.gutenberg.org)... 152.19.134.47, 2610:28:3090:3000:0:bad:cafe:47
  209. Connecting to www.gutenberg.org (www.gutenberg.org)|152.19.134.47|:443... connected.
  210. HTTP request sent, awaiting response... 200 OK
  211. Length: 173595 (170K) [text/plain]
  212. Saving to: 'alice.txt’
  213.  
  214. alice.txt 100%[=============================================================================================================>] 169.53K 51.4KB/s in 3.3s
  215.  
  216. 2019-10-24 16:16:51 (51.4 KB/s) - 'alice.txt saved [173595/173595]
  217.  
  218. hadoop@node-master:~$
  219. hadoop@node-master:~$
  220. hadoop@node-master:~$ wget -O holmes.txt https://www.gutenberg.org/files/1661/1661-0.txt
  221. --2019-10-24 16:16:56-- https://www.gutenberg.org/files/1661/1661-0.txt
  222. Resolving www.gutenberg.org (www.gutenberg.org)... 152.19.134.47, 2610:28:3090:3000:0:bad:cafe:47
  223. Connecting to www.gutenberg.org (www.gutenberg.org)|152.19.134.47|:443... connected.
  224. HTTP request sent, awaiting response... 200 OK
  225. Length: 607788 (594K) [text/plain]
  226. Saving to: 'holmes.txt’
  227.  
  228. holmes.txt 100%[=============================================================================================================>] 593.54K 138KB/s in 4.3s
  229.  
  230. 2019-10-24 16:17:03 (138 KB/s) - 'holmes.txt saved [607788/607788]
  231.  
  232. hadoop@node-master:~$
  233. hadoop@node-master:~$
  234. hadoop@node-master:~$ wget -O frankenstein.txt https://www.gutenberg.org/files/84/84-0.txt
  235. --2019-10-24 16:17:07-- https://www.gutenberg.org/files/84/84-0.txt
  236. Resolving www.gutenberg.org (www.gutenberg.org)... 152.19.134.47, 2610:28:3090:3000:0:bad:cafe:47
  237. Connecting to www.gutenberg.org (www.gutenberg.org)|152.19.134.47|:443... connected.
  238. HTTP request sent, awaiting response... 200 OK
  239. Length: 450783 (440K) [text/plain]
  240. Saving to: 'frankenstein.txt’
  241.  
  242. frankenstein.txt 100%[=============================================================================================================>] 440.22K 124KB/s in 3.6s
  243.  
  244. 2019-10-24 16:17:14 (124 KB/s) - 'frankenstein.txt saved [450783/450783]
  245.  
  246. hadoop@node-master:~$
  247. hadoop@node-master:~$
  248. hadoop@node-master:~$ hdfs dfs -put alice.txt holmes.txt frankenstein.txt books
  249. 2019-10-24 16:17:21,244 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  250. hadoop@node-master:~$
  251. hadoop@node-master:~$
  252. hadoop@node-master:~$ hdfs dfs -ls books
  253. 2019-10-24 16:17:29,413 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  254. Found 3 items
  255. -rw-r--r-- 1 hadoop supergroup 173595 2019-10-24 16:17 books/alice.txt
  256. -rw-r--r-- 1 hadoop supergroup 450783 2019-10-24 16:17 books/frankenstein.txt
  257. -rw-r--r-- 1 hadoop supergroup 607788 2019-10-24 16:17 books/holmes.txt
  258. hadoop@node-master:~$
  259. hadoop@node-master:~$
  260. hadoop@node-master:~$ hdfs dfs -get books/alice.txt
  261. 2019-10-24 16:17:35,328 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  262. get: `alice.txt': File exists
  263. hadoop@node-master:~$

实践日志

(1) Monitor your HDFS Cluster

主要是配置slave文件,3.0版本后改为了worker文件。

如果运行失败,就关掉,format一下,再开启。然后再运行下面命令。

  1. /usr/local/hadoop$ hdfs dfsadmin -report
  2. 19/10/22 17:24:07 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
  3. Configured Capacity: 20014161920 (18.64 GB)
  4. Present Capacity: 7294656512 (6.79 GB)
  5. DFS Remaining: 7294607360 (6.79 GB)
  6. DFS Used: 49152 (48 KB)
  7. DFS Used%: 0.00%
  8. Under replicated blocks: 0
  9. Blocks with corrupt replicas: 0
  10. Missing blocks: 0  
  11. Missing blocks (with replication factor 1): 0
  12.  
  13. -------------------------------------------------
  14. Live datanodes (2):
  15.  
  16. Name: 192.168.56.102:50010 (worker2-VirtualBox)
  17. Hostname: worker2-VirtualBox
  18. Decommission Status : Normal
  19. Configured Capacity: 10007080960 (9.32 GB)
  20. DFS Used: 24576 (24 KB)
  21. Non DFS Used: 6359736320 (5.92 GB)
  22. DFS Remaining: 3647320064 (3.40 GB)
  23. DFS Used%: 0.00%
  24. DFS Remaining%: 36.45%
  25. Configured Cache Capacity: 0 (0 B)
  26. Cache Used: 0 (0 B)
  27. Cache Remaining: 0 (0 B)
  28. Cache Used%: 100.00%
  29. Cache Remaining%: 0.00%
  30. Xceivers: 1
  31. Last contact: Tue Oct 22 17:24:06 AEDT 2019
  32.  
  33. Name: 192.168.56.101:50010 (worker1-VirtualBox)
  34. Hostname: worker1-VirtualBox
  35. Decommission Status : Normal
  36. Configured Capacity: 10007080960 (9.32 GB)
  37. DFS Used: 24576 (24 KB)
  38. Non DFS Used: 6359769088 (5.92 GB)
  39. DFS Remaining: 3647287296 (3.40 GB)
  40. DFS Used%: 0.00%
  41. DFS Remaining%: 36.45%
  42. Configured Cache Capacity: 0 (0 B)
  43. Cache Used: 0 (0 B)
  44. Cache Remaining: 0 (0 B)
  45. Cache Used%: 100.00%
  46. Cache Remaining%: 0.00%
  47. Xceivers: 1
  48. Last contact: Tue Oct 22 17:24:06 AEDT 2019

(2) 图形化监控

Goto: Yarn http://192.168.56.1:8088/

Goto: http://192.168.56.1:50070

Spark安装


一、安装方法

Goto: Install, Configure, and Run Spark on Top of a Hadoop YARN Cluster

Goto: https://anaconda.org/conda-forge/pyspark

    • hadoop-3.1.2.tar.gz
    • scala-2.12.10.deb
    • spark-2.4.4-bin-without-hadoop.tgz

二、一些可能的问题

Ref: 6 2 2Spark配置安装实验二:集群版

Ref: Spark multinode environment setup on yarn

Ref: SBT Error: “Failed to construct terminal; falling back to unsupported…”【.bashrc添加句柄】

Ref: Getting “cat: /release: No such file or directory” when running scala【使用高版本2.12.2+】

Ref: Using Spark's "Hadoop Free" Build【需要指定已安装hadoop的位置】

三、测试

四、远程 Notebook

Goto: Jupyter notebook远程访问服务器(实践出真知版)

End.

[Spark] 00 - Install Hadoop & Spark的更多相关文章

  1. Docker搭建大数据集群 Hadoop Spark HBase Hive Zookeeper Scala

    Docker搭建大数据集群 给出一个完全分布式hadoop+spark集群搭建完整文档,从环境准备(包括机器名,ip映射步骤,ssh免密,Java等)开始,包括zookeeper,hadoop,hiv ...

  2. install scala & spark env

    安装Scala 1,到http://www.scala-lang.org/download/ 下载与Spark版本对应的Scala.Spark1.2对应于Scala2.10的版本.这里下载scala- ...

  3. ubuntu下hadoop,spark配置

    转载来自:http://www.cnblogs.com/spark-china/p/3941878.html 在VMWare 中准备第二.第三台运行Ubuntu系统的机器:   在VMWare中构建第 ...

  4. hadoop+spark集群搭建入门

    忽略元数据末尾 回到原数据开始处 Hadoop+spark集群搭建 说明: 本文档主要讲述hadoop+spark的集群搭建,linux环境是centos,本文档集群搭建使用两个节点作为集群环境:一个 ...

  5. hadoop+spark集群搭建

    1.选取三台服务器(CentOS系统64位) 114.55.246.88 主节点 114.55.246.77 从节点 114.55.246.93 从节点 之后的操作如果是用普通用户操作的话也必须知道r ...

  6. WSL2+Ubuntu配置Java Maven Hadoop Spark环境

    所需文件: 更新日期为2021/5/8: Linux 内核更新包 JDK1.8 maven3.8.1 hadoop3.3.0 spark3.1.1 WSL?WSL2? WSL是适用于 Linux 的 ...

  7. Apache Spark 1.6 Hadoop 2.6 Mac下单机安装配置

    一. 下载资料 1. JDK 1.6 + 2. Scala 2.10.4 3. Hadoop 2.6.4 4. Spark 1.6 二.预先安装 1. 安装JDK 2. 安装Scala 2.10.4 ...

  8. 大数据项目实践:基于hadoop+spark+mongodb+mysql+c#开发医院临床知识库系统

    一.前言 从20世纪90年代数字化医院概念提出到至今的20多年时间,数字化医院(Digital Hospital)在国内各大医院飞速的普及推广发展,并取得骄人成绩.不但有数字化医院管理信息系统(HIS ...

  9. 哈,我自己翻译的小书,马上就完成了,是讲用python处理大数据框架hadoop,spark的

    花了一些时间, 但感觉很值得. Big Data, MapReduce, Hadoop, and Spark with Python Master Big Data Analytics and Dat ...

随机推荐

  1. VS引用文件出现黄色感叹号丢失文件,应该如何解决?

    VS是微软开发的一款超级强大的IDE,深受广大.net开发者喜爱. 但是再强大,也会有它的bug和缺点. 多人协同开发时,不知道你有没有遇到一个这样的情况:第二天上班,早早来到公司,打开电脑,拉取一下 ...

  2. Vue+springboot管理系统

    About 此项目是vue+element-ui 快速开发的物资管理系统,后台用的java springBoot 所有数据都是从服务器实时获取的数据,具有登陆,注册,对数据进行管理,打印数据等功能 说 ...

  3. sea.js的同步魔法

    前些时间也是想写点关于CMD模块规范的文字,以便帮助自己理解.今天看到一篇知乎回答,算是给了我一点启发. 同步写法却不阻塞? 先上一个sea.js很经典的模块写法: // 定义一个模块 define( ...

  4. Asp.NetCore源码学习[1-2]:配置[Option]

    Asp.NetCore源码学习[1-2]:配置[Option] 在上一篇文章中,我们知道了可以通过IConfiguration访问到注入的ConfigurationRoot,但是这样只能通过索引器IC ...

  5. javascript 基础知识汇总(一)

    1.<script> 标签 1) 可以通过<script> 标签将javaScript 代码添加到页面中 (type 和language 属性不是必须的) 2)外部的脚本可以通 ...

  6. 洛谷 P3627 【抢掠计划】

    题库:洛谷 题号:3627 题目:抢掠计划 link:https://www.luogu.org/problem/P3627 思路 : 这道题是一道Tarjan + 最长路的题.首先,我们用Tarja ...

  7. C++中的I/O输入输出问题

    C++ I/O navigation: 1.文件输入输出 2.string流 1.输入输出 C++语言不直接处理输入输出,而是通过一些标准库中类型.从设备(文件,控制台,内存)中读取数据,向设备中写入 ...

  8. 模板汇总——Tarjian

    1. 单向边  + 新图建边 int belong[N], dfn[N], low[N], now_time, scc_cnt; stack<int> s; void dfs(int u) ...

  9. codeforces 509 D. Restoring Numbers(数学+构造)

    题目链接:http://codeforces.com/problemset/problem/509/D 题意:题目给出公式w[i][j]= (a[i] + b[j])% k; 给出w,要求是否存在这样 ...

  10. codeforces 459 C. Pashmak and Buses(思维)

    题目链接:http://codeforces.com/problemset/problem/459/C 题意:有n个人,k辆车,d天要求没有两个人在d天都坐在一起.输出坐的方法. 题解:这题很有意思, ...