本文总结了Hadoop生态系统中各个组件使用的端口,包括了HDFS,Map Reduce,HBase,Hive,Spark,WebHCat,Impala,Alluxio,Sqoop等,后续会持续更新。

HDFS Ports

Service

Servers

Default Ports Used

Protocol

Description

Need End User Access?

Configuration Parameters

NameNode WebUI

Master Nodes (NameNode and any back-up NameNodes)

http

Web UI to look at current status of HDFS, explore file system

Yes (Typically admins, Dev/Support teams)

dfs.http.address

https

Secure http service

dfs.https.address

NameNode metadata service

Master Nodes (NameNode and any back-up NameNodes)

8020/9000

IPC

File system metadata operations

Yes (All clients who directly need to interact with the HDFS)

Embedded in URI specified by fs.default.name

DataNode

All Slave Nodes

http

DataNode WebUI to access the status, logs etc.

Yes (Typically admins, Dev/Support teams)

dfs.datanode.http.address

https

Secure http service

dfs.datanode.https.address

 

Data transfer

 

dfs.datanode.address

IPC

Metadata operations

No

dfs.datanode.ipc.address

Secondary NameNode

Secondary NameNode and any backup Secondary NameNode

http

Checkpoint for NameNode metadata

No

dfs.secondary.http.address

Map Reduce Ports:

Service

Servers

Default Ports Used

Protocol

Description

Need End User Access?

Configuration Parameters

JobTracker  WebUI

Master Nodes (JobTracker Node and any back-up Job­Tracker node )

http

Web UI for JobTracker

Yes

mapred.job.tracker.http.address

JobTracker

Master Nodes (JobTracker Node)

IPC

For job submissions

Yes (All clients who need to submit the MapReduce jobs  including Hive, Hive server, Pig)

Embedded in URI specified by mapred.job.tracker

Task­Tracker Web UI and Shuffle

All Slave Nodes

http

DataNode Web UI to access status, logs, etc.

Yes (Typically admins, Dev/Support teams)

mapred.task.tracker.http.address

History Server WebUI

 

http

Web UI for Job History

Yes

mapreduce.history.server.http.address

HBase Ports:

Service

Servers

Default Ports Used

Protocol

Description

Need End User Access?

Configuration Parameters

HMaster

Master Nodes (HBase Master Node and any back-up HBase Master node)

   

Yes

hbase.master.port

HMaster Info Web UI

Master Nodes (HBase master Node and back up HBase Master node if any)

http

The port for the HBase­Master web UI. Set to -1 if you do not want the info server to run.

Yes

hbase.master.info.port

Region Server

All Slave Nodes

   

Yes (Typically admins, dev/support teams)

hbase.regionserver.port

Region Server

All Slave Nodes

http

 

Yes (Typically admins, dev/support teams)

hbase.regionserver.info.port

 

All ZooKeeper Nodes

 

Port used by ZooKeeper peers to talk to each other.Seehere for more information.

No

hbase.zookeeper.peerport

 

All ZooKeeper Nodes

 

Port used by ZooKeeper peers to talk to each other.Seehere for more information.

 

hbase.zookeeper.leaderport

     

Property from ZooKeeper's config zoo.cfg. The port at which the clients will connect.

 

hbase.zookeeper.property.clientPort

Hive Ports:

Service

Servers

Default Ports Used

Protocol

Description

Need End User Access?

Configuration Parameters

Hive Server2

Hive Server machine (Usually a utility machine)

thrift

Service for programatically (Thrift/JDBC) connecting to Hive

Yes (Clients who need to connect to Hive either programatically or through UI SQL tools that use JDBC)

ENV Variable HIVE_PORT

Hive Metastore

 

thrift

 

Yes (Clients that run Hive, Pig and potentially M/R jobs that use HCatalog)

hive.metastore.uris

WebHCat Ports:

Service

Servers

Default Ports Used

Protocol

Description

Need End User Access?

WebHCat Server

Any utility machine

http

Web API on top of HCatalog and other Hadoop services

Yes

Spark Ports:

Service

Servers

Default Ports Used

Description

Spark GUI

Nodes running spark

Spark web interface for monitoring and troubleshooting

Impala Ports:

Service

Servers

Default Ports Used

Description

Impala Daemon

Nodes running impala daemon

Used by transmit commands and receive results by impala-shell

Impala Daemon

Nodes running impala daemon

Used by applications through JDBC

Impala Daemon

Nodes running impala daemon

Impala web interface for monitoring and troubleshooting

Impala StateStore Daemon

Nodes running impala StateStore daemon

StateStore web interface for monitoring and troubleshooting

Impala Catalog Daemon

Nodes running impala catalog daemon

Catalog service web interface for monitoring and troubleshooting

Alluxio Ports:

Service

Servers

Default Ports Used

Protocol

Description

Need End User Access?

Alluxio Web GUI

Any utility machine

http

Web GUI to check alluxio status

Yes

Alluxio API

Any utility machine

Tcp

Api to access data on alluxio

No

Sqoop Ports:

Service

Servers

Default Ports Used

Description

Sqoop server

Nodes running Sqoop

Used by Sqoop client to access the sqoop server

Hadoop Ecosystem related ports的更多相关文章

  1. Hadoop ecosystem notes Outline - TODO

    Motivation Sometimes I fell like giving up, then I remember I have a lot of motherfuckers to prove w ...

  2. Hadoop ecosystem

    How did it all start- huge data on the web! Nutch built to crawl this web data Huge data had to save ...

  3. Hadoop ecosystem 生态圈

    Cascading: hadoop上面的workflow Sqoop(发音:skup)是一款开源的工具,主要用于在Hadoop(Hive)与传统的数据库(mysql.postgresql...)间进行 ...

  4. hadoop发行版本

    Azure HDInsight Azure HDInsight is Microsoft's distribution of Hadoop. The Azure HDInsight ecosystem ...

  5. Hadoop HDFS 用户指南

    This document is a starting point for users working with Hadoop Distributed File System (HDFS) eithe ...

  6. 关于hadoop

    hadoop 是什么? 1. 适合海量数据的分布式存储与计算平台. 海量: 是指 1T 以上数据. 分布式: 任务分配到多态虚拟机上进行计算. 2. 多个任务是怎么被分配到多个虚拟机当中的? 分配是需 ...

  7. 使用Windows Azure的VM安装和配置CDH搭建Hadoop集群

    本文主要内容是使用Windows Azure的VIRTUAL MACHINES和NETWORKS服务安装CDH (Cloudera Distribution Including Apache Hado ...

  8. Hadoop入门进阶课程10--HBase介绍、安装与应用案例

    本文版权归作者和博客园共有,欢迎转载,但未经作者同意必须保留此段声明,且在文章页面明显位置给出原文连接,博主为石山园,博客地址为 http://www.cnblogs.com/shishanyuan  ...

  9. [Hadoop 周边] Hadoop技术生态圈

    Hadoop版本演进 当前Hadoop有两大版本:Hadoop 1.0和Hadoop 2.0. Hadoop1.0被称为第一代Hadoop,由分布式文件系统HDFS和分布式计算框架MapReduce组 ...

随机推荐

  1. Ubuntu 16.04 安装jdk

    Ubuntu 16.04 安装jdk 准备工作 安装版本:jdk-8u91-linux-x64.tar.gz 官方下载 创建目录作为JDK的安装目录,这里选择安装位置为:/usr/java/ sudo ...

  2. DELPHI XE5 UP2 无真机输出 APP并转换为IPA(实践整理)

    1.在Mac上配置开发环境(具体步骤请百度)   XCODE5.1+IOS7.1SDK+COMMAND LINE TOOLS   安装PlatformAssistant   买一个真机调试账号(实际测 ...

  3. Newtonsoft.Json.Linq

    var json = "{\"name\":\"ok1\",\"sex\":\"man\"}"; / ...

  4. if 判断

    语法一: if 条件: #条件成立时执行的字代码块 代码1 代码2 代码3 示例: sex='female' age=18 is_beautiful=True if sex == 'female' a ...

  5. wcf文件上传时碰到的配置问题

    1.远程服务器返回了意外相应:(413) Request Entity Too Large 修改客户端配置maxReceivedMessageSize="2147483647" & ...

  6. 数组Byte [] 和 string 相互转换

    using System; using System.Collections.Generic; using System.Text; namespace NET.MST.Fourth.StringBy ...

  7. C#在线运行--cmd方法

       此次C#在线运行采用cmd.exe用csc对文件进行编译,然后再运行的思路实现在线运行的效果.不过会生成二个文件(.cs和.exe),可能需要定期清除临时文件夹. 首先利用时间戳生成唯一文件名, ...

  8. 如何彻底删除TFS上的团队项目 For VS 2017

    参考 Visual Studio 2017 TFSDeleteProject.exe 位置 X:\Program Files (x86)\Microsoft Visual Studio\2017\En ...

  9. php代码审计6审计xss漏洞

    跨站脚本攻击(Cross Site Scripting)是指攻击者利用网站程序对用户输入过滤不足,输入可以显示在页面上对其他用户造成影响的html代码,从而盗取用户资料,利用用户身份进行某种动作或者对 ...

  10. 文件参数化-utp框架之根据yaml文件自动生成python文件+utp运行用例

    根据yaml文件自动生成python文件 utp框架: bin目录:存放执行文件(run.py) cases目录:存放生成的用例的python文件(该目录下的文件为根据data目录下的测试用例生成的p ...