Hadoop Ecosystem related ports
本文总结了Hadoop生态系统中各个组件使用的端口,包括了HDFS,Map Reduce,HBase,Hive,Spark,WebHCat,Impala,Alluxio,Sqoop等,后续会持续更新。
HDFS Ports:
|
Service |
Servers |
Default Ports Used |
Protocol |
Description |
Need End User Access? |
Configuration Parameters |
|
NameNode WebUI |
Master Nodes (NameNode and any back-up NameNodes) |
http |
Web UI to look at current status of HDFS, explore file system |
Yes (Typically admins, Dev/Support teams) |
dfs.http.address |
|
|
https |
Secure http service |
dfs.https.address |
||||
|
NameNode metadata service |
Master Nodes (NameNode and any back-up NameNodes) |
8020/9000 |
IPC |
File system metadata operations |
Yes (All clients who directly need to interact with the HDFS) |
Embedded in URI specified by fs.default.name |
|
DataNode |
All Slave Nodes |
http |
DataNode WebUI to access the status, logs etc. |
Yes (Typically admins, Dev/Support teams) |
dfs.datanode.http.address |
|
|
https |
Secure http service |
dfs.datanode.https.address |
||||
|
Data transfer |
dfs.datanode.address |
|||||
|
IPC |
Metadata operations |
No |
dfs.datanode.ipc.address |
|||
|
Secondary NameNode |
Secondary NameNode and any backup Secondary NameNode |
http |
Checkpoint for NameNode metadata |
No |
dfs.secondary.http.address |
Map Reduce Ports:
|
Service |
Servers |
Default Ports Used |
Protocol |
Description |
Need End User Access? |
Configuration Parameters |
|
JobTracker WebUI |
Master Nodes (JobTracker Node and any back-up JobTracker node ) |
http |
Web UI for JobTracker |
Yes |
mapred.job.tracker.http.address |
|
|
JobTracker |
Master Nodes (JobTracker Node) |
IPC |
For job submissions |
Yes (All clients who need to submit the MapReduce jobs including Hive, Hive server, Pig) |
Embedded in URI specified by mapred.job.tracker |
|
|
TaskTracker Web UI and Shuffle |
All Slave Nodes |
http |
DataNode Web UI to access status, logs, etc. |
Yes (Typically admins, Dev/Support teams) |
mapred.task.tracker.http.address |
|
|
History Server WebUI |
http |
Web UI for Job History |
Yes |
mapreduce.history.server.http.address |
HBase Ports:
|
Service |
Servers |
Default Ports Used |
Protocol |
Description |
Need End User Access? |
Configuration Parameters |
|
HMaster |
Master Nodes (HBase Master Node and any back-up HBase Master node) |
Yes |
hbase.master.port |
|||
|
HMaster Info Web UI |
Master Nodes (HBase master Node and back up HBase Master node if any) |
http |
The port for the HBaseMaster web UI. Set to -1 if you do not want the info server to run. |
Yes |
hbase.master.info.port |
|
|
Region Server |
All Slave Nodes |
Yes (Typically admins, dev/support teams) |
hbase.regionserver.port |
|||
|
Region Server |
All Slave Nodes |
http |
Yes (Typically admins, dev/support teams) |
hbase.regionserver.info.port |
||
|
All ZooKeeper Nodes |
Port used by ZooKeeper peers to talk to each other.Seehere for more information. |
No |
hbase.zookeeper.peerport |
|||
|
All ZooKeeper Nodes |
Port used by ZooKeeper peers to talk to each other.Seehere for more information. |
hbase.zookeeper.leaderport |
||||
|
Property from ZooKeeper's config zoo.cfg. The port at which the clients will connect. |
hbase.zookeeper.property.clientPort |
Hive Ports:
|
Service |
Servers |
Default Ports Used |
Protocol |
Description |
Need End User Access? |
Configuration Parameters |
|
Hive Server2 |
Hive Server machine (Usually a utility machine) |
thrift |
Service for programatically (Thrift/JDBC) connecting to Hive |
Yes (Clients who need to connect to Hive either programatically or through UI SQL tools that use JDBC) |
ENV Variable HIVE_PORT |
|
|
Hive Metastore |
thrift |
Yes (Clients that run Hive, Pig and potentially M/R jobs that use HCatalog) |
hive.metastore.uris |
WebHCat Ports:
|
Service |
Servers |
Default Ports Used |
Protocol |
Description |
Need End User Access? |
|
WebHCat Server |
Any utility machine |
http |
Web API on top of HCatalog and other Hadoop services |
Yes |
Spark Ports:
|
Service |
Servers |
Default Ports Used |
Description |
|
Spark GUI |
Nodes running spark |
Spark web interface for monitoring and troubleshooting |
Impala Ports:
|
Service |
Servers |
Default Ports Used |
Description |
|
Impala Daemon |
Nodes running impala daemon |
Used by transmit commands and receive results by impala-shell |
|
|
Impala Daemon |
Nodes running impala daemon |
Used by applications through JDBC |
|
|
Impala Daemon |
Nodes running impala daemon |
Impala web interface for monitoring and troubleshooting |
|
|
Impala StateStore Daemon |
Nodes running impala StateStore daemon |
StateStore web interface for monitoring and troubleshooting |
|
|
Impala Catalog Daemon |
Nodes running impala catalog daemon |
Catalog service web interface for monitoring and troubleshooting |
Alluxio Ports:
|
Service |
Servers |
Default Ports Used |
Protocol |
Description |
Need End User Access? |
|
Alluxio Web GUI |
Any utility machine |
http |
Web GUI to check alluxio status |
Yes |
|
|
Alluxio API |
Any utility machine |
Tcp |
Api to access data on alluxio |
No |
Sqoop Ports:
|
Service |
Servers |
Default Ports Used |
Description |
|
Sqoop server |
Nodes running Sqoop |
Used by Sqoop client to access the sqoop server |
Hadoop Ecosystem related ports的更多相关文章
- Hadoop ecosystem notes Outline - TODO
Motivation Sometimes I fell like giving up, then I remember I have a lot of motherfuckers to prove w ...
- Hadoop ecosystem
How did it all start- huge data on the web! Nutch built to crawl this web data Huge data had to save ...
- Hadoop ecosystem 生态圈
Cascading: hadoop上面的workflow Sqoop(发音:skup)是一款开源的工具,主要用于在Hadoop(Hive)与传统的数据库(mysql.postgresql...)间进行 ...
- hadoop发行版本
Azure HDInsight Azure HDInsight is Microsoft's distribution of Hadoop. The Azure HDInsight ecosystem ...
- Hadoop HDFS 用户指南
This document is a starting point for users working with Hadoop Distributed File System (HDFS) eithe ...
- 关于hadoop
hadoop 是什么? 1. 适合海量数据的分布式存储与计算平台. 海量: 是指 1T 以上数据. 分布式: 任务分配到多态虚拟机上进行计算. 2. 多个任务是怎么被分配到多个虚拟机当中的? 分配是需 ...
- 使用Windows Azure的VM安装和配置CDH搭建Hadoop集群
本文主要内容是使用Windows Azure的VIRTUAL MACHINES和NETWORKS服务安装CDH (Cloudera Distribution Including Apache Hado ...
- Hadoop入门进阶课程10--HBase介绍、安装与应用案例
本文版权归作者和博客园共有,欢迎转载,但未经作者同意必须保留此段声明,且在文章页面明显位置给出原文连接,博主为石山园,博客地址为 http://www.cnblogs.com/shishanyuan ...
- [Hadoop 周边] Hadoop技术生态圈
Hadoop版本演进 当前Hadoop有两大版本:Hadoop 1.0和Hadoop 2.0. Hadoop1.0被称为第一代Hadoop,由分布式文件系统HDFS和分布式计算框架MapReduce组 ...
随机推荐
- vmware10安装Arch
必须保证机器能够上网! 1.vmware10中创建虚拟机(跟虚拟机中安装其他系统同样的操作). 2.开启上一步中创建的虚拟机. 3选择第一项进入 4自动进入root命令行 5进入 /dev (进行分区 ...
- Alpha项目复审
队名 优点 缺点 排名 拉登是我罩的 1.最底层.从无到有实现的软硬件结合的俄罗斯方块游戏. 2.从画电路原理图.PCB电路板设计.接线.操作系统(没用到操作系统).驱动程序.应用程序开发,串口通信. ...
- SNMP协议学习笔记
什么是SNMP协议? SNMP协议是以UDP为基础的应用层协议,全称为 简单网络管理协议,用于网络管理系统与被管设备(路由器,交换机,服务器等设备)进行通信. 应用场景 随着网络设备的增多,需要单独的 ...
- query聚类技术
query聚类 目的 query聚类主要有以下两个目的 解决query空间稀疏问题(长尾query) 挖掘用户意图(一条行为包含的意图是稀疏的,当有一簇行为时,意图更明确) 可以说聚类是构建内容模型的 ...
- 关于使用idea的一些小技巧
1:idea与git同步以后查看修改变化: file --setting--versioncontorller
- 【连载】redis库存操作,分布式锁的四种实现方式[二]--基于Redisson实现分布式锁
一.redisson介绍 redisson实现了分布式和可扩展的java数据结构,支持的数据结构有:List, Set, Map, Queue, SortedSet, ConcureentMap, L ...
- 搭建linux虚拟机
一.VMware 9 安装CentOS 7 -> 创建新虚拟机, 选择稍后安装操作系统 -> 选择Linux系统Centos 64位 -> 填写虚拟机名称, 选择虚拟机安 ...
- pdo + 事务处理 处理线性事务
/* * 事物处理线性操作. * 以转账为例 */ header('Content-type:text/html;charset=utf-8'); $opt = array(PDO::ATTR_PER ...
- C# 密封
到目前位置所说的都是让类如何如何进行继承啊 ,重写啊,巴不得类有十多个继承,超级多的重写. 但是,今天我们来说说不允许继承和不允许重写! 这个不允许继承是包括类和方法. 这种情况好比: 爸爸有私房钱, ...
- 汇编工具安装一:MASM32的安装!
MASM32是一款汇编开发工具,网址:http://www.masm32.com/ ,它是集合了很多开发汇编程序的工具,windows下汇编程序的开发,很多人都在使用它. 直接下载:http://we ...