1、package org.apache.hadoop.mapred这四个协议都在这个包下。

2、从最简单的AdminOperationsProtocol看,

void refreshQueues() throws IOException;

void refreshNodes() throws IOException;

boolean setSafeMode(JobTracker.SafeModeAction safeModeAction)  throws IOException;

3、相对于hdfs里边那几个协议,这几个协议都经常修改,不是很稳定

InterTrackerProtocol

HeartbeatResponse heartbeat(TaskTrackerStatus status,
                            boolean restarted,
                            boolean initialContact,
                            boolean acceptNewTasks,
                            short responseId)
  throws IOException;

public String getFilesystemName() throws IOException;

public void reportTaskTrackerError(String taskTracker,
                                   String errorClass,
                                   String errorMessage) throws IOException;

TaskCompletionEvent[] getTaskCompletionEvents(JobID jobid, int fromEventId
    , int maxEvents) throws IOException;

public String getSystemDir();

public String getBuildVersion() throws IOException;

public String getVIVersion() throws IOException;

JobSubmissionProtocol

public JobID getNewJobId() throws IOException;

public JobStatus submitJob(JobID jobName, String jobSubmitDir, Credentials ts)
throws IOException;

public ClusterStatus getClusterStatus(boolean detailed) throws IOException;

public AccessControlList getQueueAdmins(String queueName) throws IOException;

public void killJob(JobID jobid) throws IOException;

public void setJobPriority(JobID jobid, String priority)
                                                    throws IOException;

public boolean killTask(TaskAttemptID taskId, boolean shouldFail) throws IOException;

public JobProfile getJobProfile(JobID jobid) throws IOException;

public JobStatus getJobStatus(JobID jobid) throws IOException;

public Counters getJobCounters(JobID jobid) throws IOException;

public TaskReport[] getMapTaskReports(JobID jobid) throws IOException;

public TaskReport[] getReduceTaskReports(JobID jobid) throws IOException;

public TaskReport[] getCleanupTaskReports(JobID jobid) throws IOException;

public TaskReport[] getSetupTaskReports(JobID jobid) throws IOException;

public String getFilesystemName() throws IOException;

public JobStatus[] jobsToComplete() throws IOException;

public JobStatus[] getAllJobs() throws IOException;

public TaskCompletionEvent[] getTaskCompletionEvents(JobID jobid
    , int fromEventId, int maxEvents) throws IOException;

public String[] getTaskDiagnostics(TaskAttemptID taskId)
throws IOException;
  public String getSystemDir();

public String getStagingAreaDir() throws IOException;

public JobQueueInfo[] getQueues() throws IOException;

public JobQueueInfo getQueueInfo(String queue) throws IOException;

public JobStatus[] getJobsFromQueue(String queue) throws IOException;

public QueueAclsInfo[] getQueueAclsForCurrentUser() throws IOException;

public
Token<DelegationTokenIdentifier> getDelegationToken(Text renewer
                                                    ) throws IOException,
                                                        InterruptedException;

public long renewDelegationToken(Token<DelegationTokenIdentifier> token
                                 ) throws IOException,
                                          InterruptedException;

public void cancelDelegationToken(Token<DelegationTokenIdentifier> token
                                  ) throws IOException,InterruptedException;

TaskUmbilicalProtocol

JvmTask getTask(JvmContext context) throws IOException;

boolean statusUpdate(TaskAttemptID taskId, TaskStatus taskStatus,
    JvmContext jvmContext) throws IOException, InterruptedException;

void reportDiagnosticInfo(TaskAttemptID taskid, String trace,
    JvmContext jvmContext) throws IOException;

void reportNextRecordRange(TaskAttemptID taskid, SortedRanges.Range range,
    JvmContext jvmContext) throws IOException;

boolean ping(TaskAttemptID taskid, JvmContext jvmContext) throws IOException;

void done(TaskAttemptID taskid, JvmContext jvmContext) throws IOException;
void commitPending(TaskAttemptID taskId, TaskStatus taskStatus,
    JvmContext jvmContext) throws IOException, InterruptedException;

boolean canCommit(TaskAttemptID taskid, JvmContext jvmContext) throws IOException;

void shuffleError(TaskAttemptID taskId, String message, JvmContext jvmContext)
    throws IOException;

void fsError(TaskAttemptID taskId, String message, JvmContext jvmContext)
    throws IOException;

void fatalError(TaskAttemptID taskId, String message, JvmContext jvmContext)
    throws IOException;

MapTaskCompletionEventsUpdate getMapCompletionEvents(JobID jobId,
                                                     int fromIndex,
                                                     int maxLocs,
                                                     TaskAttemptID id,
                                                     JvmContext jvmContext)
throws IOException;

void updatePrivateDistributedCacheSizes(org.apache.hadoop.mapreduce.JobID jobId,
                                        long[] sizes) throws IOException;

从协议VersionedProtocol开始4——AdminOperationsProtocol、InterTrackerProtocol、JobSubmissionProtocol、TaskUmbilicalProtocol的更多相关文章

  1. 从协议VersionedProtocol开始

    VersionedProtocol协议是Hadoop的最顶层协议接口的抽象:5--3--3共11个协议,嘿嘿 1)HDFS相关 ClientDatanodeProtocol:client与datano ...

  2. 从协议VersionedProtocol开始3——ClientProtocol、DatanodeProtocol、NamenodeProtocol、RefreshAuthorizationPolicyProtocol、RefreshUserMappingsProtocol

    1.ClientProtocol这个玩意的版本号是61L:DatanodeProtocol 是26L:NamenodeProtocol是 3L;RefreshAuthorizationPolicyPr ...

  3. 从协议VersionedProtocol开始2——ClientDatanodeProtocol和InterDatanodeProtocol

    1.首先,我看的是hadoop1.2.1 这个里边,有点奇怪ClientDatanodeProtocol的versionID是4,但是InterDatanodeProtocol的versionID是3 ...

  4. 从协议VersionedProtocol开始1

    Phase 0: Make a plan You must first decide what steps you're going to have in your process. It sound ...

  5. Hadoop系列番外篇之一文搞懂Hadoop RPC框架及细节实现

    @ 目录 Hadoop RPC 框架解析 1.Hadoop RPC框架概述 1.1 RPC框架特点 1.2 Hadoop RPC框架 2.Java基础知识回顾 2.1 Java反射机制与动态代理 2. ...

  6. MapReduce剖析笔记之四:TaskTracker通过心跳机制获取任务的流程

    上一节分析到了JobTracker把作业从队列里取出来并进行了初始化,所谓的初始化,主要是获取了Map.Reduce任务的数量,并统计了哪些DataNode所在的服务器可以处理哪些Split等等,将这 ...

  7. MapReduce剖析笔记之二:Job提交的过程

    上一节以WordCount分析了MapReduce的基本执行流程,但并没有从框架上进行分析,这一部分工作在后续慢慢补充.这一节,先剖析一下作业提交过程. 在分析之前,我们先进行一下粗略的思考,如果要我 ...

  8. hadoop之JobTracker功能分析

    JobTracker是整个MapReduce计算框架中的主服务,相当于集群的“管理者”,负责整个集群的作业控制和资源管理.本文对JobTracker的启动过程及心跳接收与应答两个主要功能进行分析. 1 ...

  9. 【Hadoop代码笔记】通过JobClient对Jobtracker的调用详细了解Hadoop RPC

    Hadoop的各个服务间,客户端和服务间的交互采用RPC方式.关于这种机制介绍的资源很多,也不难理解,这里不做背景介绍.只是尝试从Jobclient向JobTracker提交作业这个最简单的客户端服务 ...

随机推荐

  1. Maven——eclipse中使用Maven创建Web项目

    原文:http://www.cnblogs.com/xdp-gacl/p/4054814.html 一.创建Web项目 1.1 选择建立Maven Project 选择File -> New - ...

  2. Java Date与SimpleDateFormat

    最近在弄一些涉及到时间处理的项目.本来自己写了一个时间转换函数,虽然能用但是过于麻烦而且不够规范,于是学习了下java自带的时间处理的类. public class Timechg { public ...

  3. WCF配置详解

    前面一篇文章<WCF 学习总结1 -- 简单实例> 一股脑儿展示了几种WCF部署方式,其中配置文件(App.config/Web.config)都是IDE自动生成,省去了我们不少功夫.现在 ...

  4. window下安装nodejs

    下载nodejs 去https://nodejs.org/en/download/下载nodejs,有.mis(安装版)和.exe(二进制版) .mis(安装版) 一般下载这个就行,简单方便,自带np ...

  5. Linux下安装最新的Eclipse

    欢迎关注我的社交账号: 博客园地址: http://www.cnblogs.com/jiangxinnju/p/4781259.html GitHub地址: https://github.com/ji ...

  6. laravel实现第三方登录

    https://github.com/laravel/socialite 这是官方的第三方登录包,支持很多国外的第三方登录 https://github.com/overtrue/socialite ...

  7. 微信JS SDK Demo 官方案例

    微信JS-SDK是微信公众平台面向网页开发者提供的基于微信内的网页开发工具包. 通过使用微信JS-SDK,网页开发者可借助微信高效地使用拍照.选图.语音.位置等手机系统的能力,同时可以直接使用微信分享 ...

  8. 全面理解面向对象的 JavaScript

    前言 当今 JavaScript 大行其道,各种应用对其依赖日深.web 程序员已逐渐习惯使用各种优秀的 JavaScript 框架快速开发 Web 应用,从而忽略了对原生 JavaScript 的学 ...

  9. ubuntu修改更新源为163

    安装ubuntu后为了使下载及更新软件包更快,需要将自带的源更换为国内的源.以下以163为例: 1.首先备份下原始自带源: sudo cp sources.list /etc/apt/sources. ...

  10. fork&exec

    进程是系统进行资源分配和调度的基本单位,包括代码.数据和PCB进程控制块等资源. fork函数通过系统调用创建一个与原进程相同的子进程. 在调用进程(父进程)中返回一次,返回子进程ID:在子进程返回0 ...