环境篇:CM+CDH6.3.2环境搭建(全网最全)
环境篇:CM+CDH6.3.2环境搭建(全网最全)
一 环境准备
1.1 三台虚拟机准备
Master( 32g内存 + 100g硬盘 + 4cpu + 每个cpu2核)
2台Slave( 12g内存 + 100g硬盘 + 4cpu + 每个cpu1核)
- 参考地址:
https://www.cnblogs.com/ttzzyy/p/12566281.html
1、准备的机器只要网络IP不冲突通并且可以正常访问网络即可,如机器资源不够,可自行分配所需要的资源,或者使用云服务
2、关闭防火墙,云服务还需要开启对应的IP端口
3、机器hostname修改,配置hosts,并配置SHH免密
4、时间同步
注意:如果拿不稳可以做快照的做快照,这样恢复起来容易。
1.2 常用yum源更新,gcc,G++,C++等环境(可以跳过)
yum -y install chkconfig python bind-utils psmisc libxslt zlib sqlite cyrus-sasl-plain cyrus-sasl-gssapi fuse fuse-libs redhat-lsb postgresql* portmap mod_ssl openssl openssl-devel python-psycopg2 MySQL-python python-devel telnet pcre-devel gcc gcc-c++
1.3 配置本地yum云
注意此处只需要在主节点机器上执行
1.3.1 更新yum源httpd
#更新yum源httpd
yum -y install httpd
#查看httpd状态
systemctl status httpd.service
#启动httpd
service httpd start
#配置httpd永久生效(重启生效)
chkconfig httpd on
1.3.2 更新yum源yum-utils
#更新yum源yum-utils createrepo
yum -y install yum-utils createrepo
#进入yum源路径
cd /var/www/html/
#创建cm文件夹
mkdir cm
1.3.3 将下载好的资源添加进本地yum云
在“2.1CM下载”中,下载需要的文件,后面要放在本地yum云中,提供访问
在添加完需要资源后,使用工具和本地的yum源创建联系,其中repodata是yum联系创建的
createrepo /var/www/html/cm/
1.3.4 创建本地repo文件
vim /etc/yum.repos.d/cloudera-manager.repo
---->添加如下内容(注意IP改成本地yum主机IP,只能是ip)
[cloudera-manager]
name=Cloudera Manager, Version yum
baseurl=http://192.168.37.10/cm
gpgcheck=0
enabled=1
----<
- 更新yum源
yum clean all
yum makecache
- 验证
yum list | grep cloudera-manager
1.3.5 将本地yum文件分发至从节点
scp -r /etc/yum.repos.d/cloudera-manager.repo root@cdh02.cm:/etc/yum.repos.d/
scp -r /etc/yum.repos.d/cloudera-manager.repo root@cdh03.cm:/etc/yum.repos.d/
1.4 安装JDK(所有节点)
#查看命令
rpm -qa | grep java
#删除命令(删除所有openjdk)
rpm -e --nodeps xxx
- 将oracle-j2sdk1.8-1.8.0+update181-1.x86_64.rpm上传至每个节点安装
rpm -ivh oracle-j2sdk1.8-1.8.0+update181-1.x86_64.rpm
- 修改配置文件
vim /etc/profile
--->添加
export JAVA_HOME=/usr/java/jdk1.8.0_181-cloudera
export PATH=$JAVA_HOME/bin:$PATH
export CLASSPATH=.:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar
---<
- 刷新源
source /etc/profile
- 检验
java
javac
1.5 安装mysql
用于管理CM数据,故安装在主节点即可
- 此处先把mysql-connector-java-5.1.47.jar传到每台服务器的/usr/share/java(如果目录不存在需要先创建)目录下,并更名为mysql-connector-java.jar,如果不更名后面会报错
- 把mysql-5.7.11-linux-glibc2.5-x86_64.tar.gz解压到/usr/local目录下并改名为mysql,然后执行如下命令
#创建归档,数据,临时文件夹
mkdir /usr/local/mysql/arch /usr/local/mysql/data /usr/local/mysql/tmp
#创建用户,赋予权限
useradd mysql
chown -R mysql.mysql /usr/local/mysql
#更新yum
yum -y install perl perl-devel
- 修改vi /etc/my.cnf
删除原有配置,增加如下配置
[client]
port=3306
socket=/usr/local/mysql/data/mysql.sock
default-character-set=utf8mb4
[mysqld]
port=3306
socket=/usr/local/mysql/data/mysql.sock
skip-slave-start
skip-external-locking
key_buffer_size=256M
sort_buffer_size=2M
read_buffer_size=2M
read_rnd_buffer_size=4M
query_cache_size=32M
max_allowed_packet=16M
myisam_sort_buffer_size=128M
tmp_table_size=32M
table_open_cache=512
thread_cache_size=8
wait_timeout=86400
interactive_timeout=86400
max_connections=600
# Try number of CPU's*2 for thread_concurrency
#thread_concurrency=32
#isolation level and default engine
default-storage-engine=INNODB
transaction-isolation=READ-COMMITTED
server-id=1739
basedir=/usr/local/mysql
datadir=/usr/local/mysql/data
pid-file=/usr/local/mysql/data/hostname.pid
#open performance schema
log-warnings
sysdate-is-now
binlog_format=ROW
log_bin_trust_function_creators=1
log-error=/usr/local/mysql/data/hostname.err
log-bin=/usr/local/mysql/arch/mysql-bin
expire_logs_days=7
innodb_write_io_threads=16
relay-log=/usr/local/mysql/relay_log/relay-log
relay-log-index=/usr/local/mysql/relay_log/relay-log.index
relay_log_info_file=/usr/local/mysql/relay_log/relay-log.info
log_slave_updates=1
gtid_mode=OFF
enforce_gtid_consistency=OFF
# slave
slave-parallel-type=LOGICAL_CLOCK
slave-parallel-workers=4
master_info_repository=TABLE
relay_log_info_repository=TABLE
relay_log_recovery=ON
#other logs
#general_log=1
#general_log_file=/usr/local/mysql/data/general_log.err
#slow_query_log=1
#slow_query_log_file=/usr/local/mysql/data/slow_log.err
#for replication slave
sync_binlog=500
#for innodb options
innodb_data_home_dir=/usr/local/mysql/data/
innodb_data_file_path=ibdata1:1G;ibdata2:1G:autoextend
innodb_log_group_home_dir=/usr/local/mysql/arch
innodb_log_files_in_group=4
innodb_log_file_size=1G
innodb_log_buffer_size=200M
#根据生产需要,调整pool size
innodb_buffer_pool_size=2G
#innodb_additional_mem_pool_size=50M #deprecated in 5.6
tmpdir=/usr/local/mysql/tmp
innodb_lock_wait_timeout=1000
#innodb_thread_concurrency=0
innodb_flush_log_at_trx_commit=2
innodb_locks_unsafe_for_binlog=1
#innodb io features: add for mysql5.5.8
performance_schema
innodb_read_io_threads=4
innodb-write-io-threads=4
innodb-io-capacity=200
#purge threads change default(0) to 1 for purge
innodb_purge_threads=1
innodb_use_native_aio=on
#case-sensitive file names and separate tablespace
innodb_file_per_table=1
lower_case_table_names=1
[mysqldump]
quick
max_allowed_packet=128M
[mysql]
no-auto-rehash
default-character-set=utf8mb4
[mysqlhotcopy]
interactive-timeout
[myisamchk]
key_buffer_size=256M
sort_buffer_size=256M
read_buffer=2M
write_buffer=2M
- 配置服务及开机自启动
cd /usr/local/mysql
#将服务文件拷贝到init.d下,并重命名为mysql
cp support-files/mysql.server /etc/rc.d/init.d/mysql
#赋予可执行权限
chmod +x /etc/rc.d/init.d/mysql
#删除服务
chkconfig --del mysql
#添加服务
chkconfig --add mysql
chkconfig --level 345 mysql on
#添加快捷方式
ln -s /usr/local/mysql/bin/mysql /usr/bin/
#检查服务
chkconfig --list | grep mysql
- 安装mysql的初始db
#输入如下命令等待
/usr/local/mysql/bin/mysqld \
--defaults-file=/etc/my.cnf \
--user=mysql \
--basedir=/usr/local/mysql/ \
--datadir=/usr/local/mysql/data/ \
--initialize
在初始化时如果加上 –initial-insecure,则会创建空密码的 root@localhost 账号,否则会创建带密码的 root@localhost 账号,密码直接写在 log-error 日志文件中
(在5.6版本中是放在 ~/.mysql_secret 文件里,更加隐蔽,不熟悉的话可能会无所适从)
- 启动
#下面命令执行完记得敲回车
/usr/local/mysql/bin/mysqld_safe --defaults-file=/etc/my.cnf &
#启动mysql(如果上面命令没有启动mysql),停止为service mysql stop
#service mysql start
#查看mysql状态
service mysql status
- 查看临时密码
cat /usr/local/mysql/data/hostname.err |grep password
- 登录及修改用户密码
mysql -u root -p
#输入刚刚日志中的密码进入
#设置密码为:root
set password for 'root'@'localhost'=password('root');
#配置远程可以访问
grant all privileges on *.* to 'root'@'%' identified by 'root' with grant option;
use mysql
delete from user where host!='%';
#刷新
flush privileges;
#退出
quit
- 登录创建cm需要的库的用户
服务名 | 数据库名 | 用户名 |
---|---|---|
Cloudera Manager Server | scm | scm |
Activity Monitor | amon | amon |
Reports Manager | rman | rman |
Hue | hue | hue |
Hive Metastore Server | metastore | hive |
Sentry Server | sentry | sentry |
Cloudera Navigator Audit Server | nav | nav |
Cloudera Navigator Metadata Server | navms | navms |
Oozie | oozie | oozie |
#进入mysql(输入修改后的密码)
mysql -u root -p
#scm库和权限暂时不创建,后面指定数据库,会自动创建
#CREATE DATABASE scm DEFAULT CHARACTER SET utf8 DEFAULT COLLATE utf8_general_ci;
CREATE DATABASE amon DEFAULT CHARACTER SET utf8 DEFAULT COLLATE utf8_general_ci;
CREATE DATABASE rman DEFAULT CHARACTER SET utf8 DEFAULT COLLATE utf8_general_ci;
CREATE DATABASE metastore DEFAULT CHARACTER SET utf8 DEFAULT COLLATE utf8_general_ci;
CREATE DATABASE hue DEFAULT CHARACTER SET utf8 DEFAULT COLLATE utf8_general_ci;
CREATE DATABASE oozie DEFAULT CHARACTER SET utf8 DEFAULT COLLATE utf8_general_ci;
#GRANT ALL ON scm.* TO 'scm'@'%' IDENTIFIED BY 'scm';
GRANT ALL ON amon.* TO 'amon'@'%' IDENTIFIED BY 'amon';
GRANT ALL ON rman.* TO 'rman'@'%' IDENTIFIED BY 'rman';
GRANT ALL ON metastore.* TO 'hive'@'%' IDENTIFIED BY 'hive';
GRANT ALL ON hue.* TO 'hue'@'%' IDENTIFIED BY 'hue';
GRANT ALL ON oozie.* TO 'oozie'@'%' IDENTIFIED BY 'oozie';
#####注意此处再授权一个本主机名地址,不然web页面配置很容易出错,注意修改本地主机名hostname
GRANT ALL ON amon.* TO 'amon'@'本主机名' IDENTIFIED BY 'amon';
#刷新源
FLUSH PRIVILEGES;
#检查权限是否正确
show grants for 'amon'@'%';
show grants for 'rman'@'%';
show grants for 'hive'@'%';
show grants for 'hue'@'%';
show grants for 'oozie'@'%';
#退出
quit
#重启服务
service mysql restart
二 CM+CDH安装
2.1 CM下载
红框中都需要下载
2.2 通过yum安装daemons,agent,server
主节点
yum list | grep cloudera-manager
yum -y install cloudera-manager-daemons cloudera-manager-agent cloudera-manager-server
从节点
yum list | grep cloudera-manager
yum -y install cloudera-manager-daemons cloudera-manager-agent
注意从节点不能查到,可能由防火墙引起,确认关闭
2.3 上传cloudera-manager-installer.bin到主节点,赋予权限
mkdir /usr/software
cd /usr/software
chmod +x cloudera-manager-installer.bin
2.4 CDH下载
下载地址:https://archive.cloudera.com/cdh6/6.3.2/parcels/
红框中都需要下载
2.5 配置本地Parcel存储库
- 将CDH下载的文件,上传到指定文件夹(主节点)
cd /opt/cloudera/parcel-repo/
- 校验文件是否下载完全
sha1sum CDH-6.3.2-1.cdh6.3.2.p0.1605554-el7.parcel | awk '{ print $1 }' > CDH-6.3.2-1.cdh6.3.2.p0.1605554-el7.parcel.sha
2.6 执行安装
cd /usr/software
./cloudera-manager-installer.bin
rm -f /etc/cloudera-scm-server/db.properties
#再次执行脚本(一路yes)
./cloudera-manager-installer.bin
此处便安装完成,可以登录主机IP+7180端口登录web页面,用户名密码admin/admin,但是此处暂时不要急于登录网页并配置参数,先改掉数据库连接方式重启cloudera-scm-server后再进行配置,操作如下:
- 1) 执行脚本scm_prepare_database.sh
#设置Cloudera Manager 数据库
/opt/cloudera/cm/schema/scm_prepare_database.sh mysql -uroot -p'root' scm scm scm
#进如mysql(这里是上面改的密码root)
mysql -uroot -proot
GRANT ALL ON scm.* TO 'scm'@'%' IDENTIFIED BY 'scm';
FLUSH PRIVILEGES;
show grants for 'scm'@'%';
quit
- 2) 停止ClouderaManager服务
service cloudera-scm-server stop
service cloudera-scm-server-db stop
- 3) 删除内嵌的默认数据库PostgreSQL的配置
rm -f /etc/cloudera-scm-server/db.mgmt.properties
- 4) 启动ClouderaManager服务
service cloudera-scm-server start
- 如果有问题查看日志
vim /var/log/cloudera-scm-server/cloudera-scm-server.log
2.7 CM->CDH安装(账号:admin,密码:admin)
处理相关警告:所有机器都要修改
- 问题一
#临时
sysctl vm.swappiness=10
#永久
echo 'vm.swappiness=10'>> /etc/sysctl.conf
- 问题二
echo never > /sys/kernel/mm/transparent_hugepage/defrag
echo never > /sys/kernel/mm/transparent_hugepage/enabled
#修改启动脚本
vim /etc/rc.local
--->添加
echo never > /sys/kernel/mm/transparent_hugepage/defrag
echo never > /sys/kernel/mm/transparent_hugepage/enabled
---<
重新运行检查
2.8 第一次组件安装
选择组件,为了简单安全有效,选择HDFS,YARN,ZOOKEEPER
这里可以将SecondaryNameNode和NameNode放在不同的机器,可以在NameNode异常时竟可能的确保数据正确性
下图为基本配置,不需要管,下一步
安装完成
2.9 问题处理
2.9.1 Clock Offset及 Network Interface Speed
2.9.2 HDFS Canary
直接抑制
2.9.3 Java Heap Size of NameNode in Bytes
堆大小,使用默认值4G即可
类似警告还有:
Java Heap Size of Host Monitor in Bytes
Maximum Non-Java Memory of Host Monitor
Java Heap Size of Service Monitor in Bytes
Maximum Non-Java Memory of Service Monitor
2.9.4 Memory Overcommit Validation Threshold
内存被调拨过度,所有服务安装完后,总的分配内存是 22.7G ,总内存是 23.4G
内存调拨为22.7/23.4=0.97>默认值0.8,改成1.0就OK了。
2.9.5 NameNode and SecondaryNameNode have different heapsizes
NameNode和SecondaryNameNode具有不同的堆大小
- 搜索 Java Heap Size
2.9.6 Erasure Coding Policy Verification Test
擦除编码策略验证测试:直接抑制
2.9.7 文件系统检查点
不良 : 文件系统检查点已有 20 小时,40 分钟。占配置检查点期限 1 小时的 2,068.25%。 临界阈值:400.00%。 自上个文件系统检查点以来已发生 255 个事务。这是 1,000,000 的已配置检查点事务目标的 0.03%。
解决方法:
1、namenode的Cluster ID 与 secondnamenode的Cluster ID 不一致,对比/dfs/nn/current/VERSION 和/dfs/snn/current/VERSION中的Cluster ID 来确认,如果不一致改成一致后重启节点应该可以解决。
2、修改之后还出现这个状况,查看secondnamenode 日志,报
ERROR: Exception in doCheckpoint java.io.IOException: Inconsistent checkpoint field
这个错误,直接删除 /dfs/snn/current/下所有文件,重启snn节点
2.10 正常状态
可以使用添加服务来添加需要的组件
三 安装Kafka
基础配置走起
四 安装Flume
五 安装Hbase
基础配置走起
六 安装Spark
默认配置走起
七 安装hive
基础配置走起
- 解决中文乱码
#进入mysql(输入修改后的密码)
mysql -u root -p
#进入hive元数据库
use metastore
#hive中文乱码解决
alter table COLUMNS_V2 modify column COMMENT varchar(256) character set utf8;
alter table TABLE_PARAMS modify column PARAM_VALUE varchar(4000) character set utf8;
alter table PARTITION_PARAMS modify column PARAM_VALUE varchar(4000) character set utf8;
alter table PARTITION_KEYS modify column PKEY_COMMENT varchar(4000) character set utf8;
alter table INDEX_PARAMS modify column PARAM_VALUE varchar(4000) character set utf8;
#刷新
flush privileges;
#退出
quit
八 安装Oozie
默认配置走起
九 安装Sqoop
十 安装Impala
十一 安装hue
- hue常用连接器配置
https://docs.gethue.com/administrator/configuration/connectors/
Hue通过本机或SqlAlchemy连接器连接到任何数据库或仓库。完成HUE-8758之后,可以通过UI配置连接,直到此之前需要将它们添加到Hue hue_safety_valve.ini文件中。
十二 cdh 6.3.2引用maven
- pom.xml
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd">
<repositories>
<repository>
<id>cloudera</id>
<url>https://repository.cloudera.com/artifactory/cloudera-repos/</url>
</repository>
</repositories>
</project>
Project | groupId | artifactId | version |
---|---|---|---|
Apache Avro | org.apache.avro | avro | 1.8.2-cdh6.3.2 |
org.apache.avro | avro-compiler | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-guava-dependencies | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-ipc | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-mapred | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-maven-plugin | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-protobuf | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-service-archetype | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-thrift | 1.8.2-cdh6.3.2 | |
org.apache.avro | avro-tools | 1.8.2-cdh6.3.2 | |
org.apache.avro | trevni-avro | 1.8.2-cdh6.3.2 | |
org.apache.avro | trevni-core | 1.8.2-cdh6.3.2 | |
Apache Crunch | org.apache.crunch | crunch-archetype | 0.11.0-cdh6.3.2 |
org.apache.crunch | crunch-contrib | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-core | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-examples | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-hbase | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-hive | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-scrunch | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-spark | 0.11.0-cdh6.3.2 | |
org.apache.crunch | crunch-test | 0.11.0-cdh6.3.2 | |
Apache Flume 1.x | org.apache.flume | flume-ng-auth | 1.9.0-cdh6.3.2 |
org.apache.flume | flume-ng-configuration | 1.9.0-cdh6.3.2 | |
org.apache.flume | flume-ng-core | 1.9.0-cdh6.3.2 | |
org.apache.flume | flume-ng-embedded-agent | 1.9.0-cdh6.3.2 | |
org.apache.flume | flume-ng-node | 1.9.0-cdh6.3.2 | |
org.apache.flume | flume-ng-sdk | 1.9.0-cdh6.3.2 | |
org.apache.flume | flume-ng-tests | 1.9.0-cdh6.3.2 | |
org.apache.flume | flume-tools | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-channels | flume-file-channel | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-channels | flume-jdbc-channel | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-channels | flume-kafka-channel | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-channels | flume-spillable-memory-channel | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-clients | flume-ng-log4jappender | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-configfilters | flume-ng-config-filter-api | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-configfilters | flume-ng-environment-variable-config-filter | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-configfilters | flume-ng-external-process-config-filter | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-configfilters | flume-ng-hadoop-credential-store-config-filter | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-legacy-sources | flume-avro-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-legacy-sources | flume-thrift-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-dataset-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-hdfs-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-hive-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-http-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-irc-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-ng-hbase2-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-ng-kafka-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sinks | flume-ng-morphline-solr-sink | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sources | flume-jms-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sources | flume-kafka-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sources | flume-scribe-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sources | flume-taildir-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-ng-sources | flume-twitter-source | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-shared | flume-shared-kafka | 1.9.0-cdh6.3.2 | |
org.apache.flume.flume-shared | flume-shared-kafka-test | 1.9.0-cdh6.3.2 | |
GCS Connector | com.google.cloud.bigdataoss | gcs-connector | hadoop3-1.9.10-cdh6.3.2 |
com.google.cloud.bigdataoss | gcsio | 1.9.10-cdh6.3.2 | |
com.google.cloud.bigdataoss | util | 1.9.10-cdh6.3.2 | |
com.google.cloud.bigdataoss | util-hadoop | hadoop3-1.9.10-cdh6.3.2 | |
Apache Hadoop | org.apache.hadoop | hadoop-aliyun | 3.0.0-cdh6.3.2 |
org.apache.hadoop | hadoop-annotations | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-archive-logs | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-archives | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-assemblies | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-auth | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-aws | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-azure | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-azure-datalake | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-build-tools | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-client | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-client-api | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-client-integration-tests | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-client-minicluster | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-client-runtime | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-cloud-storage | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-common | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-datajoin | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-distcp | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-extras | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-gridmix | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-hdfs | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-hdfs-client | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-hdfs-httpfs | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-hdfs-native-client | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-hdfs-nfs | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-kafka | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-kms | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-app | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-common | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-core | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-hs | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-hs-plugins | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-jobclient | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-nativetask | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-shuffle | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-client-uploader | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-mapreduce-examples | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-maven-plugins | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-minicluster | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-minikdc | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-nfs | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-openstack | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-resourceestimator | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-rumen | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-sls | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-streaming | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-api | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-applications-distributedshell | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-applications-unmanaged-am-launcher | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-client | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-common | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-registry | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-applicationhistoryservice | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-common | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-nodemanager | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-resourcemanager | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-router | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-sharedcachemanager | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-tests | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-timeline-pluginstorage | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-timelineservice | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-timelineservice-hbase | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-timelineservice-hbase-tests | 3.0.0-cdh6.3.2 | |
org.apache.hadoop | hadoop-yarn-server-web-proxy | 3.0.0-cdh6.3.2 | |
Apache HBase | org.apache.hbase | hbase-annotations | 2.1.0-cdh6.3.2 |
org.apache.hbase | hbase-checkstyle | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-client | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-client-project | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-common | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-endpoint | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-examples | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-external-blockcache | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-hadoop-compat | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-hadoop2-compat | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-http | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-it | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-mapreduce | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-metrics | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-metrics-api | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-procedure | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-protocol | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-protocol-shaded | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-replication | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-resource-bundle | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-rest | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-rsgroup | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-server | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-shaded-client | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-shaded-client-byo-hadoop | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-shaded-client-project | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-shaded-mapreduce | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-shell | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-spark | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-spark-it | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-testing-util | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-thrift | 2.1.0-cdh6.3.2 | |
org.apache.hbase | hbase-zookeeper | 2.1.0-cdh6.3.2 | |
HBase Indexer | com.ngdata | hbase-indexer-all | 1.5-cdh6.3.2 |
com.ngdata | hbase-indexer-cli | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-common | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-demo | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-dist | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-engine | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-model | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-morphlines | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-mr | 1.5-cdh6.3.2 | |
com.ngdata | hbase-indexer-server | 1.5-cdh6.3.2 | |
com.ngdata | hbase-sep-api | 1.5-cdh6.3.2 | |
com.ngdata | hbase-sep-demo | 1.5-cdh6.3.2 | |
com.ngdata | hbase-sep-impl | 1.5-cdh6.3.2 | |
com.ngdata | hbase-sep-tools | 1.5-cdh6.3.2 | |
Apache Hive | org.apache.hive | hive-accumulo-handler | |
org.apache.hive | hive-ant | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-beeline | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-classification | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-cli | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-common | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-contrib | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-exec | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-hbase-handler | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-hplsql | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-jdbc | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-kryo-registrator | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-llap-client | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-llap-common | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-llap-ext-client | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-llap-server | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-llap-tez | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-metastore | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-orc | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-serde | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-service | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-service-rpc | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-shims | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-spark-client | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-storage-api | 2.1.1-cdh6.3.2 | |
org.apache.hive | hive-testutils | 2.1.1-cdh6.3.2 | |
org.apache.hive.hcatalog | hive-hcatalog-core | 2.1.1-cdh6.3.2 | |
org.apache.hive.hcatalog | hive-hcatalog-pig-adapter | 2.1.1-cdh6.3.2 | |
org.apache.hive.hcatalog | hive-hcatalog-server-extensions | 2.1.1-cdh6.3.2 | |
org.apache.hive.hcatalog | hive-hcatalog-streaming | 2.1.1-cdh6.3.2 | |
org.apache.hive.hcatalog | hive-webhcat | 2.1.1-cdh6.3.2 | |
org.apache.hive.hcatalog | hive-webhcat-java-client | 2.1.1-cdh6.3.2 | |
org.apache.hive.shims | hive-shims-0.23 | 2.1.1-cdh6.3.2 | |
org.apache.hive.shims | hive-shims-common | 2.1.1-cdh6.3.2 | |
org.apache.hive.shims | hive-shims-scheduler | 2.1.1-cdh6.3.2 | |
Apache Kafka | org.apache.kafka | connect-api | 2.2.1-cdh6.3.2 |
org.apache.kafka | connect-basic-auth-extension | 2.2.1-cdh6.3.2 | |
org.apache.kafka | connect-file | 2.2.1-cdh6.3.2 | |
org.apache.kafka | connect-json | 2.2.1-cdh6.3.2 | |
org.apache.kafka | connect-runtime | 2.2.1-cdh6.3.2 | |
org.apache.kafka | connect-transforms | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-clients | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-examples | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-log4j-appender | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-streams | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-streams-examples | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-streams-scala_2.11 | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-streams-scala_2.12 | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-streams-test-utils | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka-tools | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka_2.11 | 2.2.1-cdh6.3.2 | |
org.apache.kafka | kafka_2.12 | 2.2.1-cdh6.3.2 | |
Kite SDK | org.kitesdk | kite-data-core | 1.0.0-cdh6.3.2 |
org.kitesdk | kite-data-crunch | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-data-hbase | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-data-hive | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-data-mapreduce | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-data-oozie | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-data-s3 | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-data-spark | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-hadoop-compatibility | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-maven-plugin | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-minicluster | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-avro | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-core | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-hadoop-core | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-hadoop-parquet-avro | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-hadoop-rcfile | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-hadoop-sequencefile | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-json | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-maxmind | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-metrics-scalable | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-metrics-servlets | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-protobuf | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-saxon | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-solr-cell | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-solr-core | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-tika-core | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-tika-decompress | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-twitter | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-morphlines-useragent | 1.0.0-cdh6.3.2 | |
org.kitesdk | kite-tools | 1.0.0-cdh6.3.2 | |
Apache Kudu | org.apache.kudu | kudu-backup-tools | 1.10.0-cdh6.3.2 |
org.apache.kudu | kudu-backup2_2.11 | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-client | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-client-tools | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-flume-sink | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-hive | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-mapreduce | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-spark2-tools_2.11 | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-spark2_2.11 | 1.10.0-cdh6.3.2 | |
org.apache.kudu | kudu-test-utils | 1.10.0-cdh6.3.2 | |
Apache Oozie | org.apache.oozie | oozie-client | 5.1.0-cdh6.3.2 |
org.apache.oozie | oozie-core | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-examples | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-fluent-job-api | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-fluent-job-client | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-server | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-distcp | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-git | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-hcatalog | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-hive | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-hive2 | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-oozie | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-pig | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-spark | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-sqoop | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-sharelib-streaming | 5.1.0-cdh6.3.2 | |
org.apache.oozie | oozie-tools | 5.1.0-cdh6.3.2 | |
org.apache.oozie.test | oozie-mini | 5.1.0-cdh6.3.2 | |
Apache Pig | org.apache.pig | pig | 0.17.0-cdh6.3.2 |
org.apache.pig | piggybank | 0.17.0-cdh6.3.2 | |
org.apache.pig | pigsmoke | 0.17.0-cdh6.3.2 | |
org.apache.pig | pigunit | 0.17.0-cdh6.3.2 | |
Cloudera Search | com.cloudera.search | search-crunch | 1.0.0-cdh6.3.2 |
com.cloudera.search | search-mr | 1.0.0-cdh6.3.2 | |
Apache Sentry | com.cloudera.cdh | solr-upgrade | 1.0.0-cdh6.3.2 |
org.apache.sentry | sentry-binding-hbase-indexer | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-binding-hive | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-binding-hive-common | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-binding-hive-conf | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-binding-hive-follower | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-binding-kafka | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-binding-solr | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-core-common | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-core-model-db | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-core-model-indexer | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-core-model-kafka | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-core-model-solr | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-dist | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-hdfs-common | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-hdfs-dist | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-hdfs-namenode-plugin | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-hdfs-service | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-policy-common | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-policy-engine | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-provider-cache | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-provider-common | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-provider-db | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-provider-file | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-service-api | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-service-client | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-service-providers | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-service-server | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-service-web | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-shaded | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-shaded-miscellaneous | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-spi | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-tests-hive | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-tests-kafka | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-tests-solr | 2.1.0-cdh6.3.2 | |
org.apache.sentry | sentry-tools | 2.1.0-cdh6.3.2 | |
org.apache.sentry | solr-sentry-handlers | 2.1.0-cdh6.3.2 | |
Apache Solr | org.apache.lucene | lucene-analyzers-common | 7.4.0-cdh6.3.2 |
org.apache.lucene | lucene-analyzers-icu | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-kuromoji | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-morfologik | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-nori | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-opennlp | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-phonetic | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-smartcn | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-stempel | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-analyzers-uima | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-backward-codecs | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-benchmark | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-classification | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-codecs | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-core | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-demo | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-expressions | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-facet | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-grouping | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-highlighter | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-join | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-memory | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-misc | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-queries | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-queryparser | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-replicator | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-sandbox | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-spatial | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-spatial-extras | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-spatial3d | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-suggest | 7.4.0-cdh6.3.2 | |
org.apache.lucene | lucene-test-framework | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-analysis-extras | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-analytics | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-cell | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-clustering | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-core | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-dataimporthandler | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-dataimporthandler-extras | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-jetty-customizations | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-langid | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-ltr | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-prometheus-exporter | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-security-util | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-sentry-audit-logging | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-solrj | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-test-framework | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-uima | 7.4.0-cdh6.3.2 | |
org.apache.solr | solr-velocity | 7.4.0-cdh6.3.2 | |
Apache Spark | org.apache.spark | spark-avro_2.11 | 2.4.0-cdh6.3.2 |
org.apache.spark | spark-catalyst_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-core_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-graphx_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-hadoop-cloud_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-hive_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-kubernetes_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-kvstore_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-launcher_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-mllib-local_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-mllib_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-network-common_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-network-shuffle_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-network-yarn_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-repl_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-sketch_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-sql-kafka-0-10_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-sql_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-streaming-flume-assembly_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-streaming-flume-sink_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-streaming-flume_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-streaming-kafka-0-10-assembly_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-streaming-kafka-0-10_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-streaming_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-tags_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-unsafe_2.11 | 2.4.0-cdh6.3.2 | |
org.apache.spark | spark-yarn_2.11 | 2.4.0-cdh6.3.2 | |
Apache Sqoop | org.apache.sqoop | sqoop | 1.4.7-cdh6.3.2 |
Apache ZooKeeper | org.apache.zookeeper | zookeeper | 3.4.5-cdh6.3.2 |
环境篇:CM+CDH6.3.2环境搭建(全网最全)的更多相关文章
- ①CM+CDH6.2.0安装(全网最全)
CM+CDH6.2.0环境准备 一 虚拟机及CentOs7配置 CentOS下载地址 master(16g+80g+2cpu+2核)+2台slave(8g+60g+2cpu+2核) 1.1 打开&qu ...
- 环境篇:Atlas2.0.0兼容CDH6.2.0部署
环境篇:Atlas2.0.0兼容CDH6.2.0部署 Atlas 是什么? Atlas是一组可扩展和可扩展的核心基础治理服务,使企业能够有效地满足Hadoop中的合规性要求,并允许与整个企业数据生态系 ...
- 环境篇:Kylin3.0.1集成CDH6.2.0
环境篇:Kylin3.0.1集成CDH6.2.0 Kylin是什么? Apache Kylin™是一个开源的.分布式的分析型数据仓库,提供Hadoop/Spark 之上的 SQL 查询接口及多维分析( ...
- 环境篇:Atlas2.1.0兼容CDH6.3.2部署
环境篇:Atlas2.1.0兼容CDH6.3.2部署 Atlas 是什么? Atlas是一组可扩展和可扩展的核心基础治理服务,使企业能够有效地满足Hadoop中的合规性要求,并允许与整个企业数据生态系 ...
- Android总结篇系列:Android开发环境搭建
工欲善其事必先利其器. 1.安装并配置Java环境进入Java oracle官网,当前网址如下:http://www.oracle.com/technetwork/java/javase/downlo ...
- 环境篇:Zeppelin
环境篇:Zeppelin Zeppelin 是什么 Apache Zeppelin 是一个让交互式数据分析变得可行的基于网页的开源框架.Zeppelin提供了数据分析.数据可视化等功能. Zeppel ...
- 环境篇:呕心沥血@CDH线上调优
环境篇:呕心沥血@线上调优 为什么出这篇文章? 近期有很多公司开始引入大数据,由于各方资源有限,并不能合理分配服务器资源,和服务器选型,小叶这里将工作中的总结出来,给新入行的小伙伴带个方向,不敢说一定 ...
- or1200下raw-os(仿真环境篇)
貌似最近都在公司混日子过了,怎么办?哎哎哎~罪过啊罪过,不过也是的,加工资居然没我份,顶领导个肺的,叫我怎么继续活啊~哎哎哎~ 算了,不谈这些鸟事情了,说多了都是泪啊,这篇blog开始我们进入raw- ...
- 搭建windows环境下(nginx+mysql+php)开发环境
搭建windows环境下(nginx+mysql+php)开发环境 1. 所需准备应用程序包 1.1 nginx 程序包nginx-1.0.4.zip或其他版本(下载地址: http ...
随机推荐
- python实现摇骰子猜大小函数升级没把加注及三大运行商短信验证过滤
摇骰子游戏升级 此次更改增加下注功能,启动资金1000元,每次赔率都是一倍,钱输光退出. 源码: #!/user/bin/env python #-*-coding:utf-8 -*- #Author ...
- EOS基础全家桶(十二)智能合约IDE-VSCode
简介 上一篇我们介绍了EOS的专用IDE工具EOS Studio,该工具的优势是简单,易上手,但是灵活性低,且对系统资源开销大,依赖多,容易出现功能异常.那么我们开发人员最容易使用的,可能还是深度定制 ...
- 案例 (一)如何把python项目部署到linux服务器上
一.背景 用Python写了个脚本,需要部署到Linux环境的服务器上,由于服务器linux系统(centos,redhat等)自带的是python2,现在的python萌新都是从python3开 ...
- ABAP基础4:模块化
子程序定义 以form开始,以endform结束,用perform语句调用,from语句可以在程序内部/外部,perform一定要写在前面 perform. from. 子程序模块 endform. ...
- mysql 查询获取排名的方法(绝对有效)
http://blog.csdn.net/k8080880/article/details/11253305 select case when pid=0 then case when @prevTy ...
- Hyperledger Fabric——balance transfer(三)创建和加入Channel
详细解析blance transfer示例的创建通道(Channel)和加入节点到通道的过程. 创建Channel 1.首先看app.js的路由函数 var createChannel = requi ...
- 求平均成绩(hdu2023)
注意:要心细,不要错在小细节上.如int c[6];double agve; c[j]=agve:这是错误的. #include<stdio.h> #include<cmath> ...
- 四使用浮动div布局
刚开始学习的小白,如有不足之处还请各位补充,感激涕零.在html中有两种方式布局<table>表格和<div>,个人剧的使用表格布局可以避免bug产生,并且表格布局相对来说要容 ...
- Js 改变时间格式输出格式
朋友看到的方法,非js原生的 自己封装到 function date2str(x,y) { var z={y:x.getFullYear(),M:x.getMonth()+1,d:x.getDate( ...
- Java 泛型与集合
1.List练习,请用泛型的写法来完成. 已知有一个Worker 类如下: public class Worker { private int age; private String name; p ...