win10下搭建storm环境
原文:https://blog.csdn.net/lu_wei_wei/article/details/80843365
1.下载storm;
http://mirror.bit.edu.cn/apache/storm/apache-storm-1.2.2/apache-storm-1.2.2.zip
2.下载zookeeper;
http://mirror.bit.edu.cn/apache/zookeeper/current/zookeeper-3.4.12.tar.gz
3.下载python;
4.启动zookeeper;
(1)解压zookeeper-3.4.12;
(2)进入zookeeper-3.4.12/conf;
(3)复制zoo_sample.cfg,重命名为zoo.cfg,不需要修改里面的配置;
(4)进入zookeeper-3.4.12/bin;
(5)启动zookeeper
命令:zkServer.cmd
上面代表已经启动成功!
5.启动storm相关;
(1)配置文件
进入apache-storm-1.2.2\conf storm.yaml 如下
# Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License. ########### These MUST be filled in for a storm configuration
# storm.zookeeper.servers:
# - "server1"
# - "server2"
storm.zookeeper.servers:
- "127.0.0.1"
#
# nimbus.seeds: ["host1", "host2", "host3"]
nimbus.seeds: ["127.0.0.1"]
storm.local.dir: "D:\\storm-local\\data3"
supervisor.slots.ports:
- 6700
- 6701
- 6702
- 6703
#
#
# ##### These may optionally be filled in:
#
## List of custom serializations
# topology.kryo.register:
# - org.mycompany.MyType
# - org.mycompany.MyType2: org.mycompany.MyType2Serializer
#
## List of custom kryo decorators
# topology.kryo.decorators:
# - org.mycompany.MyDecorator
#
## Locations of the drpc servers
# drpc.servers:
# - "server1"
# - "server2" ## Metrics Consumers
## max.retain.metric.tuples
## - task queue will be unbounded when max.retain.metric.tuples is equal or less than 0.
## whitelist / blacklist
## - when none of configuration for metric filter are specified, it'll be treated as 'pass all'.
## - you need to specify either whitelist or blacklist, or none of them. You can't specify both of them.
## - you can specify multiple whitelist / blacklist with regular expression
## expandMapType: expand metric with map type as value to multiple metrics
## - set to true when you would like to apply filter to expanded metrics
## - default value is false which is backward compatible value
## metricNameSeparator: separator between origin metric name and key of entry from map
## - only effective when expandMapType is set to true
# topology.metrics.consumer.register:
# - class: "org.apache.storm.metric.LoggingMetricsConsumer"
# max.retain.metric.tuples: 100
# parallelism.hint: 1
# - class: "org.mycompany.MyMetricsConsumer"
# max.retain.metric.tuples: 100
# whitelist:
# - "execute.*"
# - "^__complete-latency$"
# parallelism.hint: 1
# argument:
# - endpoint: "metrics-collector.mycompany.org"
# expandMapType: true
# metricNameSeparator: "." ## Cluster Metrics Consumers
# storm.cluster.metrics.consumer.register:
# - class: "org.apache.storm.metric.LoggingClusterMetricsConsumer"
# - class: "org.mycompany.MyMetricsConsumer"
# argument:
# - endpoint: "metrics-collector.mycompany.org"
#
# storm.cluster.metrics.consumer.publish.interval.secs: 60 # Event Logger
# topology.event.logger.register:
# - class: "org.apache.storm.metric.FileBasedEventLogger"
# - class: "org.mycompany.MyEventLogger"
# arguments:
# endpoint: "event-logger.mycompany.org" # Metrics v2 configuration (optional)
#storm.metrics.reporters:
# # Graphite Reporter
# - class: "org.apache.storm.metrics2.reporters.GraphiteStormReporter"
# daemons:
# - "supervisor"
# - "nimbus"
# - "worker"
# report.period: 60
# report.period.units: "SECONDS"
# graphite.host: "localhost"
# graphite.port: 2003
#
# # Console Reporter
# - class: "org.apache.storm.metrics2.reporters.ConsoleStormReporter"
# daemons:
# - "worker"
# report.period: 10
# report.period.units: "SECONDS"
# filter:
# class: "org.apache.storm.metrics2.filters.RegexFilter"
# expression: ".*my_component.*emitted.*"
启动storm:分别启动Nimbus、Supervisor、Storm UI Daemons
进如apache-storm-1.2.2\bin
启动 Nimbus:
storm.py nimbus
启动 Supervisor:
storm.py supervisor
启动 Storm UI
storm.py ui
启动完毕,输入http://127.0.0.1:8080/访问
1.下载storm; http://mirror.bit.edu.cn/apache/storm/apache-storm-1.2.2/apache-storm-1.2.2.zip 2.下载zookeeper; http://mirror.bit.edu.cn/apache/zookeeper/current/zookeeper-3.4.12.tar.gz 3.下载python; 4.启动zookeeper; (1)解压zookeeper-3.4.12;
(2)进入zookeeper-3.4.12/conf; (3)复制zoo_sample.cfg,重命名为zoo.cfg,不需要修改里面的配置; (4)进入zookeeper-3.4.12/bin; (5)启动zookeeper 命令:zkServer.cmd
上面代表已经启动成功! 5.启动storm相关; (1)配置文件 进入apache-storm-1.2.2\conf storm.yaml 如下
# Licensed to the Apache Software Foundation (ASF) under one# or more contributor license agreements. See the NOTICE file# distributed with this work for additional information# regarding copyright ownership. The ASF licenses this file# to you under the Apache License, Version 2.0 (the# "License"); you may not use this file except in compliance# with the License. You may obtain a copy of the License at## http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing, software# distributed under the License is distributed on an "AS IS" BASIS,# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.# See the License for the specific language governing permissions and# limitations under the License.
########### These MUST be filled in for a storm configuration# storm.zookeeper.servers:# - "server1"# - "server2"storm.zookeeper.servers: - "127.0.0.1"# # nimbus.seeds: ["host1", "host2", "host3"]nimbus.seeds: ["127.0.0.1"]storm.local.dir: "D:\\storm-local\\data3"supervisor.slots.ports: - 6700 - 6701 - 6702 - 6703# # # ##### These may optionally be filled in:# ## List of custom serializations# topology.kryo.register:# - org.mycompany.MyType# - org.mycompany.MyType2: org.mycompany.MyType2Serializer### List of custom kryo decorators# topology.kryo.decorators:# - org.mycompany.MyDecorator### Locations of the drpc servers# drpc.servers:# - "server1"# - "server2"
## Metrics Consumers## max.retain.metric.tuples## - task queue will be unbounded when max.retain.metric.tuples is equal or less than 0.## whitelist / blacklist## - when none of configuration for metric filter are specified, it'll be treated as 'pass all'.## - you need to specify either whitelist or blacklist, or none of them. You can't specify both of them.## - you can specify multiple whitelist / blacklist with regular expression## expandMapType: expand metric with map type as value to multiple metrics## - set to true when you would like to apply filter to expanded metrics## - default value is false which is backward compatible value## metricNameSeparator: separator between origin metric name and key of entry from map## - only effective when expandMapType is set to true# topology.metrics.consumer.register:# - class: "org.apache.storm.metric.LoggingMetricsConsumer"# max.retain.metric.tuples: 100# parallelism.hint: 1# - class: "org.mycompany.MyMetricsConsumer"# max.retain.metric.tuples: 100# whitelist:# - "execute.*"# - "^__complete-latency$"# parallelism.hint: 1# argument:# - endpoint: "metrics-collector.mycompany.org"# expandMapType: true# metricNameSeparator: "."
## Cluster Metrics Consumers# storm.cluster.metrics.consumer.register:# - class: "org.apache.storm.metric.LoggingClusterMetricsConsumer"# - class: "org.mycompany.MyMetricsConsumer"# argument:# - endpoint: "metrics-collector.mycompany.org"## storm.cluster.metrics.consumer.publish.interval.secs: 60
# Event Logger# topology.event.logger.register:# - class: "org.apache.storm.metric.FileBasedEventLogger"# - class: "org.mycompany.MyEventLogger"# arguments:# endpoint: "event-logger.mycompany.org"
# Metrics v2 configuration (optional)#storm.metrics.reporters:# # Graphite Reporter# - class: "org.apache.storm.metrics2.reporters.GraphiteStormReporter"# daemons:# - "supervisor"# - "nimbus"# - "worker"# report.period: 60# report.period.units: "SECONDS"# graphite.host: "localhost"# graphite.port: 2003## # Console Reporter# - class: "org.apache.storm.metrics2.reporters.ConsoleStormReporter"# daemons:# - "worker"# report.period: 10# report.period.units: "SECONDS"# filter:# class: "org.apache.storm.metrics2.filters.RegexFilter"# expression: ".*my_component.*emitted.*"123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115启动storm:分别启动Nimbus、Supervisor、Storm UI Daemons 进如apache-storm-1.2.2\bin 启动 Nimbus:
storm.py nimbus1
启动 Supervisor:
storm.py supervisor1
启动 Storm UI
storm.py ui1
启动完毕,输入http://127.0.0.1:8080/访问
--------------------- 作者:-奋斗的小鹿- 来源:CSDN 原文:https://blog.csdn.net/lu_wei_wei/article/details/80843365 版权声明:本文为博主原创文章,转载请附上博文链接!
win10下搭建storm环境的更多相关文章
- win10下Spark的环境搭建
win10下Spark的环境搭建 2018-08-19 18:36:45 一.jdk 1.8.0 安装与配置 二.scala 2.11.8 安装与配置http://www.scala-lang.or ...
- win10下搭建jz2440v3(arm s3c2440)开发及gdb调试环境【转】
本文转载自:https://blog.csdn.net/newjay03/article/details/72835758 本来打算完全在Ubuntu下开发的,但是水平有限,没有在Ubuntu下找到合 ...
- win10下搭建vue开发环境
特别说明:下面任何命令都是在windows的命令行工具下进行输入,打开命令行工具的快捷方式如下图: 详细的安装步骤如下: 一.安装node.js 说明:安装node.js的windows版本后 ...
- 1.WIN10下搭建vue开发环境
WIN10下搭建vue开发环境 详细的安装步骤如下: 一.安装node.js 说明:安装node.js的windows版本后,会自动安装好node以及包管理工具npm,我们后续的安装将依赖npm工具. ...
- win10下搭建深度学习--总结【学习笔记】
win10 下搭建深度学习开发环境总结: 1.本人环境如下:win10,GTX1050TI.i7,anaconda3,vs2015,cuda9.0,cudnn7.1.4,tensorflow-gpu= ...
- 在windows10下搭建ubuntu环境
虽然win10下搞了一个ubuntu子系统,但是还是各种不习惯,经过一番研究,我还是选择下面的组合来搭建: Git Bash + ConEmu + MinGW15.3 + vim + chocolat ...
- Linux 14.04lts 环境下搭建交叉编译环境arm-linux-gcc-4.5.1
交叉编译工具链是为了编译.链接.处理和调试跨平台体系结构的程序代码,在该环境下编译出嵌入式Linux系统所需要的操作系统.应用程序等,然后再上传到目标板上. 首 先要明确gcc 和arm-linux- ...
- Sublime Text 2下搭建Python环境常见错误
Sublime Text 2下搭建Python环境时,最容易出的错误就是Python环境配置错误,导致build(Ctrl+B)后没有任何反应. 关于Python编程环境的配置,网上很容易搜索到.先默 ...
- Android学习——windows下搭建Cygwin环境
在上一篇博文<Android学习——windows下搭建NDK_r9环境>中,我们详细的讲解了在windows下进行Android NDK开发环境的配置,我们也讲到了在NDk r7以后,我 ...
随机推荐
- c_数据结构_栈的实现
#include<stdio.h> #include<stdlib.h> #define STACK_INIT_SIZE 100 #define STACKINCREMENT ...
- Python_subprocess模块
subprocess中,允许生成新的进程,连接到input/output/error管道,并获取他们的返回(状态)码,主要用于替换os.system/os.spawn*几个旧的模块和方法 subpro ...
- hdu1202解题报告
#include<stdio.h>int main(){ int n,i; double sum,s,p,sum1; while(scanf("%d" ...
- webpack的build的时候时间长处理方案
观察第一次build的时间比较长,之后的编译时间较短,可以通过webpack -watch 监测性能 1, 将webpack升级到4.0,build 的速度提升很多 2,用webpack -watch ...
- JavaEE 之 RESTful
1.RESTful a.定义:一种软件架构风格,设计风格而不是标准,只是提供了一组设计原则和约束条件.它主要用于客户端和服务器交互类的软件.基于这个风格设计的软件可以更简洁,更有层次,更易于实现缓存等 ...
- 《Android高级进阶》读书笔记
<Android高级进阶>是据我所知的市面上唯一一本技术工具书,比较的高大全,作者的目的是为了对全领域有个初步的概念 No1: 在Android系统中,拥有事件传递处理能力的类有以下三种 ...
- Tree Reconstruction Gym - 101911G(构造)
---恢复内容开始--- Monocarp has drawn a tree (an undirected connected acyclic graph) and then has given ea ...
- netty简单NIO模型
首先是使用java原生nio类库编写的例子,开发一套nio框架不简单,所以选择了netty,该例完成后,是netty举例. package com.smkj.netty; public class T ...
- muduo学习笔记(六) 多线程的TcpServer
目录 前言 多线程TcpServer EventLoopThreadPool 线程池设计模式 muduo中的使用 连接的建立.消息.销毁 on_connection on_message on_clo ...
- EF Core使用CodeFirst在MySql中创建新数据库以及已有的Mysql数据库如何使用DB First生成域模型
官方教程:https://docs.microsoft.com/en-us/aspnet/core/data/?view=aspnetcore-2.1 使用EF CodeFirst在MySql中创建新 ...