1 DataStreamAPI
1.1 DataStream Data Sources

source是程序的数据源输入,你可以通过StreamExecutionEnvironment.addSource(sourceFunction)来为你的程序添加一个source。

flink提供了大量的已经实现好的source方法,可以自定义source
    通过实现sourceFunction接口来自定义无并行度的source。

1 使用并行度为1的source
      public class MyNoParalleSource implements SourceFunction<Long>{
      
          private long count = 1L;
      
          private boolean isRunning = true;
      
          /**
           * 主要的方法
           * 启动一个source
           * 大部分情况下,都需要在这个run方法中实现一个循环,这样就可以循环产生数据了
           *
           * @param ctx
           * @throws Exception
           */
          @Override
          public void run(SourceContext<Long> ctx) throws Exception {
              while(isRunning){
                  ctx.collect(count);
                  count++;
                  //每秒产生一条数据
                  Thread.sleep(1000);
              }
          }
          * 取消一个cancel的时候会调用的方法
          @Override
          public void cancel() {
              isRunning = false;
          }
      }
      
      2 Main方法执行
      public class StreamingDemoWithMyNoPralalleSource {
          public static void main(String[] args) throws Exception {
          //获取Flink的运行环境
          StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
          //获取数据源
          DataStreamSource<Long> text = env.addSource(new MyNoParalleSource()).setParallelism(1); //注意:针对此source,并行度只能设置为1
          DataStream<Long> num = text.map(new MapFunction<Long, Long>() {
              @Override
              public Long map(Long value) throws Exception {
                  System.out.println("接收到数据:" + value);
                  return value;
              }
          });

//每2秒钟处理一次数据
      DataStream<Long> sum = num.timeWindowAll(Time.seconds(2)).sum(0);
      //打印结果
      sum.print().setParallelism(1);
      String jobName = StreamingDemoWithMyNoPralalleSource.class.getSimpleName();
      env.execute(jobName);
       }
      }

可以通过实现ParallelSourceFunction接口或者继承RichParallelSourceFunction来自定义有并行度的source。继承RichParallelSourceFunction的那些SourceFunction意味着它们都是并行执行的并且可能有一些资源需要open/close

public class MyParalleSource implements ParallelSourceFunction<Long> {
      private long count = 1L;
      private boolean isRunning = true;

/**
       * 主要的方法
       * 启动一个source
       * 大部分情况下,都需要在这个run方法中实现一个循环,这样就可以循环产生数据了
       *
       * @param ctx
       * @throws Exception
       */
      @Override
      public void run(SourceContext<Long> ctx) throws Exception {
          while(isRunning){
              ctx.collect(count);
              count++;
              //每秒产生一条数据
              Thread.sleep(1000);
          }
      }
      /**
       * 取消一个cancel的时候会调用的方法
       *
       */
      @Override
      public void cancel() {
          isRunning = false;
       }
    }

public class StreamingDemoWithMyPralalleSource {
        public static void main(String[] args) throws Exception {
          //获取Flink的运行环境
          StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();

//获取数据源
          DataStreamSource<Long> text = env.addSource(new MyParalleSource()).setParallelism(2);

DataStream<Long> num = text.map(new MapFunction<Long, Long>() {
              @Override
              public Long map(Long value) throws Exception {
                  System.out.println("接收到数据:" + value);
                  return value;
              }
          });

//每2秒钟处理一次数据
          DataStream<Long> sum = num.timeWindowAll(Time.seconds(2)).sum(0);

//打印结果
          sum.print().setParallelism(1);

String jobName = StreamingDemoWithMyPralalleSource.class.getSimpleName();
          env.execute(jobName);
          }
       }

public class MyRichParalleSource extends RichParallelSourceFunction<Long> {
          private long count = 1L;
          private boolean isRunning = true;
          /**
           * 主要的方法
           * 启动一个source
           * 大部分情况下,都需要在这个run方法中实现一个循环,这样就可以循环产生数据了
           * @param ctx
           * @throws Exception
           */
          @Override
          public void run(SourceContext<Long> ctx) throws Exception {
              while(isRunning){
                  ctx.collect(count);
                  count++;
                  //每秒产生一条数据
                  Thread.sleep(1000);
              }
          }
          /**
           * 取消一个cancel的时候会调用的方法
           *
           */
          @Override
          public void cancel() {
              isRunning = false;
          }
          /**
           * 这个方法只会在最开始的时候被调用一次
           * 实现获取链接的代码
           * @param parameters
           * @throws Exception
           */
          @Override
          public void open(Configuration parameters) throws Exception {
              System.out.println("open.............");
              super.open(parameters);
          }
      
          /**
           * 实现关闭链接的代码
           * @throws Exception
           */
          @Override
          public void close() throws Exception {
              super.close();
          }
      }
      
       public class StreamingDemoWithMyRichPralalleSource {
       
           public static void main(String[] args) throws Exception {
              //获取Flink的运行环境
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
      
              //获取数据源
              DataStreamSource<Long> text = env.addSource(new MyRichParalleSource()).setParallelism(2);
      
              DataStream<Long> num = text.map(new MapFunction<Long, Long>() {
                  @Override
                  public Long map(Long value) throws Exception {
                      System.out.println("接收到数据:" + value);
                      return value;
                  }
              });
      
              //每2秒钟处理一次数据
              DataStream<Long> sum = num.timeWindowAll(Time.seconds(2)).sum(0);
      
              //打印结果
              sum.print().setParallelism(1);
      
              String jobName = StreamingDemoWithMyRichPralalleSource.class.getSimpleName();
              env.execute(jobName);
              }
          }

基于文件
    readTextFile(path)
    读取文本文件,文件遵循TextInputFormat 读取规则,逐行读取并返回。

基于socket
    socketTextStream从socker中读取数据,元素可以通过一个分隔符切开。

public class SocketDemoFullCount {
          public static void main(String[] args) throws Exception{
              //获取需要的端口号
              int port;
              try {
                  ParameterTool parameterTool = ParameterTool.fromArgs(args);
                  port = parameterTool.getInt("port");
              }catch (Exception e){
                  System.err.println("No port set. use default port 9010--java");
                  port = 9010;
              }
              //获取flink的运行环境
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
              String hostname = "SparkMaster";
              String delimiter = "\n";
              //连接socket获取输入的数据
              DataStreamSource<String> text = env.socketTextStream(hostname, port, delimiter);
      
              DataStream<Tuple2<Integer,Integer>> intData = text.map(new MapFunction<String, Tuple2<Integer,Integer>>() {
                  @Override
                  public Tuple2<Integer,Integer> map(String value) throws Exception {
                      return new Tuple2<>(1,Integer.parseInt(value));
                  }
              });
      
              intData.keyBy(0)
                      .timeWindow(Time.seconds(5))
                      .process(new ProcessWindowFunction<Tuple2<Integer,Integer>, String, Tuple, TimeWindow>() {
                          @Override
                          public void process(Tuple key, Context context, Iterable<Tuple2<Integer, Integer>> elements, Collector<String> out) throws Exception {
                              System.out.println("执行process");
                              long count = 0;
                               for(Tuple2<Integer,Integer> element: elements){
                                  count++;
                              }
                              out.collect("window:"+context.window()+",count:"+count);
                          }
                      }).print();
              //这一行代码一定要实现,否则程序不执行
              env.execute("Socket window count");
          }
      }

基于集合
    fromCollection(Collection)
    通过java 的collection集合创建一个数据流,集合中的所有元素必须是相同类型的。

public class StreamingFromCollection {
          public static void main(String[] args) throws Exception {
              //获取Flink的运行环境
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
              ArrayList<Integer> data = new ArrayList<>();
              data.add(10);
              data.add(15);
              data.add(20);
      
              //指定数据源
              DataStreamSource<Integer> collectionData = env.fromCollection(data);
              //通map对数据进行处理
              DataStream<Integer> num = collectionData.map(new MapFunction<Integer, Integer>() {
                  @Override
                  public Integer map(Integer value) throws Exception {
                      return value + 1;
                  }
              });
              //直接打印
              num.print().setParallelism(1);
              env.execute("StreamingFromCollection");
          }
    }

自定义输入
    addSource 可以实现读取第三方数据源的数据
    系统内置提供了一批connectors,连接器会提供对应的source支持【kafka】

1.2 DataStream Transformations

map:输入一个元素,然后返回一个元素,中间可以做一些清洗转换等操作

flatmap:输入一个元素,可以返回零个,一个或者多个元素

keyBy:根据指定的key进行分组,相同key的数据会进入同一个分区

dataStream.keyBy("someKey") // 指定对象中的 "someKey"字段作为分组key
      dataStream.keyBy(0) //指定Tuple中的第一个元素作为分组key
      
      注意:以下类型是无法作为key的
      1:一个实体类对象,没有重写hashCode方法,并且依赖object的hashCode方法
      2:一个任意形式的数组类型
      3:基本数据类型,int,long

filter:过滤函数,对传入的数据进行判断,符合条件的数据会被留下。

public class StreamingDemoFilter {
          public static void main(String[] args) throws Exception {
              //获取Flink的运行环境
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
              //获取数据源
              DataStreamSource<Long> text = env.addSource(new MyNoParalleSource()).setParallelism(1);//注意:针对此source,并行度只能设置为1
              DataStream<Long> num = text.map(new MapFunction<Long, Long>() {
                  @Override
                  public Long map(Long value) throws Exception {
                      System.out.println("原始接收到数据:" + value);
                      return value;
                  }
              });
      
              //执行filter过滤,满足条件的数据会被留下
              DataStream<Long> filterData = num.filter(new FilterFunction<Long>() {
                  //把所有的奇数过滤掉
                  @Override
                  public boolean filter(Long value) throws Exception {
                      return value % 2 == 0;
                  }
              });
      
              DataStream<Long> resultData = filterData.map(new MapFunction<Long, Long>() {
                  @Override
                  public Long map(Long value) throws Exception {
                      System.out.println("过滤之后的数据:" + value);
                      return value;
                  }
              });
      
      
              //每2秒钟处理一次数据
              DataStream<Long> sum = resultData.timeWindowAll(Time.seconds(2)).sum(0);
      
              //打印结果
              sum.print().setParallelism(1);
      
              String jobName = StreamingDemoFilter.class.getSimpleName();
              env.execute(jobName);
          }
      }

reduce:对数据进行聚合操作,结合当前元素和上一次reduce返回的值进行聚合操作,然后返回一个新的值

aggregations:sum(),min(),max()等

window:在后面单独详解

Union:合并多个流,新的流会包含所有流中的数据,但是union是一个限制,就是所有合并的流类型必须是一致的。

public class StreamingDemoUnion {
      public static void main(String[] args) throws Exception {
          //获取Flink的运行环境
          StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();

//获取数据源
          DataStreamSource<Long> text1 = env.addSource(new MyNoParalleSource()).setParallelism(1);//注意:针对此source,并行度只能设置为1

DataStreamSource<Long> text2 = env.addSource(new MyNoParalleSource()).setParallelism(1);

//把text1和text2组装到一起
          DataStream<Long> text = text1.union(text2);

DataStream<Long> num = text.map(new MapFunction<Long, Long>() {
              @Override
              public Long map(Long value) throws Exception {
                  System.out.println("原始接收到数据:" + value);
                  return value;
              }
          });
          //每2秒钟处理一次数据
          DataStream<Long> sum = num.timeWindowAll(Time.seconds(2)).sum(0);
          //打印结果
          sum.print().setParallelism(1);
          String jobName = StreamingDemoUnion.class.getSimpleName();
          env.execute(jobName);
      }
    }

Connect:和union类似,但是只能连接两个流,两个流的数据类型可以不同,会对两个流中的数据应用不同的处理方法。

CoMap, CoFlatMap:在ConnectedStreams中需要使用这种函数,类似于map和flatmap

public class StreamingDemoConnect {
      
          public static void main(String[] args) throws Exception {
              //获取Flink的运行环境
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
              //获取数据源
              DataStreamSource<Long> text1 = env.addSource(new MyNoParalleSource()).setParallelism(1);//注意:针对此source,并行度只能设置为1
              DataStreamSource<Long> text2 = env.addSource(new MyNoParalleSource()).setParallelism(1);
              SingleOutputStreamOperator<String> text2_str = text2.map(new MapFunction<Long, String>() {
                  @Override
                  public String map(Long value) throws Exception {
                      return "str_" + value;
                  }
              });
      
              ConnectedStreams<Long, String> connectStream = text1.connect(text2_str);
      
              SingleOutputStreamOperator<Object> result = connectStream.map(new CoMapFunction<Long, String, Object>() {
                  @Override
                  public Object map1(Long value) throws Exception {
                      return value;
                  }
                  @Override
                  public Object map2(String value) throws Exception {
                      return value;
                  }
              });
              //打印结果
              result.print().setParallelism(1);
              String jobName = StreamingDemoConnect.class.getSimpleName();
              env.execute(jobName);
          }
      }

Split:根据规则把一个数据流切分为多个流:

Select:和split配合使用,选择切分后的流

public class StreamingDemoSplit {
          public static void main(String[] args) throws Exception {
              //获取Flink的运行环境
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
      
              //获取数据源
              DataStreamSource<Long> text = env.addSource(new MyNoParalleSource()).setParallelism(1);//注意:针对此source,并行度只能设置为1
      
              //对流进行切分,按照数据的奇偶性进行区分
              SplitStream<Long> splitStream = text.split(new OutputSelector<Long>() {
                  @Override
                  public Iterable<String> select(Long value) {
                      ArrayList<String> outPut = new ArrayList<>();
                      if (value % 2 == 0) {
                          outPut.add("even");//偶数
                      } else {
                          outPut.add("odd");//奇数
                      }
                      return outPut;
                  }
              });
              
              //选择一个或者多个切分后的流
              DataStream<Long> evenStream = splitStream.select("even");
              DataStream<Long> oddStream = splitStream.select("odd");
      
              DataStream<Long> moreStream = splitStream.select("odd","even");
      
      
              //打印结果
              moreStream.print().setParallelism(1);
      
              String jobName = StreamingDemoSplit.class.getSimpleName();
              env.execute(jobName);
          }
      }

1.3 DataStream API之partition

Random partitioning:随机分区

dataStream.shuffle()

Rebalancing:对数据集进行再平衡,重分区,消除数据倾斜

dataStream.rebalance()

Rescaling:如果上游操作有2个并发,而下游操作有4个并发,那么上游的一个并发结果分配给下游的两个并发操作,另外的一个并发结果分配给了下游的另外两个并发操作。另一方面,下游有两个并发操作而上游又4个并发操作,那么上游的其中两个操作的结果分配给下游的一个并发操作而另外两个并发操作的结果则分配给另外一个并发操作。

Rescaling与Rebalancing的区别:
    Rebalancing会产生全量重分区,而Rescaling不会。

dataStream.rescale()

Custom partitioning:自定义分区需要实现Partitioner接口

dataStream.partitionCustom(partitioner, “someKey”) 针对对象

dataStream.partitionCustom(partitioner, 0) 针对Tuple

public class MyPartition implements Partitioner<Long> {
          @Override
          public int partition(Long key, int numPartitions) {
              System.out.println("分区总数:"+numPartitions);
              if(key % 2 == 0){
                  return 0;
              }else{
                  return 1;
              }
          }
      }

public class SteamingDemoWithMyParitition {
          public static void main(String[] args) throws Exception{
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
              env.setParallelism(2);
              DataStreamSource<Long> text = env.addSource(new MyNoParalleSource());
      
              //对数据进行转换,把long类型转成tuple1类型
              DataStream<Tuple1<Long>> tupleData = text.map(new MapFunction<Long, Tuple1<Long>>() {
                  @Override
                  public Tuple1<Long> map(Long value) throws Exception {
                      return new Tuple1<>(value);
                  }
              });
              //分区之后的数据
              DataStream<Tuple1<Long>> partitionData= tupleData.partitionCustom(new MyPartition(), 0);
              DataStream<Long> result = partitionData.map(new MapFunction<Tuple1<Long>, Long>() {
                  @Override
                  public Long map(Tuple1<Long> value) throws Exception {
                      System.out.println("当前线程id:" + Thread.currentThread().getId() + ",value: " + value);
                      return value.getField(0);
                  }
              });
      
              result.print().setParallelism(1);
      
              env.execute("SteamingDemoWithMyParitition");
      
          }
      }

Broadcasting:在后面单独详解

1.4 DataStream Sink

writeAsText():将元素以字符串形式逐行写入,这些字符串通过调用每个元素的toString()方法来获取

print() / printToErr():打印每个元素的toString()方法的值到标准输出或者标准错误输出流中

自定义输出addSink【kafka、redis】

<dependency>
          <groupId>org.apache.bahir</groupId>
          <artifactId>flink-connector-redis_2.11</artifactId>
          <version>1.0</version>
      </dependency>

public class StreamingDemoToRedis {
          public static void main(String[] args) throws Exception{
              StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
              DataStreamSource<String> text = env.socketTextStream("hadoop100", 9000, "\n");

//对数据进行组装,把string转化为tuple2<String,String>
              DataStream<Tuple2<String, String>> l_wordsData = text.map(new MapFunction<String, Tuple2<String, String>>() {
                  @Override
                  public Tuple2<String, String> map(String value) throws Exception {
                      return new Tuple2<>("l_words", value);
                  }
              });
      
              //创建redis的配置
              FlinkJedisPoolConfig conf = new FlinkJedisPoolConfig.Builder().setHost("SparkMaster").setPort(6379).build();
      
              //创建redissink
              RedisSink<Tuple2<String, String>> redisSink = new RedisSink<>(conf, new MyRedisMapper());
      
              l_wordsData.addSink(redisSink);
      
              env.execute("StreamingDemoToRedis");
          }
      
          public static class MyRedisMapper implements RedisMapper<Tuple2<String, String>>{
              //表示从接收的数据中获取需要操作的redis key
              @Override
              public String getKeyFromData(Tuple2<String, String> data) {
                  return data.f0;
              }
              //表示从接收的数据中获取需要操作的redis value
              @Override
              public String getValueFromData(Tuple2<String, String> data) {
                  return data.f1;
              }
      
              @Override
              public RedisCommandDescription getCommandDescription() {
                  return new RedisCommandDescription(RedisCommand.LPUSH);
              }
          }
      }

2 DataSet API
2.1 DataSet Sources

基于文件
    readTextFile(path)

public class BatchWordCountJava {
          public static void main(String[] args) throws Exception{
          
              String inputPath = "D:\\data\\file";
              String outPath = "D:\\data\\result";
      
              //获取运行环境
              ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
              //获取文件中的内容
              DataSource<String> text = env.readTextFile(inputPath);
      
              DataSet<Tuple2<String, Integer>> counts = text.flatMap(new Tokenizer()).groupBy(0).sum(1);
              counts.writeAsCsv(outPath,"\n"," ").setParallelism(1);
              env.execute("batch word count");
      
          }
          public static class Tokenizer implements FlatMapFunction<String,Tuple2<String,Integer>>{
              public void flatMap(String value, Collector<Tuple2<String, Integer>> out) throws Exception {
                  String[] tokens = value.toLowerCase().split("\\W+");
                  for (String token: tokens) {
                      if(token.length()>0){
                          out.collect(new Tuple2<String, Integer>(token,1));
                      }
                  }
              }
          }
      }

基于集合
    fromCollection(Collection)

2.2 DataSet Transformations

Map:输入一个元素,然后返回一个元素,中间可以做一些清洗转换等操作

FlatMap:输入一个元素,可以返回零个,一个或者多个元素

MapPartition:类似map,一次处理一个分区的数据【如果在进行map处理的时候需要获取第三方资源链接,建议使用MapPartition】

public class BatchDemoMapPartition {
      public static void main(String[] args) throws Exception{
          //获取运行环境
          ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();

ArrayList<String> data = new ArrayList<>();
          data.add("hello you");
          data.add("hello me");
          DataSource<String> text = env.fromCollection(data);
          /*text.map(new MapFunction<String, String>() {
              @Override
              public String map(String value) throws Exception {
                  //获取数据库连接--注意,此时是每过来一条数据就获取一次链接
                  //处理数据
                  //关闭连接
                  return value;
              }
          });*/

DataSet<String> mapPartitionData = text.mapPartition(new MapPartitionFunction<String, String>() {
              @Override
              public void mapPartition(Iterable<String> values, Collector<String> out) throws Exception {
                  //获取数据库连接--注意,此时是一个分区的数据获取一次连接【优点,每个分区获取一次链接】
                  //values中保存了一个分区的数据
                  //处理数据
                  Iterator<String> it = values.iterator();
                  while (it.hasNext()) {
                      String next = it.next();
                      String[] split = next.split("\\W+");
                      for (String word : split) {
                          out.collect(word);
                      }
                  }
                  //关闭链接
              }
          });
          mapPartitionData.print();
      }
    }

Filter:过滤函数,对传入的数据进行判断,符合条件的数据会被留下

Reduce:对数据进行聚合操作,结合当前元素和上一次reduce返回的值进行聚合操作,然后返回一个新的值

Aggregate:sum、max、min等

Distinct:返回一个数据集中去重之后的元素,data.distinct()

public class BatchDemoDistinct {
              public static void main(String[] args) throws Exception{
                  //获取运行环境
                  ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
                  ArrayList<String> data = new ArrayList<>();
                  data.add("hello you");
                  data.add("hello me");
          
                  DataSource<String> text = env.fromCollection(data);
                  FlatMapOperator<String, String> flatMapData = text.flatMap(new FlatMapFunction<String, String>() {
                      @Override
                      public void flatMap(String value, Collector<String> out) throws Exception {
                          String[] split = value.toLowerCase().split("\\W+");
                          for (String word : split) {
                              System.out.println("单词:"+word);
                              out.collect(word);
                          }
                      }
                  });
                  flatMapData.distinct()// 对数据进行整体去重
                          .print();
              }
       }
    Join:内连接

public class BatchDemoJoin {
      public static void main(String[] args) throws Exception{
          //获取运行环境
          ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
          //tuple2<用户id,用户姓名>
          ArrayList<Tuple2<Integer, String>> data1 = new ArrayList<>();
          data1.add(new Tuple2<>(1,"zs"));
          data1.add(new Tuple2<>(2,"ls"));
          data1.add(new Tuple2<>(3,"ww"));

//tuple2<用户id,用户所在城市>
          ArrayList<Tuple2<Integer, String>> data2 = new ArrayList<>();
          data2.add(new Tuple2<>(1,"beijing"));
          data2.add(new Tuple2<>(2,"shanghai"));
          data2.add(new Tuple2<>(3,"guangzhou"));

DataSource<Tuple2<Integer, String>> text1 = env.fromCollection(data1);
          DataSource<Tuple2<Integer, String>> text2 = env.fromCollection(data2);
          
          text1.join(text2).where(0)//指定第一个数据集中需要进行比较的元素角标
                          .equalTo(0)//指定第二个数据集中需要进行比较的元素角标
                          .with(new JoinFunction<Tuple2<Integer,String>, Tuple2<Integer,String>, Tuple3<Integer,String,String>>() {
                              @Override
                              public Tuple3<Integer, String, String> join(Tuple2<Integer, String> first, Tuple2<Integer, String> second)
                                      throws Exception {
                                  return new Tuple3<>(first.f0,first.f1,second.f1);
                              }
                          }).print();

//注意,这里用map和上面使用的with最终效果是一致的。
          /*text1.join(text2).where(0)//指定第一个数据集中需要进行比较的元素角标
                  .equalTo(0)//指定第二个数据集中需要进行比较的元素角标
                  .map(new MapFunction<Tuple2<Tuple2<Integer,String>,Tuple2<Integer,String>>, Tuple3<Integer,String,String>>() {
                      @Override
                      public Tuple3<Integer, String, String> map(Tuple2<Tuple2<Integer, String>, Tuple2<Integer, String>> value) throws Exception {
                          return new Tuple3<>(value.f0.f0,value.f0.f1,value.f1.f1);
                      }
                  }).print();*/
      }
    }

OuterJoin:外链接

public class BatchDemoOuterJoin {

public static void main(String[] args) throws Exception{
      
              //获取运行环境
              ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
      
              //tuple2<用户id,用户姓名>
              ArrayList<Tuple2<Integer, String>> data1 = new ArrayList<>();
              data1.add(new Tuple2<>(1,"zs"));
              data1.add(new Tuple2<>(2,"ls"));
              data1.add(new Tuple2<>(3,"ww"));
      
      
              //tuple2<用户id,用户所在城市>
              ArrayList<Tuple2<Integer, String>> data2 = new ArrayList<>();
              data2.add(new Tuple2<>(1,"beijing"));
              data2.add(new Tuple2<>(2,"shanghai"));
              data2.add(new Tuple2<>(4,"guangzhou"));
      
      
              DataSource<Tuple2<Integer, String>> text1 = env.fromCollection(data1);
              DataSource<Tuple2<Integer, String>> text2 = env.fromCollection(data2);
      
              /**
               * 左外连接
               *
               * 注意:second这个tuple中的元素可能为null
               *
               */
              text1.leftOuterJoin(text2)
                      .where(0)
                      .equalTo(0)
                      .with(new JoinFunction<Tuple2<Integer,String>, Tuple2<Integer,String>, Tuple3<Integer,String,String>>() {
                          @Override
                          public Tuple3<Integer, String, String> join(Tuple2<Integer, String> first, Tuple2<Integer, String> second) throws Exception {
                              if(second==null){
                                  return new Tuple3<>(first.f0,first.f1,"null");
                              }else{
                                  return new Tuple3<>(first.f0,first.f1,second.f1);
                              }
      
                          }
                      }).print();
      
              /**
               * 右外连接
               *
               * 注意:first这个tuple中的数据可能为null
               *
               */
              text1.rightOuterJoin(text2)
                      .where(0)
                      .equalTo(0)
                      .with(new JoinFunction<Tuple2<Integer,String>, Tuple2<Integer,String>, Tuple3<Integer,String,String>>() {
                          @Override
                          public Tuple3<Integer, String, String> join(Tuple2<Integer, String> first, Tuple2<Integer, String> second) throws Exception {
                              if(first==null){
                                  return new Tuple3<>(second.f0,"null",second.f1);
                              }
                              return new Tuple3<>(first.f0,first.f1,second.f1);
                          }
                      }).print();
              /**
               * 全外连接
               *
               * 注意:first和second这两个tuple都有可能为null
               *
               */
              text1.fullOuterJoin(text2)
                      .where(0)
                      .equalTo(0)
                      .with(new JoinFunction<Tuple2<Integer,String>, Tuple2<Integer,String>, Tuple3<Integer,String,String>>() {
                          @Override
                          public Tuple3<Integer, String, String> join(Tuple2<Integer, String> first, Tuple2<Integer, String> second) throws Exception {
                              if(first==null){
                                  return new Tuple3<>(second.f0,"null",second.f1);
                              }else if(second == null){
                                  return new Tuple3<>(first.f0,first.f1,"null");
                              }else{
                                  return new Tuple3<>(first.f0,first.f1,second.f1);
                              }
                          }
                      }).print();
          }
      }
    Cross:获取两个数据集的笛卡尔积

public class BatchDemoCross {
      public static void main(String[] args) throws Exception{
          //获取运行环境
          ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();

//tuple2<用户id,用户姓名>
          ArrayList<String> data1 = new ArrayList<>();
          data1.add("zs");
          data1.add("ww");
          //tuple2<用户id,用户所在城市>
          ArrayList<Integer> data2 = new ArrayList<>();
          data2.add(1);
          data2.add(2);
          DataSource<String> text1 = env.fromCollection(data1);
          DataSource<Integer> text2 = env.fromCollection(data2);
          CrossOperator.DefaultCross<String, Integer> cross = text1.cross(text2);
          cross.print();
      }

Union:返回两个数据集的总和,数据类型需要一致

First-n:获取集合中的前N个元素

Sort Partition:在本地对数据集的所有分区进行排序,通过sortPartition()的链接调用来完成对多个字段的排序

public class BatchDemoFirstN {
      public static void main(String[] args) throws Exception{
          //获取运行环境
          ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
          ArrayList<Tuple2<Integer, String>> data = new ArrayList<>();
          data.add(new Tuple2<>(2,"zs"));
          data.add(new Tuple2<>(4,"ls"));
          data.add(new Tuple2<>(3,"ww"));
          data.add(new Tuple2<>(1,"xw"));
          data.add(new Tuple2<>(1,"aw"));
          data.add(new Tuple2<>(1,"mw"));

DataSource<Tuple2<Integer, String>> text = env.fromCollection(data);
          //获取前3条数据,按照数据插入的顺序
          text.first(3).print();
          System.out.println("==============================");

//根据数据中的第一列进行分组,获取每组的前2个元素
          text.groupBy(0).first(2).print();
          System.out.println("==============================");

//根据数据中的第一列分组,再根据第二列进行组内排序[升序],获取每组的前2个元素
          text.groupBy(0).sortGroup(1, Order.ASCENDING).first(2).print();
          System.out.println("==============================");

//不分组,全局排序获取集合中的前3个元素,针对第一个元素升序,第二个元素倒序
          text.sortPartition(0,Order.ASCENDING).sortPartition(1,Order.DESCENDING).first(3).print();
      }
    }

2.3 DataSet partition

Rebalance:对数据集进行再平衡,重分区,消除数据倾斜

Hash-Partition:根据指定key的哈希值对数据集进行分区

partitionByHash()

Range-Partition:根据指定的key对数据集进行范围分区
    .partitionByRange()

Custom Partitioning:自定义分区规则

自定义分区需要实现Partitioner接口

partitionCustom(partitioner, “someKey”)

partitionCustom(partitioner, 0)

public class BatchDemoHashRangePartition {
          public static void main(String[] args) throws Exception{
              //获取运行环境
              ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();
      
              ArrayList<Tuple2<Integer, String>> data = new ArrayList<>();
              data.add(new Tuple2<>(1,"hello1"));
              data.add(new Tuple2<>(2,"hello2"));
              data.add(new Tuple2<>(2,"hello3"));
              data.add(new Tuple2<>(3,"hello4"));
              data.add(new Tuple2<>(3,"hello5"));
              data.add(new Tuple2<>(3,"hello6"));
              data.add(new Tuple2<>(4,"hello7"));
              data.add(new Tuple2<>(4,"hello8"));
              data.add(new Tuple2<>(4,"hello9"));
              data.add(new Tuple2<>(4,"hello10"));
              data.add(new Tuple2<>(5,"hello11"));
              data.add(new Tuple2<>(5,"hello12"));
              data.add(new Tuple2<>(5,"hello13"));
              data.add(new Tuple2<>(5,"hello14"));
              data.add(new Tuple2<>(5,"hello15"));
              data.add(new Tuple2<>(6,"hello16"));
              data.add(new Tuple2<>(6,"hello17"));
              data.add(new Tuple2<>(6,"hello18"));
              data.add(new Tuple2<>(6,"hello19"));
              data.add(new Tuple2<>(6,"hello20"));
              data.add(new Tuple2<>(6,"hello21"));
      
      
              DataSource<Tuple2<Integer, String>> text = env.fromCollection(data);
      
              /*text.partitionByHash(0).mapPartition(new MapPartitionFunction<Tuple2<Integer,String>, Tuple2<Integer,String>>() {
                  @Override
                  public void mapPartition(Iterable<Tuple2<Integer, String>> values, Collector<Tuple2<Integer, String>> out) throws Exception {
                      Iterator<Tuple2<Integer, String>> it = values.iterator();
                      while (it.hasNext()){
                          Tuple2<Integer, String> next = it.next();
                          System.out.println("当前线程id:"+Thread.currentThread().getId()+","+next);
                      }
      
                  }
              }).print();*/
      
              text.partitionByRange(0).mapPartition(new MapPartitionFunction<Tuple2<Integer,String>, Tuple2<Integer,String>>() {
                  @Override
                  public void mapPartition(Iterable<Tuple2<Integer, String>> values, Collector<Tuple2<Integer, String>> out) throws Exception {
                      Iterator<Tuple2<Integer, String>> it = values.iterator();
                      while (it.hasNext()){
                          Tuple2<Integer, String> next = it.next();
                          System.out.println("当前线程id:"+Thread.currentThread().getId()+","+next);
                      }
                  }
              }).print();
          }
      }

2.4 DataSet Sink

writeAsText():将元素以字符串形式逐行写入,这些字符串通过调用每个元素的toString()方法来获取
    writeAsCsv():将元组以逗号分隔写入文件中,行及字段之间的分隔是可配置的。每个字段的值来自对象的toString()方法
    print():打印每个元素的toString()方法的值到标准输出或者标准错误输出流中

3 Flink 序列化器

Flink自带了针对诸如int,long,String等标准类型的序列化器

针对Flink无法实现序列化的数据类型,我们可以交给Avro和Kryo

使用方法:ExecutionEnvironment env = ExecutionEnvironment.getExecutionEnvironment();

使用avro序列化:env.getConfig().enableForceAvro();
      使用kryo序列化:env.getConfig().enableForceKryo();
      使用自定义序列化:env.getConfig().addDefaultKryoSerializer(Class<?> type, Class<? extends Serializer<?>> serializerClass)
      https://ci.apache.org/projects/flink/flink-docs-release-1.6/dev/custom_serializers.html
        1
        2
        3
        4

4 Flink 数据类型

Java Tuple 和 Scala case class

Java POJOs:java实体类

Primitive Types

默认支持java和scala基本数据类型

General Class Types

默认支持大多数java和scala class

Hadoop Writables
    支持hadoop中实现了org.apache.hadoop.Writable的数据类型

Special Types

例如scala中的Either Option 和Try

https://ci.apache.org/projects/flink/flink-docs-release-1.6/dev/api_concepts.html#supported-data-types

Flink 的datastreamAPI 以及function函数的调用,性能调优的更多相关文章

  1. Spring AOP在函数接口调用性能分析及其日志处理方面的应用

    面向切面编程可以实现在不修改原来代码的情况下,增加我们所需的业务处理逻辑,比如:添加日志.本文AOP实例是基于Aspect Around注解实现的,我们需要在调用API函数的时候,统计函数调用的具体信 ...

  2. 【翻译】Flink Table Api & SQL — 性能调优 — 流式聚合

    本文翻译自官网:Streaming Aggregation  https://ci.apache.org/projects/flink/flink-docs-release-1.9/dev/table ...

  3. [性能调优]在PeopleSoft中使用函数索引

    那些没有在PeopleSoft系统遇到性能问题的人,特别是基于Oracle数据库的PeopleSoft,可能不知道基于函数的索引. 根据定义,基于函数的索引是使用如下方法定义的: 基于表达式,例如算术 ...

  4. Flink调优

      第1章 资源配置调优 Flink性能调优的第一步,就是为任务分配合适的资源,在一定范围内,增加资源的分配与性能的提升是成正比的,实现了最优的资源配置后,在此基础上再考虑进行后面论述的性能调优策略. ...

  5. VS开发】C中调用C++文件中定义的function函数

    [VS开发]C中调用C++文件中定义的function函数 标签(空格分隔): [VS开发] 声明:引用请注明出处http://blog.csdn.net/lg1259156776/ 精要一揽 C调用 ...

  6. js中使用function定义类、实例化,函数的调用方法

    function Test002(name, age){ name, age, this.printInfo = function(){ //定义的公有方法 console.log(name, age ...

  7. Entity Framework 6 Recipes 2nd Edition(11-4)译 -> 在”模型定义”函数里调用另一个”模型定义”函数

    11-4.在”模型定义”函数里调用另一个”模型定义”函数 问题 想要用一个”模型定义”函数去实现另一个”模型定义”函数 解决方案 假设我们已有一个公司合伙人关系连同它们的结构模型,如Figure 11 ...

  8. TypeScript Function(函数)

    在JavaScript中,函数是构成任何应用程序的基础块.通过函数,你得以实现建立抽象层.模仿类.信息隐藏和模块化.在TypeScript中,虽然已经存在类和模块化,但是函数依旧在如何去"处 ...

  9. JS中函数的调用和this的值

    调用每一个函数会暂停当前函数的执行,传递控制权和参数给新函数.除了声明时定义的形式参数,每个函数还接收两个附加的参数:this 和 arguments. 参数this在面向对象编程中非常重要,他的值取 ...

随机推荐

  1. 菜鸟入门【ASP.NET Core】3:准备CentOS和Nginx环境转(转)

    基本软件 VMware虚拟机 centos:http://isoredirect.centos.org/centos/7/isos/x86_64/CentOS-7-x86_64-Minimal-170 ...

  2. LeetCode-数组操作-Python<三>

    上一篇:LeetCode链表相加-Python<二> 以前没怎么做过算法题,来来去去都是那些循环,所以先从数组简单题开始做. 这两天最大心得: 总在边界里考虑不周到,导致错误 做晕的时候, ...

  3. 将第三方包安装到maven本地仓库

    今天在做jasper report生成pdf文档的时候,需要引入亚洲字体jar包.maven仓库是有这个jar包,但是在项目pom文件始终不能下载.无奈只有将jar包安装到maven本地仓库. 1 将 ...

  4. elasticsearch6.7 05. Document APIs(9)Bulk API

    8.Bulk API 可以把多个index或delete操作放在单个bulk API中执行.这样可以极大地提高索引速度. /_bulkAPI使用如下的JSON结构: action_and_meta_d ...

  5. HighCharts使用更多图表HighChartsMore

    添加highcharts-moreimport HighCharts from 'highcharts'import highchartsMore from 'highcharts/highchart ...

  6. 【代码笔记】Web-HTML-图像

    一,效果图. 二,代码. <!DOCTYPE html> <html> <head> <meta charset="utf-8"> ...

  7. 在vue中赋值的路径没有被编译

    当我们跑起来的时候,f12会看到相对路径,但是此时会报错,说找不到图片,这时候有其中一个办法,直接 require进去. 这时候就可以成功显示图片,但是路径会不一样,因为编译出来. 至于如何props ...

  8. Windows下使用Rtools编译R语言包

    使用devtools安装github中的R源代码时,经常会出各种错误,索性搜了一下怎么在Windows下直接打包,网上的资料也是参差不齐,以下是自己验证通过的. 一.下载Rtools 下载地址:htt ...

  9. 安卓开发之ScrollView

    当界面不足以将所有的内容显示出来的时候便导致下面的部分内容无法显示出来 所有加上ScrollView 来讲要显示的内容放入之中便可以实现上下滚动界面内容 但是当要显示多个控件的时候会出错  原因是Sc ...

  10. 表单元素的required,autocomplete,list用法

    required: 当在一个表单元素如:input中加上required属性时,点击提交表单按钮,针对input会弹出一个默认的警告信息,如下图: 代码如下: <form id="my ...