Work Queues

  • using the Java Client

  • In the first tutorial we wrote programs to send and receive messages from a named queue. In this one we'll create a Work Queue that will be used to distribute time-consuming tasks among multiple workers.

  • The main idea behind Work Queues (aka: Task Queues) is to avoid doing a resource-intensive task immediately and having to wait for it to complete. Instead we schedule the task to be done later.We encapsulate a task as a message and send it to a queue.A worker process running in the background will pop the tasks and eventually execute the job.When you run many workers the tasks will be shared between them.

  • This concept is especially useful in web applications where it's impossible to handle a complex task during a short HTTP request window.

Preparation

  • In the previous part of this tutorial we sent a message containing "Hello World!". Now we'll be sending strings that stand for complex tasks.We don't have a real-world task, like images to be resized or pdf files to be rendered, so let's fake it by just pretending we're busy - by using the Thread.sleep() function.We'll take the number of dots in the string as its complexity; every dot will account for one second of "work". For example, a fake task described by Hello... will take three seconds.

  • We will slightly modify the Send.java code from our previous example, to allow arbitrary messages to be sent from the command line.This program will schedule tasks to our work queue, so let's name it NewTask.java:

  •   public class NewTask {
    
      private static final String TASK_QUEUE_NAME = "task_queue";
    
      public static void main(String[] argv) throws Exception {
    ConnectionFactory factory = new ConnectionFactory();
    factory.setHost("localhost");
    Connection connection = factory.newConnection();
    Channel channel = connection.createChannel(); channel.queueDeclare(TASK_QUEUE_NAME, true, false, false, null); String message="message";
    //String message = getMessage(argv);
    for (int i=0;i<5;i++) {
    message+=".";
    channel.basicPublish("", TASK_QUEUE_NAME,
    MessageProperties.PERSISTENT_TEXT_PLAIN,
    message.getBytes("UTF-8"));
    System.out.println(" [x] Sent '" + message + "'");
    }
    channel.close();
    connection.close();
    }
  • Our old Recv.java program also requires some changes: it needs to fake a second of work for every dot in the message body. It will handle delivered messages and perform the task, so let's call it Worker.java:

    •   public class Worker {
      private static final String TASK_QUEUE_NAME = "task_queue"; public static void main(String[] argv) throws Exception {
      ConnectionFactory factory = new ConnectionFactory();
      factory.setHost("localhost");
      final Connection connection = factory.newConnection();
      final Channel channel = connection.createChannel(); channel.queueDeclare(TASK_QUEUE_NAME, true, false, false, null);
      System.out.println(" [*] Waiting for messages. To exit press CTRL+C"); channel.basicQos(1); final Consumer consumer = new DefaultConsumer(channel) {
      @Override
      public void handleDelivery(String consumerTag, Envelope envelope, AMQP.BasicProperties properties, byte[] body) throws IOException {
      String message = new String(body, "UTF-8"); System.out.println(" [x] Received '" + message + "'");
      try {
      doWork(message);
      } finally {
      System.out.println(" [x] Done");
      //channel.basicAck(envelope.getDeliveryTag(), false);
      }
      }
      };
      boolean autoAck = true; // acknowledgment is covered below
      channel.basicConsume(TASK_QUEUE_NAME, autoAck, consumer);
      } private static void doWork(String task) {
      for (char ch : task.toCharArray()) {
      if (ch == '.') {
      try {
      Thread.sleep(10000);
      } catch (InterruptedException _ignored) {
      Thread.currentThread().interrupt();
      }
      }
      }
      }

    }

  • result:

Round-robin dispatching循环分发

  • One of the advantages of using a Task Queue is the ability to easily parallelise work.If we are building up a backlog of work,we can just add more workers and that way,scale easily.
  • First,run two worker instances at the same time.
  • Sencond,publish new tasks.
  • 结果如上图。
  • By default,RabbitMQ will send each message to the next consumer,in sequence.On average every consumer will get the same number of messages.This way of distributing message is called round-robin.

Message acknowledgment

  • Doing a task can take a few seconds.You may wonder what happens if one of the consumers starts a long task and dies with it only partly done.With our current code,once RabbitMQ delivers a message to the costomer it immediately marks it for detetion.In this case ,if you kill a worker we will lose the message it was just processing.We'll also lose all the messsages that were dispatched to this particular worker but were not yet handled.
  • But we don't want to lose any tasks.If a worker dies,we'd like the task to be delivered to another workere.
  • In order to make sure a message is never lost.RabbitMQ supports message acknowledgments.An ack is sent back by the consumer to tell RabbitMQ that a particular message has been received,processed and that RabbitMQ and that RabbitMQ is free to delete it.
  • If a consumer dies(its channel is closed,connection is closed,or TCP connection is lost) without sending an ack.RabbitMQ will understand that a message wasn't processed fully and re-queue it.If there are other consumers online at the same time ,it will then quickly redeliver it to another consumer.That way you can be sure that no message is lost,even if workers occasionally die.
  • There aren't any message timeouts;RabbitMQ will redeliver the message when the consumer dies.It's fine even if processing a message takes a very,very long time .
  • Manual message acknowledgments are turned on by default .In previous examples we explicitly turned them off via the autoAck-true flag.It's time to set this flag to false and send a proper acknowledgment form the worker,once we're done with a task.
  •   public static void main(String[] argv) throws Exception {
    ConnectionFactory factory = new ConnectionFactory();
    factory.setHost("localhost");
    final Connection connection = factory.newConnection();
    final Channel channel = connection.createChannel(); channel.queueDeclare(TASK_QUEUE_NAME, true, false, false, null);
    System.out.println(" [*] Waiting for messages. To exit press CTRL+C"); channel.basicQos(1); final Consumer consumer = new DefaultConsumer(channel) {
    @Override
    public void handleDelivery(String consumerTag, Envelope envelope, AMQP.BasicProperties properties, byte[] body) throws IOException {
    String message = new String(body, "UTF-8"); System.out.println(" [x] Received '" + message + "'");
    try {
    doWork(message);
    } finally {
    System.out.println(" [x] Done");
    channel.basicAck(envelope.getDeliveryTag(), false);
    }
    }
    };
    //boolean autoAck = true; // acknowledgment is covered below
    channel.basicConsume(TASK_QUEUE_NAME, false, consumer);
    }
  • Using this code we can be sure that even if you kill a worker while it processing a message,nothing will be lost.Soon after the worker dies all unacknowledged message will be redelivered.
  • Acknowledgment must be sent on same channel the delivery it is for was received on.Attempts to acknowledgment using a different channel will result in a channel-level protocol exception..
    • Forgotten acknowledgments
    • It's a common mistake to miss the basicAck .It's an easy error,but the consequences are serious.Messages will be redelivered when your client quits,but RabbitMQ will eat more and more memory as it won't be able to release any unacked messages.

Message durability

  • We have learned how to make sure that even if the consumer dies,the task isn't lost.But our tasks will still be lost if RabbitMQ server stops.
  • When RabbitMQ quits or crashs it will forget the queues and messages unless you tell it not to.Two things are required to make sure that messages aren't lost:we need to mark both the queue and messages as durable.
    • First,we need to make sure that RabbitMQ will nerver lose our queue .In order to do so,we need to declare it as durable:
    •   boolean durable = true;
      channel.queueDeclare("hello", durable, false, false, null);
    • Although this command is correct by itself ,it won't work in our present setup.That's because we've already defined a queue called 'hello' which is not durable.RabbitMQ doesn't allow you to redefine an existing queue with different parameters and will return an error to any programs that tries to do that.But there is a quick workaround -Let's declare a queue with different name,for example task_queue.
    • This queueDeclare change needs to be applied to both the producer and consumer code .
    • As this point we're sure that the task_queue queue won't be lost even if RabbitMQ restarts.Now we need to mark our messages as persistent-by setting MessageProperties to the value PERSISTENT_TEXT_PLAIN.
    •   import com.rabbitmq.client.MessageProperties;
      
        channel.basicPublish("", "task_queue",
      MessageProperties.PERSISTENT_TEXT_PLAIN,
      message.getBytes());
  • Note on message persistence
    • Marking messages as persistent doesn't fully guarantee that a message won't be lost.Although it tells RabbitMQ to save the message to disk,there is still a short time window when RabbitMQ has accepted a message and hasn't saved it yet .Also,RabbitMQ doesn't do fsync(2) for every message it may be just saved to cache and not really written to the disk.The persistence guarantee aren't strong,but it's more than enough for our simple task queue.

Fair dispatch

  • You might have noticed that the dispatching still doesn't work exactly as we want.For example in a situation with two workers,when all odd messages are heavy and even message are light,one worker will be constantly busy and the other one will do hardly any work.Well,RabbitMQ doesn't konw anything that and will still dispatch messages evently.
  • This happens because RabbitMQ just dispatches a message when the message enters the queue.It doesn't look at the number of unacknowledged messages for a consumer.It just blindly dispatches every n-th message to the n-th consumer.
  • In order to defeat that we can use the basicQos method with the prefetchCount=1 setting.This tells RabbitMQ not to give more than one message to a worker at a time.Or in other words,don't dispatch a new message to a worker until it has processed an acknowledged the previous one.Instead ,it will dispatch it to the next worker that is not still busy.
  •   int prefetchCount = 1;
    channel.basicQos(prefetchCount);

总结

  • Task Queue
  • 循环分发 Round-robin dispatch
  • 消息确认
  • 消息持久化
    • 队列持久化channel.queueDeclare(...,true,...)声明队列时指定
    • 消息持久化channel.basicPublish(...,MessageProperties.PERSISTENT_TEXT_PLAIN,...)发送消息时通过MessageProperties指定
  • 平均分发 fair dispatch (可能导致一个worker busy 另一个free)
    • channel.basicQos(1)//等得到ack时在分发下一条消息

4、Work-Queue的更多相关文章

  1. Python进阶(3)_进程与线程中的lock(线程中互斥锁、递归锁、信号量、Event对象、队列queue)

    1.同步锁 (Lock) 当全局资源(counter)被抢占的情况,问题产生的原因就是没有控制多个线程对同一资源的访问,对数据造成破坏,使得线程运行的结果不可预期.这种现象称为“线程不安全”.在开发过 ...

  2. STL学习笔记6 -- 栈stack 、队列queue 和优先级priority_queue 三者比较

    栈stack  .队列queue  和优先级priority_queue 三者比较 默认下stack 和queue 基于deque 容器实现,priority_queue 则基于vector 容器实现 ...

  3. Python之路(第四十五篇)线程Event事件、 条件Condition、定时器Timer、线程queue

    一.事件Event Event(事件):事件处理的机制:全局定义了一个内置标志Flag,如果Flag值为 False,那么当程序执行 event.wait方法时就会阻塞,如果Flag值为True,那么 ...

  4. 并发编程(五)——GIL全局解释器锁、死锁现象与递归锁、信号量、Event事件、线程queue

    GIL.死锁现象与递归锁.信号量.Event事件.线程queue 一.GIL全局解释器锁 1.什么是全局解释器锁 GIL本质就是一把互斥锁,相当于执行权限,每个进程内都会存在一把GIL,同一进程内的多 ...

  5. Python并发编程06 /阻塞、异步调用/同步调用、异步回调函数、线程queue、事件event、协程

    Python并发编程06 /阻塞.异步调用/同步调用.异步回调函数.线程queue.事件event.协程 目录 Python并发编程06 /阻塞.异步调用/同步调用.异步回调函数.线程queue.事件 ...

  6. [操作系统知识储备,进程相关概念,开启进程的两种方式、 进程Queue介绍]

    [操作系统知识储备,进程相关概念,开启进程的两种方式.进程Queue介绍] 操作系统知识回顾 为什么要有操作系统. 程序员无法把所有的硬件操作细节都了解到,管理这些硬件并且加以优化使用是非常繁琐的工作 ...

  7. Message、Handler、Message Queue、Looper 之间的关系

    单线程模型中Message.Handler.Message Queue.Looper之间的关系 1.Message Message即为消息,可以理解为线程间交流的信息.处理数据后台线程需要更新UI,你 ...

  8. [数据结构]——链表(list)、队列(queue)和栈(stack)

    在前面几篇博文中曾经提到链表(list).队列(queue)和(stack),为了更加系统化,这里统一介绍着三种数据结构及相应实现. 1)链表 首先回想一下基本的数据类型,当需要存储多个相同类型的数据 ...

  9. Python成长笔记 - 基础篇 (十一)----RabbitMQ、Redis 、线程queue

    本节内容: 1.RabbitMQ 消息队列 2.Redis 3.Mysql   PY 中的线程queue(threading Queue):用于多个线程之间进行数据交换,不能在进程间进行通信 进程qu ...

  10. 单线程模型中Message、Handler、Message Queue、Looper之间的关系

    1. Android进程 在了解Android线程之前得先了解一下Android的进程.当一个程序第一次启动的时候,Android会启动一个LINUX进程和一个主线程.默认的情况下,所有该程序的组件都 ...

随机推荐

  1. Python自动化运维之高级函数

    本帖最后由 陈泽 于 2018-6-20 17:31 编辑 一.协程 1.1协程的概念 协程,又称微线程,纤程.英文名Coroutine.一句话说明什么是线程:协程是一种用户态的轻量级线程.(其实并没 ...

  2. 环形链表 II

    给定一个链表,返回链表开始入环的第一个节点. 如果链表无环,则返回 null. 为了表示给定链表中的环,我们使用整数 pos 来表示链表尾连接到链表中的位置(索引从 0 开始). 如果 pos 是 - ...

  3. day66_10_10,vue项目环境搭建

    一.下载. 首先去官网查看网址. 下载vue环境之前需要先下载node,使用应用商城npm下载,可以将其下载源改成cnpm: """ node ~~ python:nod ...

  4. Redis Pipelining

    Redis是一种基于客户端-服务端模型以及请求/响应协议的TCP服务.这意味着通常情况下一个请求会遵循以下步骤: 客户端向服务端发送一个查询请求,并监听Socket返回,通常是以阻塞模式,等待服务端响 ...

  5. CentOS设置主机名称

    1.查看主机名,使用命令hostname: 2.修改主机名,修改/etc/hostname文件

  6. redis命令之 ----key(键)

    DEL DEL key [key ...] 删除给定的一个或多个 key . 不存在的 key 会被忽略. DUMP DUMP key 序列化给定 key ,并返回被序列化的值,使用 RESTORE  ...

  7. 将Excel表格数据转换成Datatable

    /// <summary> /// 将Excel表格数据转换成Datatable /// </summary> /// <param name="fileUrl ...

  8. Lambda 表达式构建初级示例(不完整)

    直接贴代码了: using System; using System.Collections.Generic; using System.Linq; using System.Linq.Express ...

  9. Microsoft.Extensions.DependencyInjection 之二:使用诊断工具观察内存占用

    目录 准备工作 大量接口与实现类的生成 elasticsearch+kibana+apm asp.net core 应用 请求与快照 Kibana 上的请求记录 请求耗时的分析 请求内存的分析 第2次 ...

  10. Mysql权限整理及授权命令

    1.创建用户sql> use mysql;sql> create user 'Ruthless'@'%' identified by '123456';注意:Ruthless -> ...