A multi-core processor includes logical partitions that have respective processor cores, memory areas, and Ethernet controllers. At least one of the Ethernet controllers is disabled for external communication and is assigned as an inter-partition Ethernet controller for inter-partition communication. The inter-partition Ethernet controller is configured in loopback mode. A transmitting partition addresses a message through a send buffer in a private memory area to the inter-partition Ethernet controller assigned to a receiving partition. The receiving inter-partition Ethernet controller copies the received message to a receive buffer in the receiving partition's memory area. The receive Ethernet controller returns the received message to the sending partition and the sending partition resumes control of the memory space of the send buffer, or alternatively, the receive Ethernet controller frees the memory space of the send buffer to the private memory of the sending partition.

BACKGROUND OF THE INVENTION

The present invention is directed to multi-core processors and, more particularly, to inter-partition communication in a multi-core processor.

A multi-core processor is a single computing component with two or more independent processor cores, which can run separate instructions in parallel, increasing overall speed. The cores may be included in a single integrated circuit (IC) or in more than one IC but in a single package. The different processor cores may run codes in the same operating system (OS) and may be scheduled to run code in parallel (symmetrical multi-processing—'SMP'), sharing common memory, provided that each task in the system is not in execution on two or more cores at the same time. SMP systems can move tasks between cores to balance the workload efficiently. Alternatively, different cores may be restricted as concerns sharing specific memory and input/output (I/O) ports and may run different code in the same OS or may run different OSs (asymmetrical multi-processing—'AMP'). A core may be dedicated to a specific OS or may be capable of working in more than one OS or may even run without an OS.

Multi-core processors may be used in many applications such as general-purpose embedded computing systems, embedded and network communications including routers, switches, media gateways, base station and radio network controllers, digital signal processing (DSP), and graphics and video processing, for example. Multi-core processors typically contain many Ethernet controllers, which are level 2 (L2) devices in the Open Systems Interconnection (OSI) model. These Ethernet controllers are usually configurable to be connected to any of the Physical layer (PHY) devices present in the multi-core processors. It is common to have more Ethernet controllers than the number of PHY devices.

A multi-core processor may include two or more logical partitions, usually each hosting a separate instance of an OS. Logical partitioning divides hardware resources so that specific cores, memory areas and I/O ports are allocated to the different partitions. The interaction between the partitions and the applications running may be managed by a hypervisor. A hypervisor organizes a virtual operating platform and manages the execution of multiple "guest" OSs running in parallel on the processor. Several guest OSs may share the virtualized hardware resources.

Communication is typically necessary between partitions, referred to as inter-partition communication. Inter-partition communication may take the form of messages or calls and may involve exchange of data and/or exchange of control signaling. Inter-partition communication may be implemented through memory area shared between the sending and receiving partitions. However, memory sharing reduces isolation of the partitions and increases risks to security, especially if the inter-partition communication opens up direct private memory access between the partitions. Also there is a risk of starvation where a partition(s) over allocates from the shared memory or loses or never frees up previously allocated memory. Sharing of memory makes system recovery complex in case of failure of partition(s) in the system. Such risks can be managed if a hypervisor is provided and mediates every inter-partition communication, but making hypervisor calls imposes overhead and can make communication slow. Thus, it would be advantageous to have a method for inter-partition communication that does not rely on shared memory.

BRIEF DESCRIPTION OF THE DRAWINGS

The present invention is illustrated by way of example and is not limited by embodiments thereof shown in the accompanying figures, in which like references indicate similar elements. Elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale.

FIG. 1 is a schematic block diagram of a known multi-core processor showing three logical partitions and a hypervisor;

FIG. 2 is a schematic block diagram of the multi-core processor of FIG. 1showing memory cache management and on-chip and peripheral communication connections and management;

FIG. 3 is a schematic diagram of inter-partition communication in part of a multi-core processor of the kind shown in FIGS. 1 and 2 in accordance with one embodiment of the invention, given by way of example;

FIG. 4 is a schematic diagram of inter-partition communication in part of a multi-core processor of the kind shown in FIGS. 1 and 2 in accordance with another embodiment of the invention, given by way of example; and

FIG. 5 is a flow chart of a method of operating a multi-core processor such as that shown in FIG. 3 or 4 in accordance with one embodiment of the invention, given by way of example.

DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS

FIG. 1 illustrates a known multi-core processor 100 with logical partitions 102,104 and 106 and a hypervisor 108 managing execution of instruction code by system hardware 110. The multi-core processor 100 is shown with three logical partitions but it will be appreciated that a different number of partitions may be provided. The logical partitions 102104 and 106 have respective processor cores such as 112114116 and 118, private memory areas such as 120122and 124, and input/output ('I/O') ports such as 126128130 and 132. The system hardware 110 also includes memory area 134 shared between the logical partitions 102104 and 106, a common I/O port 136, shared memory cache 138, I/O memory management unit (MMU) 140 and an interrupt controller 142.

The multi-core processor 100 may run application codes in the same operating system (OS) and may be scheduled to run a code in parallel, symmetrical multi-processing mode (SMP). In this example, the multi-core processor 100 is shown with the different logical partitions 102104 and 106 running application codes in different guest OS's in asymmetrical multi-processing mode (AMP), with the logical partition 102 executing Linux® OS, the logical partition 104 executing a third party real time OS (RTOS) and the logical partition 106 executing a lightweight executive (LWE) OS. The hypervisor 108 presents to the guest operating systems a virtual operating platform and manages the execution of the guest operating systems. The LWE OS provides run-to-completion data plane processing, so that processes do not pre-empt each other and each process must run to completion before other processes get a chance to run.

FIG. 2 illustrates in more detail hardware 200 and associated management layers in the multi-core processor 100. The hardware 200 includes a set 202 of cores such as 112 to 118 with associated private memory caches. A Corenet™ coherency fabric 204 manages coherency of the memory caches and provides on-chip and peripheral communication connections and management that supports concurrent traffic and eliminates single-point bottlenecks for non-competing resources. The Corenet™ coherency fabric 204 avoids contention and latency issues associated with scaling shared bus/shared memory architectures. The management layers for the hardware 200 include accelerator, encryption and power management modules 206, a buffer manager 208, a queue manager 209, common memory caches 210, such as the shared memory 134, memory controllers 212 and local bus controllers and interrupt control modules 214.

The hardware 200 also includes two I/O modules which include respective frame managers 216 and 218, respective 10 Gb/s Ethernet controllers 220 and 222, and respective sets 224 and 226 of four 1 Gb/s Ethernet controllers. In addition the hardware 200 includes an on-chip network 228 with a peripheral component interconnect (PCI) interface 230, a message I/O unit 232, a serial I/O unit 234 and a direct memory access (DMA) unit 236. A debug I/O unit 238 is provided for development and test work. All the I/O modules connect with a Serializer/Deserializer (SerDes) 240 for external communication, having blocks which convert data between serial data lanes and parallel interfaces in each direction. The SerDes 240 has 18 serial data lanes in this example. There may be one or more of the Ethernet controllers such as 220 to226 which are surplus and are kept disabled because there are not enough SERDES lanes to connect to them.

In operation, inter-partition communication in one known multi-core processor of the kind shown in FIG. 1 is performed using a specific virtual I/O adapter interface and I/O operation program. The I/O operation program is under the control of supervisor/hypervisor software that initializes the message routing tables in the adapter but the supervisor/hypervisor calls impose overhead and can make communication slow. The known inter-partition communication uses a data movement protocol which is dictated by the virtual I/O adapter.

FIGS. 3, 4 and 5 illustrate part of multi-core processors 300 and 400 and a method 500 of operating a multi-core processor in accordance with embodiments of the invention, given by way of example. Each of the multi-core processors 300 and 400includes a plurality of logical partitions 302 and 304 that have respective processor cores 306 and 308, memory areas 310and 312, and Ethernet controllers 314 and 316. The method 500 includes disabling for external communication at least one Ethernet controller 314 of the logical partitions 302 and 304 and assigning the Ethernet controller disabled for external communication as an inter-partition Ethernet controller for reception of inter-partition communication.

In an example of the method 500, the inter-partition Ethernet controller 314 is configured in loopback mode in which it works as a DMA device in the receive mailbox of the receiving partition 302. In an example of the method 500, one of the logical partitions is configured as a transmitting partition 302 and another of the logical partitions as a receiving partition 304 having respective private memory areas. The transmitting partition 302 addresses a message MBUF through a send buffer 318 in its private memory area 312 to the inter-partition Ethernet controller 314 assigned to the receiving partition 302, under the management of the queue manager 209 for example. The receiving partition 302 copies the message MSG received in its inter-partition Ethernet controller 314 into a receive buffer 320 in the private memory area 310 of the receiving partition.

In this example of the method 500, and as illustrated in FIG. 3, the inter-partition Ethernet controller 314 of the receiving partition 302 is configured in loopback mode and after copying the received message into said receive buffer returns the received message MSG to the application 324 in the sending partition 304, under the management of the queue manager209, notifying the sending partition of reception of the message, and the sending partition then resumes control of the memory space of the send buffer and can re-use the send buffer 318 for other messages.

In another example of the method 500, and as illustrated in FIG. 4, the inter-partition Ethernet controller 314 of the receiving partition 302 is configured in loopback mode and frees the memory space of the send buffer 318 to the private memory 312of the sending partition after copying the received message MSG into the receive buffer 320, by instructing the buffer manager 208.

In this example of the method 500, the receiving partition 302 copying the message MSG received in its inter-partition Ethernet controller 314 into a receive buffer 320 includes the receiving partition 302 allocating the receive buffer 302 from its memory area 310 and its inter-partition Ethernet controller 314 copying the message MSG received into the allocated receive buffer 302, the inter-partition Ethernet controller 314 and the receive buffer 302 forming a receive mailbox for the inter-partition message.

In more detail, in this example of the multi-core processors 300 and 400, application codes 322 and 324 are running in the partitions 302 and 304 under the LWE OS and the multi-core processor 300 is enabled for reconfigurable data path acceleration (DPA) operation. Ethernet controllers that are disabled for external communication, for example if the current device configuration leaves insufficient SerDes lanes for those Ethernet controllers, or because of device errata, are identified. A respective one of the Ethernet controllers identified as suitable for the inter-partition communication is then assigned to each of the logical partitions 302 and 304 in loopback mode. The buffer manager 208 can then allocate receive buffers such as 318 and 320 for the logical partitions 302 and 304 from their private memory 310 and 312, forming mailboxes for messages.

When the sending partition 304 defines a message to be sent to the receiving partition 302, the sending partition 304instructs the buffer manager 208 to allocate space in its private memory 312 to the send buffer 318, as shown by the arrows 326. The message to be sent MBUF is registered in the send buffer 318 and queued through the queue manager209 to the inter-partition mailbox receive Ethernet controller 314, as shown by the arrows 328. The receive Ethernet controller 314 in the receiving partition 302 then instructs the buffer manager 208 to allocate space in its private memory310 for the receive buffer 320, as shown by the arrows 330 and copies the received message MSG into the receive buffer320 as message MBUF, as shown by the arrow 332. In the processor 300, the receive Ethernet controller 314 then returns the received message MSG to the application 324 in the sending partition 304 under the management of the queue manager 209, notifying the sending partition of reception of the message, and the sending partition 304 then resumes control of the memory space of the send buffer 318. Alternatively, in the processor 400, the receive Ethernet controller 314then instructs the buffer manager 208 to free the send buffer 318 to the private memory 312 of the sending partition 304, as shown by the arrow 402. In each case, the receive Ethernet controller 314 then notifies the application 322 of the received message MBUF, as shown by the arrow 336.

The method 500 of operating a multi-core processor is summarized in the simplified flow chart of FIG. 5. The method starts at 502. At 504, the partitions and OSs of the multi-core processor are configured. At 506, Ethernet controllers 314 and 316which are disabled for external communication are identified and receive Ethernet controllers 314 are configured in loopback mode at 508. At 510, the sending partition 304 instructs the buffer manager 208 to allocate space in its private memory 312for the send buffer 318. At 512, the application 324 of the sending partition queues the message MBUF through the queue manager 209 to the receive Ethernet controller 314. The receive Ethernet controller 314 instructs the buffer manager 208 to allocate space in its private memory 310 for the receive buffer 320 at 514. At 516, the receive Ethernet controller 314copies the received message MSG into the receive buffer 320. At 518, either the receive Ethernet controller 314 returns the received message MSG through the queue manager 209 to the sending partition 304 as loopback and the sending partition304 then resumes control of the memory space of the send buffer 318 or the receive Ethernet controller 314 instructs the buffer  manager 208 to free the send buffer 318 to the private memory 312 of the sending partition 304. At 520, the receive Ethernet controller 314 notifies the application 322 of the received message MBUF and the method ends at 522.

It will be appreciated that no cycles of core processing time are necessary to copy the message for the receiving partition302. Also the inter-partition communication does not involve cycles of hypervisor time, and the allocation of buffers as mailboxes with the receive Ethernet controller is performed by the receiving partition 302 instructing the buffer manager 208. The Ethernet controllers 314 and 316 used for inter-partition communication are available in the processor 300 and are not specific hardware for inter-partition communication. The operation of copying the received message MSG into the receivebuffer 320 can be performed by any suitable data movement protocol. More than one buffer can be allocated to receive mailboxes by a receiving partition 302, if desired. Access control to sender partition memory can be enforced using Input Output Memory Management Unit ('IOMMU') enabling the receiver partition to allow its inter-partition-mailbox Ethernet port to copy messages selectively from a sending partition.

The invention may also be implemented using at least portions of processor code for performing steps of a method according to the invention when run on a programmable apparatus, such as a computer system or enabling a programmable apparatus to perform functions of a device or system according to the invention. A computer program is a list of instructions such as a particular application program and/or an operating system in processor code. The computer program may for instance include one or more of: a subroutine, a function, a procedure, an object method, an object implementation, an executable application, an applet, a servlet, a source code, an object code, a shared library/dynamic load library and/or other sequence of instructions designed for execution on a computer system.

The computer program may be stored internally on computer readable storage medium or transmitted to the computer system via a computer readable transmission medium. All or some of the computer program may be provided on computer readable media permanently, removably or remotely coupled to an information processing system. The computer readable media may include, for example and without limitation, any number of the following: magnetic storage media including disk and tape storage media; optical storage media such as compact disk media (e.g., CD-ROM, CD-R, etc.) and digital video disk storage media; nonvolatile memory storage media including semiconductor-based memory units such as FLASH memory, EEPROM, EPROM, ROM; ferromagnetic digital memories; MRAM; volatile storage media including registers, buffers or caches, main memory, RAM, etc.; and data transmission media including computer networks, point-to-point telecommunication equipment, and carrier wave transmission media, just to name a few.

In the foregoing specification, the invention has been described with reference to specific examples of embodiments of the invention. It will, however, be evident that various modifications and changes may be made therein without departing from the broader spirit and scope of the invention as set forth in the appended claims.

The connections as discussed herein may be any type of connection suitable to transfer signals from or to the respective nodes, units or devices, for example via intermediate devices. Accordingly, unless implied or stated otherwise, the connections may be direct connections or indirect connections. The connections may be illustrated or described in reference to being a single connection, a plurality of connections, unidirectional connections, or bidirectional connections. However, different embodiments may vary the implementation of the connections. For example, separate unidirectional connections may be used rather than bidirectional connections and vice versa. Also, a plurality of connections may be replaced with a single connection that transfers multiple signals serially or in a time multiplexed manner. Likewise, single connections carrying multiple signals may be separated out into various different connections carrying subsets of these signals. Therefore, many options exist for transferring signals.

SRC=https://www.google.com.hk/patents/US20130227243

Inter-partition communication in multi-core processor的更多相关文章

  1. Intel processor brand names-Xeon,Core,Pentium,Celeron----Xeon

    http://en.wikipedia.org/wiki/Comparison_of_Intel_processors Processor Series Nomenclature Code Name ...

  2. Intel processor brand names-Xeon,Core,Pentium,Celeron----Celeron

    http://en.wikipedia.org/wiki/Celeron Celeron From Wikipedia, the free encyclopedia     Celeron Produ ...

  3. solr多core的处理

    有2中配置方式,一是从Solr Admin进行multi core的配置. 在Solr Admin控制台里面选择:Core Admin 选择Add Core 然后把你准备好的路径写到里面去. name ...

  4. About Intel® Processor Numbers

    http://www.intel.com/content/www/us/en/processors/processor-numbers.html About Intel® Processor Numb ...

  5. linux kernel menuconfig【转载】

    原文网址:http://www.cnblogs.com/kulin/archive/2013/01/04/linux-core.html Linux内核裁减 (1)安装新内核: i)将新内核copy到 ...

  6. Linux内核配置选项

    http://blog.csdn.net/wdsfup/article/details/52302142 http://www.manew.com/blog-166674-12962.html Gen ...

  7. Dynamic device virtualization

    A system and method for providing dynamic device virtualization is herein disclosed. According to on ...

  8. 深入linux kernel内核配置选项

    ============================================================================== 深入linux kernel内核配置选项 ...

  9. Linux内核同步机制--转发自蜗窝科技

    Linux内核同步机制之(一):原子操作 http://www.wowotech.net/linux_kenrel/atomic.html 一.源由 我们的程序逻辑经常遇到这样的操作序列: 1.读一个 ...

随机推荐

  1. 基于element的表单渲染器 (el-form-renderer)

    基于 element-ui 封装的表单渲染器,完整继承了 element 的属性定义,并进行了简单扩展,从而用户能够通过使用一段预设的数据渲染出一个完整的 element 表单. 演示地址 项目地址 ...

  2. ubuntu12.04开启Framebuffer

    一.framebuffer概述 Framebuffer在Linux中是作为设备来实现的,它是对图形硬件的一种抽象,代表着显卡中的帧缓冲区(Framebuffer).通过Framebuffer设备,上层 ...

  3. js特效——自动滚动

    <!DOCTYPE html> <html> <head> <meta charset="UTF-8"> <title> ...

  4. phpexcel乱码问题

    php导出Excel乱码,只需在header函数前加入ob_end_clean();//清除缓冲区,避免乱码

  5. Mysql 5.7 官方文档翻译

    始于 2017年4月1日-愚人节 1.1 MySQL 5.7 新功能 本章节介绍了MySQL 5.7 新版本中新增.废弃.删除的功能. 在1.5章节 Section 1.5, "Server ...

  6. STM32系列ARM单片机介绍

    STM32系列基于专为要求高性能.低成本.低功耗的嵌入式应用专门设计的ARM Cortex-M3内核.按性能分成两个不同的系列:STM32F103"增强型"系列和STM32F101 ...

  7. 剑指offer_面试题_从上往下打印二叉树

    题目:从上往下打印出二叉树的每一个结点.同一层的结点依照从左到右的顺序打印.比如输入图4.5中的二叉树.则依次打印出8.6.10.5.7.9.11. 8 /     \ 6     10 /   \ ...

  8. php设计模式适配器模式

    php设计模式适配器模式 简介 适配器模式(有时候也称包装样式或者包装)将一个类的接口适配成用户所期待的.一个适配允许通常因为接口不兼容而不能在一起工作的类工作在一起. 其实就是通过一个转换类,这个转 ...

  9. nyoj--233--Sort it (水题)

    Sort it 时间限制:1000 ms  |  内存限制:65535 KB 难度:2 描述 You want to processe a sequence of n distinct integer ...

  10. vue平行组件传值 params

    欢迎加入前端交流群交流知识&&获取视频资料:749539640 需要传值的组件组件: <template> <div> <router-link :to= ...