US8219725B2 - Cache optimized balanced handling of initiatives in a non-uniform multiprocessor computing system - Google Patents

Cache optimized balanced handling of initiatives in a non-uniform multiprocessor computing system Download PDF

Info

Publication number
US8219725B2
US8219725B2 US13/159,572 US201113159572A US8219725B2 US 8219725 B2 US8219725 B2 US 8219725B2 US 201113159572 A US201113159572 A US 201113159572A US 8219725 B2 US8219725 B2 US 8219725B2
Authority
US
United States
Prior art keywords
processors
group
initiatives
processor
balancing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
US13/159,572
Other versions
US20110314187A1 (en
Inventor
Udo ALBRECHT
Michael Jung
Elke Nass
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALBRECHT, UDO, JUNG, MICHAEL, NASS, ELKE G.
Publication of US20110314187A1 publication Critical patent/US20110314187A1/en
Application granted granted Critical
Publication of US8219725B2 publication Critical patent/US8219725B2/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/50Allocation of resources, e.g. of the central processing unit [CPU]
    • G06F9/5083Techniques for rebalancing the load in a distributed system

Definitions

  • the present invention relates to the field of high-performance, multiprocessor computer systems which are using an optimized Input/Output (I/O) subsystem that is based on dedicated System Assist Processors (SAPs) or I/O Processors (IOPs). These I/O processors perform communication between multiple Central Processing Units (CPUs) and many peripheral I/O devices attached via multiple I/O channel paths, e.g., for database business applications. More specifically, the present invention relates to processing I/O requests issued by the SAPs, and particularly, to how a request response issued by the I/O periphery is processed in the multiprocessor system.
  • I/O Input/Output
  • SAPs System Assist Processors
  • IOPs I/O Processors
  • an “initiative” is to be understood as a response to a former initiated I/O operation or as a request coming from an I/O device via an I/O path. Such a response is typically signalled via an I/O interrupt. In the prior art method these are the interrupts from the channels to the firmware running on the I/O processors.
  • a shortcoming of this prior art method is that with an increasing number of I/O processors the handling of initiatives leads to increased cache traffic and contention due to shared data structures, which slows down the initiative handling significantly. Further, it leads to increased firmware path length and control block contention due to an increasing number of processors detecting the same initiative at the same time, but only one processor is finally handling it.
  • a method for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system wherein in each node, a plurality of I/O processors communicate with I/O devices, via a plurality of I/O paths corresponding to said plurality of nodes, and wherein said initiatives are generated by I/O hardware and/or firmware addressed by a precedent I/O request issued by a respective one of said I/O processors, and wherein said initiatives are to be processed by one or by a group of said multiple I/O processors.
  • the method comprises using at least one pre-defined, first data element indicating incoming or existing initiatives from any I/O path of all of said nodes and to be served by one of said I/O processors; using a plurality of second data structures, one of said second structures per I/O processor group, wherein each of said second data structures defines which initiative is preferably handled by which I/O processor or group of processors, respectively; using a plurality of third data structures, one of said third structures per I/O processor group, wherein multiple bits can be set in order to indicate the occurrence of respective initiatives for respective multiple I/O processors; and repeatedly balancing the workload between the groups of I/O processors by distributing initiatives to other groups in addition to the preferred group.
  • FIG. 1 is a block diagram illustrating the basic structural components of a prior art hardware and software environment
  • FIG. 2 is a block diagram illustrating basic structural components of a hardware and software environment, employed in accordance with one or more aspects of the present invention
  • FIGS. 3A-3B depict a process for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system, in accordance with one or more aspects of the present invention
  • FIG. 4 illustrates one embodiment of data structures used in accordance with one or more aspects of the present invention
  • FIG. 5 illustrates a further embodiment of data structures used in accordance with one or more aspects of the present invention.
  • FIG. 6 is a block diagram of a computer system in which one or more aspects of the present invention may be implemented.
  • FIG. 1 illustrates the most basic structural components of prior art processor node hardware and software environment used for a prior art method in IBM's System z server products.
  • This product line includes large multiprocessor systems with a high number of I/O paths 2 attached via an I/O expansion network 4 to one or more nodes 10 , which are subsystems of the overall system.
  • Each node contains multiple Processing Units 12 (PUs), which in an IBM System z server are used as general CPUs or System Assist Processors (SAPs).
  • a shared cache 8 is provided amongst the multiple processors and nodes, with the node interconnect implementing a cache coherency protocol. Shared memory 6 is distributed across all nodes.
  • Step 1 is performed once at system startup or occasionally in hotplug scenarios, and described below.
  • Steps 2 and 3 are performed continuously and with high frequency in a running system with ongoing I/O operations.
  • the SAP system initialization allocates and initializes the Global Summary Vector 20 (GSV), the Node Local Summary Masks 22 (NLSM), and the Node Local Interrupt Vectors (NLIV) 24 .
  • GSV Global Summary Vector
  • NLSM Node Local Summary Masks
  • NLIV Node Local Interrupt Vectors
  • Broken line printed circles interconnected by bidirectional broken line arrows indicate bits or paths, respectively, corresponding to each other.
  • the GSV 20 e.g., 96 bits
  • Groups of (up to 64) channels are assigned one bit in this vector.
  • the NLIVs 24 are initialized to zero. There is one vector per node, and each channel physically attached to that node is assigned a unique bit position in that vector. One double word (64 bits) in the NLIV corresponds to one bit in the GSV. The NLIV is filled in by channels to signal initiative to the SAPs.
  • One NLSM 22 is defined per node. It has the same size as the GSV. A 1-bit in the NLSM indicates that the corresponding bit in the GSV describes summary initiative for channels of that node. GSV and NLSM are logically ANDed to identify those initiatives that are, for example, handled by SAPs on the local node. However, when no node-local initiative is found, a SAP will attempt to process initiative from the other nodes.
  • the advantages of the concepts disclosed herein are achieved by reducing the balanced initiative handling to an I/O processor group instead of all I/O processors in the system.
  • the prior art implementation handles “local” work, but it also handles “remote” work when no local initiative is pending.
  • invention all I/O processors of an I/O processor group always check and handle only “local” initiatives.
  • the I/O processor groups are defined to be all I/O processors on one node sharing the cache between all I/O processors of one group. Other embodiments can even group just part of the processors of one node. As the processors of one I/O processor group use shared caches this reduces cache contention. As the number of I/O processors in one group is smaller than the number of all I/O processors, this also reduces firmware path length and control block contention.
  • the “local” property is changed from physical channel locality to “processor with functional affinity” locality. This removes the requirement for a special Functional Affinity Work Indication (FAWI) vector.
  • FWI Functional Affinity Work Indication
  • the strict reduction to “local” work might lead to reduced overall throughput in case of peak or unbalanced workload, where one group of I/O processors becomes a bottleneck. Therefore, an additional balancing algorithm between the I/O processor groups is added on top.
  • the additional balancing algorithm enables step by step spreading the initiatives (workload) across multiple I/O processor groups on a group base as soon as the I/O processor group with maximum group utilization reaches a certain high limit together with other processor groups being utilized significantly lower. Also, the additional balancing is decreased step by step again when either a certain low utilization limit is reached or the workload becomes more evenly balanced between the processor groups again.
  • FIGS. 3A & 3B depict one embodiment of a method 300 that is performed by the firmware routine.
  • step 305 it is determined if there is more then one I/O processor group available. If this is not the case then no action needs to be taken since there is no other I/O processor group available for balancing. Otherwise, for the last measurement interval the minimum utilization (minU) is calculated in step 310 with the respective I/O processor group number (minG) and the maximum utilization (maxU) is calculated with the respective I/O processor group number (maxG). If applicable, then also the second maximum group and the second minimum group utilizations are calculated for potential later usage. Then in step 315 three entrance criteria are checked before the different balancing methods are started:
  • step 320 If no criterion is fulfilled, then the method 300 simply exits. Otherwise it is checked in step 320 if the I/O processor group maxG already helps any other groups. If this is the case (case1), then the additional work caused by the currently lowest utilized I/O processor group is released in step 325 to return step by step to the original workload on the I/O processor group maxG. Otherwise, it is checked in step 330 again if this is the increase balancing case, where maxU is higher than the IHB and maxU-minU is higher than the IHD. If this condition is true, it is checked in step 335 if the I/O processor group minG does not already help the I/O processor group maxG.
  • the I/O processor group minG is allowed in step 340 to do as well the work for the I/O processor group maxG (case2). If it is determined in step 335 that the I/O processor group minG already helps the I/O processor group maxG (case3) then any further additional balancing increase would not help as the I/O processor group minG has still enough capacity to help the I/O processor group maxG: The utilization of the I/O processor group maxG can be high due to other reasons than the backend handling. The method 300 is exited in this case.
  • step 330 If it was determined in step 330 that it is not the increase balancing case, then apparently no increase balancing is required. It will then be determined in step 345 if this is one of the decrease balancing cases, where maxU is lower than the DHB or maxU ⁇ minU is lower than the DHD. If not, then the method 300 exits. Otherwise it is checked in step 350 if more than two I/O processor groups are available. If so, then it is checked in step 355 if the second maximum I/O processor group 2 nd max helps any other I/O processor groups (case4), and if so, the additional work caused by the currently lowest utilized I/O processor group is released in step 360 to return step by step to the original workload on the I/O processor group 2 nd max.
  • step 355 it was determined in step 355 that it was not the case, then it will be determined in step 365 if the number of I/O processor groups is greater than 3. If so, then it is checked in step 370 if the second minimum I/O processor group 2 nd min helps any other I/O processor groups (case5), and if so, the additional work caused by the currently lowest utilized I/O processor group is released in step 375 to return step by step to the original workload on the I/O processor group 2 nd min. If it was determined in step 370 that it is not the case, then it is finally checked in step 380 if the I/O processor group minG with the lowest utilization can be released from any other work to reach a state where all I/O processor groups do only their own work. If so, then the work will be released in step 385 incrementally.
  • FIG. 4 illustrates data structures used in the preferred embodiment of the invention.
  • a GSV 40 is shown, which captures initiatives from channels.
  • the GSV 40 exists only once in the system and access to it is shared between the different SAPs.
  • the numbers 1, . . . , 4 in some of the fields represent the four I/O processor groups that the various channels have affinity to.
  • Block 45 shows four NLSM instances in the system.
  • an NLSM instance is a data structure, which exists once per group, is essentially constant and holds the same content for all processors within one group.
  • the combination of the GSV 40 with an NLSM via a bit-wise logical AND operation gives the pending initiatives for channels with affinity to any combination of any number of I/O processor groups.
  • each I/O processor group can be encoded efficiently.
  • one mask is used for each I/O processor group respectively, wherein each of field represents a set of channels with a maximum of 64 channels which all have the same affinity.
  • each I/O processor group does only care for its own work, namely the initiatives of the channels to which it has an affinity to. If it is determined that the amount of work for one I/O processor group will be increased after a pair of groups was detected that can help each other (one I/O processor group has the maximum utilization and the other I/O processor group has the minimum utilization), the mask of the I/O processor group with the minimum utilization will be combined with the content of the mask of the I/O processor group with the maximum utilization.
  • this is implemented very efficiently by using a pre-initialized initiative balancing mask table 50 as shown in FIG. 5 .
  • the initiative balancing mask table 50 exists only once in the system and is not changed. It contains masks of initiatives for channels with affinities to any combination of any number of I/O processor groups. As in FIG. 4 , the numbers 1, . . . , 4 in some of the fields represent the four I/O processor groups that the various channels have affinity to. This allows avoiding the expensive building of such combinations during the actual execution of the work as only table lookup operations need to be performed.
  • the method adapts only indices used for the table lookup. Such table index exists once per processor group.
  • the table indices 55 shown in FIG. 5 represent the four I/O processor groups. The table indices point to the currently active initiative mask for the respective group and are modified based on the load distribution.
  • the present invention may be embodied as a system, method or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, the present invention may take the form of a computer program product embodied in any tangible medium of expression having computer usable program code embodied in the medium.
  • the computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium.
  • the computer-readable medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CDROM), an optical storage device, a transmission media such as those supporting the Internet or an intranet, or a magnetic storage device.
  • a computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory.
  • a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
  • the computer-usable medium may include a propagated data signal with the computer-usable program code embodied therewith, either in baseband or as part of a carrier wave.
  • the computer usable program code may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc.
  • Computer program code for carrying out operations of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • LAN local area network
  • WAN wide area network
  • Internet Service Provider for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • These computer program instructions may also be stored in a computer-readable medium that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable medium produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • FIG. 6 illustrates a block diagram of a computer system 1300 in which certain embodiments may be implemented.
  • the system 1300 may include a circuitry 1302 that may in certain embodiments include a microprocessor 1304 .
  • the computer system 1300 may also include a memory 1306 (e.g., a volatile memory device), and storage 1308 .
  • the storage 1308 may include a non-volatile memory device (e.g., EEPROM, ROM, PROM, RAM, DRAM, SRAM, flash, firmware, programmable logic, etc.), magnetic disk drive, optical disk drive, tape drive, etc.
  • the storage 1308 may comprise an internal storage device, an attached storage device and/or a network accessible storage device.
  • the system 1300 may include a program logic 1310 including code 1312 that may be loaded into the memory 1306 and executed by the microprocessor 1304 or circuitry 1302 .
  • the program logic 1310 including code 1312 may be stored in the storage 1308 .
  • the program logic 1310 may be implemented in the circuitry 1302 . Therefore, while FIG. 6 shows the program logic 1310 separately from the other elements, the program logic 1310 may be implemented in the memory 1306 and/or the circuitry 1302 .
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.

Abstract

A balancing process between I/O processor groups of a non-uniform multiprocessor system enables spreading of I/O workload across multiple I/O processor groups on a group base as soon as the I/O processor group with maximum group utilization reaches a certain high limit together with other processor groups being utilized significantly lower. The additional balancing is decreased step by step again when a certain low utilization limit is reached or the workload becomes more evenly balanced between the I/O processor groups. Checking if increase or decrease of the balancing is required is done periodically, but with low frequency to not affect overall performance. The checking and balancing happens asynchronously in predefined intervals. This solves the problem that with an increasing number of I/O processors the handling of initiatives leads to increased cache traffic and contention due to shared data structures, which slows down the I/O workload handling significantly.

Description

PRIOR FOREIGN APPLICATION
This application claims priority from European Patent Application No. 10166090.0, filed Jun. 16, 2010, which is hereby incorporated herein by reference in its entirety.
BACKGROUND
The present invention relates to the field of high-performance, multiprocessor computer systems which are using an optimized Input/Output (I/O) subsystem that is based on dedicated System Assist Processors (SAPs) or I/O Processors (IOPs). These I/O processors perform communication between multiple Central Processing Units (CPUs) and many peripheral I/O devices attached via multiple I/O channel paths, e.g., for database business applications. More specifically, the present invention relates to processing I/O requests issued by the SAPs, and particularly, to how a request response issued by the I/O periphery is processed in the multiprocessor system.
An exemplary non-uniform multiprocessor computing system of the above mentioned shape is described in GB Application No. 2454996 A, which teaches a method for the balanced handling of inbound initiatives. In this context, an “initiative” is to be understood as a response to a former initiated I/O operation or as a request coming from an I/O device via an I/O path. Such a response is typically signalled via an I/O interrupt. In the prior art method these are the interrupts from the channels to the firmware running on the I/O processors.
A shortcoming of this prior art method is that with an increasing number of I/O processors the handling of initiatives leads to increased cache traffic and contention due to shared data structures, which slows down the initiative handling significantly. Further, it leads to increased firmware path length and control block contention due to an increasing number of processors detecting the same initiative at the same time, but only one processor is finally handling it.
Any potential solution to these n-way effect shortcomings would have to meet the following requirements: The initiatives need to be handled with as little cache and control block contention and overhead as possible, but peak or unbalanced workloads still need to be handled without any bottlenecks on one or some processors of the system.
BRIEF SUMMARY
According to one aspect of the present invention, a method for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system are proposed, wherein in each node, a plurality of I/O processors communicate with I/O devices, via a plurality of I/O paths corresponding to said plurality of nodes, and wherein said initiatives are generated by I/O hardware and/or firmware addressed by a precedent I/O request issued by a respective one of said I/O processors, and wherein said initiatives are to be processed by one or by a group of said multiple I/O processors. The method comprises using at least one pre-defined, first data element indicating incoming or existing initiatives from any I/O path of all of said nodes and to be served by one of said I/O processors; using a plurality of second data structures, one of said second structures per I/O processor group, wherein each of said second data structures defines which initiative is preferably handled by which I/O processor or group of processors, respectively; using a plurality of third data structures, one of said third structures per I/O processor group, wherein multiple bits can be set in order to indicate the occurrence of respective initiatives for respective multiple I/O processors; and repeatedly balancing the workload between the groups of I/O processors by distributing initiatives to other groups in addition to the preferred group.
Computer program products and systems relating to the above-summarized method are also described and claimed herein.
Additional features and advantages are realized through the techniques of the present invention. Other embodiments and aspects of the invention are described in detail herein and are considered a part of the claimed invention.
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
One or more aspects of the present invention are particularly pointed out and distinctly claimed as examples in the claims at the conclusion of the specification. The foregoing and other objects, features, and advantages of the invention are apparent from the following detailed description taken in conjunction with the accompanying drawings in which:
FIG. 1 is a block diagram illustrating the basic structural components of a prior art hardware and software environment;
FIG. 2 is a block diagram illustrating basic structural components of a hardware and software environment, employed in accordance with one or more aspects of the present invention;
FIGS. 3A-3B depict a process for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system, in accordance with one or more aspects of the present invention;
FIG. 4 illustrates one embodiment of data structures used in accordance with one or more aspects of the present invention;
FIG. 5 illustrates a further embodiment of data structures used in accordance with one or more aspects of the present invention; and
FIG. 6 is a block diagram of a computer system in which one or more aspects of the present invention may be implemented.
DETAILED DESCRIPTION
One embodiment of one or more aspects of the present invention uses the prior art system and method known from GB Application No. GB 2454996 A. FIG. 1 illustrates the most basic structural components of prior art processor node hardware and software environment used for a prior art method in IBM's System z server products. This product line includes large multiprocessor systems with a high number of I/O paths 2 attached via an I/O expansion network 4 to one or more nodes 10, which are subsystems of the overall system. Each node contains multiple Processing Units 12 (PUs), which in an IBM System z server are used as general CPUs or System Assist Processors (SAPs). A shared cache 8 is provided amongst the multiple processors and nodes, with the node interconnect implementing a cache coherency protocol. Shared memory 6 is distributed across all nodes.
In general, when applying a prior art method the following steps are performed:
1. Initialization;
2. I/O initiative signaling;
3. Selection and firmware execution.
Step 1 is performed once at system startup or occasionally in hotplug scenarios, and described below. Steps 2 and 3 are performed continuously and with high frequency in a running system with ongoing I/O operations. With reference to FIG. 2, as to the initialization step, the SAP system initialization allocates and initializes the Global Summary Vector 20 (GSV), the Node Local Summary Masks 22 (NLSM), and the Node Local Interrupt Vectors (NLIV) 24. Broken line printed circles interconnected by bidirectional broken line arrows indicate bits or paths, respectively, corresponding to each other. The GSV 20 (e.g., 96 bits) is initialized to zero. Groups of (up to 64) channels are assigned one bit in this vector.
The NLIVs 24 are initialized to zero. There is one vector per node, and each channel physically attached to that node is assigned a unique bit position in that vector. One double word (64 bits) in the NLIV corresponds to one bit in the GSV. The NLIV is filled in by channels to signal initiative to the SAPs.
One NLSM 22 is defined per node. It has the same size as the GSV. A 1-bit in the NLSM indicates that the corresponding bit in the GSV describes summary initiative for channels of that node. GSV and NLSM are logically ANDed to identify those initiatives that are, for example, handled by SAPs on the local node. However, when no node-local initiative is found, a SAP will attempt to process initiative from the other nodes.
The prior art method takes the following specific characteristics into consideration:
    • Distinction of local vs. remote nodes which corresponds to the “non-uniform” aspect, and
    • The need to honor a functional affinity for specific operations;
      wherein affinity is to be understood as the requirement to do initialization or to handle initiatives from an I/O path on a specific SAP.
The advantages of the concepts disclosed herein are achieved by reducing the balanced initiative handling to an I/O processor group instead of all I/O processors in the system. The prior art implementation handles “local” work, but it also handles “remote” work when no local initiative is pending. With the concepts disclosed herein, invention all I/O processors of an I/O processor group always check and handle only “local” initiatives. In one embodiment, the I/O processor groups are defined to be all I/O processors on one node sharing the cache between all I/O processors of one group. Other embodiments can even group just part of the processors of one node. As the processors of one I/O processor group use shared caches this reduces cache contention. As the number of I/O processors in one group is smaller than the number of all I/O processors, this also reduces firmware path length and control block contention.
Further, in one embodiment, the “local” property is changed from physical channel locality to “processor with functional affinity” locality. This removes the requirement for a special Functional Affinity Work Indication (FAWI) vector. The strict reduction to “local” work might lead to reduced overall throughput in case of peak or unbalanced workload, where one group of I/O processors becomes a bottleneck. Therefore, an additional balancing algorithm between the I/O processor groups is added on top. The additional balancing algorithm enables step by step spreading the initiatives (workload) across multiple I/O processor groups on a group base as soon as the I/O processor group with maximum group utilization reaches a certain high limit together with other processor groups being utilized significantly lower. Also, the additional balancing is decreased step by step again when either a certain low utilization limit is reached or the workload becomes more evenly balanced between the processor groups again.
Checking if increase or decrease of the balancing is required is done periodically but with low frequency to not affect overall performance: The checking and balancing happens asynchronously in predefined intervals. A firmware routine is called which does this independently from handling the initiatives. This routine is executed only on a dedicated I/O processor with low frequency to keep the performance impact as low as possible. For the balancing of the preferred embodiment six different scenarios case1, . . . , case6 are checked. As soon as a valid scenario is found, the associated action is taken and the actual balancing cycle ends without checking for other scenarios, since the actual change already influences the workloads for the different I/O processor groups.
FIGS. 3A & 3B depict one embodiment of a method 300 that is performed by the firmware routine. In step 305 it is determined if there is more then one I/O processor group available. If this is not the case then no action needs to be taken since there is no other I/O processor group available for balancing. Otherwise, for the last measurement interval the minimum utilization (minU) is calculated in step 310 with the respective I/O processor group number (minG) and the maximum utilization (maxU) is calculated with the respective I/O processor group number (maxG). If applicable, then also the second maximum group and the second minimum group utilizations are calculated for potential later usage. Then in step 315 three entrance criteria are checked before the different balancing methods are started:
    • Increase balancing case: maxU is higher than a predefined Increase Help Boundary (IHB) and the difference between maxU and minU is higher than a predefined Increase Help Difference (IHD); or
    • First decrease balancing case: maxU is lower than a predefined Decrease Help Boundary (DHB); or
    • Second decrease balancing case: maxU−minU is lower than a predefined Decrease Help Difference (DHD).
If no criterion is fulfilled, then the method 300 simply exits. Otherwise it is checked in step 320 if the I/O processor group maxG already helps any other groups. If this is the case (case1), then the additional work caused by the currently lowest utilized I/O processor group is released in step 325 to return step by step to the original workload on the I/O processor group maxG. Otherwise, it is checked in step 330 again if this is the increase balancing case, where maxU is higher than the IHB and maxU-minU is higher than the IHD. If this condition is true, it is checked in step 335 if the I/O processor group minG does not already help the I/O processor group maxG. If not (case2), then the I/O processor group minG is allowed in step 340 to do as well the work for the I/O processor group maxG (case2). If it is determined in step 335 that the I/O processor group minG already helps the I/O processor group maxG (case3) then any further additional balancing increase would not help as the I/O processor group minG has still enough capacity to help the I/O processor group maxG: The utilization of the I/O processor group maxG can be high due to other reasons than the backend handling. The method 300 is exited in this case.
If it was determined in step 330 that it is not the increase balancing case, then apparently no increase balancing is required. It will then be determined in step 345 if this is one of the decrease balancing cases, where maxU is lower than the DHB or maxU−minU is lower than the DHD. If not, then the method 300 exits. Otherwise it is checked in step 350 if more than two I/O processor groups are available. If so, then it is checked in step 355 if the second maximum I/O processor group 2nd max helps any other I/O processor groups (case4), and if so, the additional work caused by the currently lowest utilized I/O processor group is released in step 360 to return step by step to the original workload on the I/O processor group 2nd max.
If it was determined in step 355 that it was not the case, then it will be determined in step 365 if the number of I/O processor groups is greater than 3. If so, then it is checked in step 370 if the second minimum I/O processor group 2nd min helps any other I/O processor groups (case5), and if so, the additional work caused by the currently lowest utilized I/O processor group is released in step 375 to return step by step to the original workload on the I/O processor group 2nd min. If it was determined in step 370 that it is not the case, then it is finally checked in step 380 if the I/O processor group minG with the lowest utilization can be released from any other work to reach a state where all I/O processor groups do only their own work. If so, then the work will be released in step 385 incrementally.
In one embodiment, four different I/O processor groups can be used. Any other number of I/O processor groups can be supported in a similar fashion. FIG. 4 illustrates data structures used in the preferred embodiment of the invention. A GSV 40 is shown, which captures initiatives from channels. The GSV 40 exists only once in the system and access to it is shared between the different SAPs. The numbers 1, . . . , 4 in some of the fields represent the four I/O processor groups that the various channels have affinity to. Block 45 shows four NLSM instances in the system. As described above, an NLSM instance is a data structure, which exists once per group, is essentially constant and holds the same content for all processors within one group. The combination of the GSV 40 with an NLSM via a bit-wise logical AND operation gives the pending initiatives for channels with affinity to any combination of any number of I/O processor groups.
Similarly, the work of each I/O processor group can be encoded efficiently. In the preferred embodiment one mask is used for each I/O processor group respectively, wherein each of field represents a set of channels with a maximum of 64 channels which all have the same affinity. With normal workload each I/O processor group does only care for its own work, namely the initiatives of the channels to which it has an affinity to. If it is determined that the amount of work for one I/O processor group will be increased after a pair of groups was detected that can help each other (one I/O processor group has the maximum utilization and the other I/O processor group has the minimum utilization), the mask of the I/O processor group with the minimum utilization will be combined with the content of the mask of the I/O processor group with the maximum utilization. This leads to a mask for the group with the minimum utilization that encodes the work for both I/O processor groups. If, on the other hand, it is decided that the amount of work will be decreased for a particular I/O processor group, then it will be checked if the amount of work for this particular I/O processor group was previously increased by the work of another I/O processor group. If that is the case, then the additional workload can simply be removed by dividing the combined mask again.
In one embodiment, this is implemented very efficiently by using a pre-initialized initiative balancing mask table 50 as shown in FIG. 5. The initiative balancing mask table 50 exists only once in the system and is not changed. It contains masks of initiatives for channels with affinities to any combination of any number of I/O processor groups. As in FIG. 4, the numbers 1, . . . , 4 in some of the fields represent the four I/O processor groups that the various channels have affinity to. This allows avoiding the expensive building of such combinations during the actual execution of the work as only table lookup operations need to be performed. The method adapts only indices used for the table lookup. Such table index exists once per processor group. The table indices 55 shown in FIG. 5 represent the four I/O processor groups. The table indices point to the currently active initiative mask for the respective group and are modified based on the load distribution.
The following are achieved in the embodiment described herein:
    • 1. As each I/O processor of an I/O processor group is generally eligible to handle any initiative for that I/O processor group, workload is balanced across the available processor capacity of that I/O processor group, thereby improving overall throughput and latency.
    • 2. Workload from high-performance I/O paths whose traffic would exceed the capacity of one I/O processor is distributed across multiple I/O processors of an I/O processor group.
    • 3. As the workload is distributed honoring hardware structures (cache), optimum overall system performance is achieved.
    • 4. A FAWI vector is no longer required as the processor with functional affinity to the I/O path is always part of the I/O processor group assigned to the I/O path.
    • 5. Additional balancing between the I/O processor groups allows handling of peak or unbalanced workload with optimum throughput.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
As will be appreciated by one skilled in the art, the present invention may be embodied as a system, method or computer program product. Accordingly, the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, the present invention may take the form of a computer program product embodied in any tangible medium of expression having computer usable program code embodied in the medium.
Any combination of one or more computer usable or computer readable medium(s) may be utilized. The computer-usable or computer-readable medium may be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium. More specific examples (a non-exhaustive list) of the computer-readable medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CDROM), an optical storage device, a transmission media such as those supporting the Internet or an intranet, or a magnetic storage device. Note that the computer-usable or computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured, via, for instance, optical scanning of the paper or other medium, then compiled, interpreted, or otherwise processed in a suitable manner, if necessary, and then stored in a computer memory. In the context of this document, a computer-usable or computer-readable medium may be any medium that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The computer-usable medium may include a propagated data signal with the computer-usable program code embodied therewith, either in baseband or as part of a carrier wave. The computer usable program code may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc.
Computer program code for carrying out operations of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). The present invention is described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable medium that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable medium produce an article of manufacture including instruction means which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
FIG. 6 illustrates a block diagram of a computer system 1300 in which certain embodiments may be implemented. The system 1300 may include a circuitry 1302 that may in certain embodiments include a microprocessor 1304. The computer system 1300 may also include a memory 1306 (e.g., a volatile memory device), and storage 1308. The storage 1308 may include a non-volatile memory device (e.g., EEPROM, ROM, PROM, RAM, DRAM, SRAM, flash, firmware, programmable logic, etc.), magnetic disk drive, optical disk drive, tape drive, etc. The storage 1308 may comprise an internal storage device, an attached storage device and/or a network accessible storage device. The system 1300 may include a program logic 1310 including code 1312 that may be loaded into the memory 1306 and executed by the microprocessor 1304 or circuitry 1302. In certain embodiments, the program logic 1310 including code 1312 may be stored in the storage 1308. In certain other embodiments, the program logic 1310 may be implemented in the circuitry 1302. Therefore, while FIG. 6 shows the program logic 1310 separately from the other elements, the program logic 1310 may be implemented in the memory 1306 and/or the circuitry 1302.
The flowchart and block diagrams in the FIG.s illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below, if any, are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention.

Claims (15)

1. A method for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system, wherein in each node, a plurality of I/O processors communicate with I/O devices, via a plurality of I/O paths corresponding to said plurality of nodes, and wherein said initiatives are generated by I/O hardware and/or firmware addressed by a precedent I/O request issued by a respective one of said I/O processors, and wherein said initiatives are to be processed by one or by a group of said multiple I/O processors, the method comprising:
using at least one pre-defined, first data element indicating incoming or existing initiatives from any I/O path of all of said nodes and to be served by one of said I/O processors;
using a plurality of second data structures, one of said second structures per I/O processor group, wherein each of said second data structures defines which initiative is preferably handled by which I/O processor or group of processors, respectively;
using a plurality of third data structures, one of said third structures per I/O processor group, wherein multiple bits can be set in order to indicate the occurrence of respective initiatives for respective multiple I/O processors; and
repeatedly balancing the workload between the groups of I/O processors by distributing initiatives to other groups of I/O processors in addition to the preferred group.
2. The method of claim 1, wherein each group of I/O processors comprises only I/O processors that use a common shared processor cache (8) for the respective group.
3. The method of claim 2, wherein each group of I/O processors comprises only I/O processors of the same node for the respective group.
4. The method of claim 1, wherein the balancing is performed by an I/O processor asynchronously to the processing of the initiatives.
5. The method of claim 1, wherein the balancing comprises:
determining the groups of I/O processors with the minimum and maximum utilization; and
distributing initiatives from the group with the maximum utilization also to the group with the minimum utilization.
6. A computer program product for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system, wherein in each node, a plurality of I/O processors communicate with I/O devices via a plurality of I/O paths corresponding to said plurality of nodes, and wherein said initiatives are generated by I/O hardware and/or firmware addressed by a precedent I/O request issued by a respective one of said I/O processors, and wherein said initiatives are to be processed by one or by a group of said multiple I/O processors, the computer program product comprising:
a computer-readable storage medium readable by a processor and storing instructions for execution by the processor for performing a method comprising:
using at least one pre-defined, first data element indicating incoming or existing initiatives from any I/O path of all of said nodes and to be served by one of said I/O processors;
using a plurality of second data structures, one of said second structures per I/O processor group, wherein each of said second data structures defines which initiative is preferably handled by which I/O processor or group of processors, respectively;
using a plurality of third data structures, one of said third structures per I/O processor group, wherein multiple bits can be set in order to indicate the occurrence of respective initiatives for respective multiple I/O processors; and
repeatedly balancing the workload between the groups of I/O processors by distributing initiatives to other groups of I/O processors in addition to the preferred group.
7. The computer program product of claim 6, wherein each group of I/O processors comprises only I/O processors that use a common shared processor cache (8) for the respective group.
8. The computer program product of claim 7, wherein each group of I/O processors comprises only I/O processors of the same node for the respective group.
9. The computer program product of claim 6, wherein the balancing is performed by an I/O processor asynchronously to the processing of the initiatives.
10. The computer program product of claim 6, wherein the balancing comprises:
determining the groups of I/O processors with the minimum and maximum utilization; and
distributing initiatives from the group with the maximum utilization also to the group with the minimum utilization.
11. A computer system comprising means for handling inbound initiatives in an input/output (I/O) subsystem of a multi-node computer system, wherein in each node, a plurality of I/O processors communicate with I/O devices, via a plurality of I/O paths corresponding to said plurality of nodes, and wherein said initiatives are generated by I/O hardware and/or firmware addressed by a precedent I/O request issued by a respective one of said I/O processors, and wherein said initiatives are to be processed by one or by a group of said multiple I/O processors, the computer system comprising:
a memory; and
a processor in communications with the memory, wherein the computer system is configured to perform a method, the method comprising:
using at least one pre-defined, first data element indicating incoming or existing initiatives from any I/O path of all of said nodes and to be served by one of said I/O processors;
using a plurality of second data structures, one of said second structures per I/O processor group, wherein each of said second data structures defines which initiative is preferably handled by which I/O processor or group of processors, respectively;
using a plurality of third data structures, one of said third structures per I/O processor group, wherein multiple bits can be set in order to indicate the occurrence of respective initiatives for respective multiple I/O processors; and
repeatedly balancing the workload between the groups of I/O processors by distributing initiatives to other groups of I/O processors in addition to the preferred group.
12. The computer system of claim 11, wherein each group of I/O processors comprises only I/O processors that use a common shared processor cache for the respective group.
13. The computer system of claim 12, wherein each group of I/O processors comprises only I/O processors of the same node for the respective group.
14. The computer system of claim 11, wherein the balancing is performed by an I/O processor asynchronously to the processing of the initiatives.
15. The computer system of claim 11, wherein the balancing comprises:
determining the groups of I/O processors with the minimum and maximum utilization; and
distributing initiatives from the group with the maximum utilization also to the group with the minimum utilization.
US13/159,572 2010-06-16 2011-06-14 Cache optimized balanced handling of initiatives in a non-uniform multiprocessor computing system Expired - Fee Related US8219725B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
EP10166090.0 2010-06-16
EP10166090 2010-06-16
EP10166090 2010-06-16

Publications (2)

Publication Number Publication Date
US20110314187A1 US20110314187A1 (en) 2011-12-22
US8219725B2 true US8219725B2 (en) 2012-07-10

Family

ID=45329684

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/159,572 Expired - Fee Related US8219725B2 (en) 2010-06-16 2011-06-14 Cache optimized balanced handling of initiatives in a non-uniform multiprocessor computing system

Country Status (1)

Country Link
US (1) US8219725B2 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8413136B2 (en) 2009-05-08 2013-04-02 Microsoft Corporation Application virtualization

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5265215A (en) 1991-04-22 1993-11-23 International Business Machines Corporation Multiprocessor system and interrupt arbiter thereof
US5307495A (en) 1987-10-23 1994-04-26 Hitachi, Ltd. Multiprocessor system statically dividing processors into groups allowing processor of selected group to send task requests only to processors of selected group
US6189065B1 (en) * 1998-09-28 2001-02-13 International Business Machines Corporation Method and apparatus for interrupt load balancing for powerPC processors
US6601153B1 (en) * 1999-12-31 2003-07-29 Unisys Corporation Method and apparatus for increasing computer performance through asynchronous memory block initialization
CN101126992A (en) 2006-08-15 2008-02-20 国际商业机器公司 Method and system for dispensing multiple tasks at multiple node of network
US7444639B2 (en) * 2001-12-20 2008-10-28 Texas Insturments Incorporated Load balanced interrupt handling in an embedded symmetric multiprocessor system
GB2454996A (en) 2008-01-23 2009-05-27 Ibm Handling inbound initiatives for a multi-processor system by its input/output subsystem using data that defines which processor is to handle it.
US20090172228A1 (en) 2007-12-28 2009-07-02 Zimmer Vincent J Method and system for handling a management interrupt event in a multi-processor computing device
US20090172423A1 (en) 2007-12-31 2009-07-02 Justin Song Method, system, and apparatus for rerouting interrupts in a multi-core processor
US20090248934A1 (en) 2008-03-26 2009-10-01 International Business Machines Corporation Interrupt dispatching method in multi-core environment and multi-core processor
US7610425B2 (en) 2005-08-22 2009-10-27 Sun Microsystems, Inc. Approach for managing interrupt load distribution
WO2010020159A1 (en) 2008-08-19 2010-02-25 中兴通讯股份有限公司 Method and device for balancing interrupt load of multicore processor

Patent Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5307495A (en) 1987-10-23 1994-04-26 Hitachi, Ltd. Multiprocessor system statically dividing processors into groups allowing processor of selected group to send task requests only to processors of selected group
US5265215A (en) 1991-04-22 1993-11-23 International Business Machines Corporation Multiprocessor system and interrupt arbiter thereof
US6189065B1 (en) * 1998-09-28 2001-02-13 International Business Machines Corporation Method and apparatus for interrupt load balancing for powerPC processors
US6601153B1 (en) * 1999-12-31 2003-07-29 Unisys Corporation Method and apparatus for increasing computer performance through asynchronous memory block initialization
US7444639B2 (en) * 2001-12-20 2008-10-28 Texas Insturments Incorporated Load balanced interrupt handling in an embedded symmetric multiprocessor system
US7610425B2 (en) 2005-08-22 2009-10-27 Sun Microsystems, Inc. Approach for managing interrupt load distribution
CN101126992A (en) 2006-08-15 2008-02-20 国际商业机器公司 Method and system for dispensing multiple tasks at multiple node of network
US8001308B2 (en) * 2007-12-28 2011-08-16 Intel Corporation Method and system for handling a management interrupt event in a multi-processor computing device
US20090172228A1 (en) 2007-12-28 2009-07-02 Zimmer Vincent J Method and system for handling a management interrupt event in a multi-processor computing device
US20090172423A1 (en) 2007-12-31 2009-07-02 Justin Song Method, system, and apparatus for rerouting interrupts in a multi-core processor
US7962771B2 (en) * 2007-12-31 2011-06-14 Intel Corporation Method, system, and apparatus for rerouting interrupts in a multi-core processor
GB2454996A (en) 2008-01-23 2009-05-27 Ibm Handling inbound initiatives for a multi-processor system by its input/output subsystem using data that defines which processor is to handle it.
US20090248934A1 (en) 2008-03-26 2009-10-01 International Business Machines Corporation Interrupt dispatching method in multi-core environment and multi-core processor
US7953915B2 (en) * 2008-03-26 2011-05-31 International Business Machines Corporation Interrupt dispatching method in multi-core environment and multi-core processor
WO2010020159A1 (en) 2008-08-19 2010-02-25 中兴通讯股份有限公司 Method and device for balancing interrupt load of multicore processor
US20110145461A1 (en) * 2008-08-19 2011-06-16 Zte Corporation Method and device for balancing interrupt load of multicore processor

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Advanced Reversing, "Chapter 8: Advanced Programmable Interrupt Controller (APIC)", vol. 3, Chapter 8, 23 pages (Aug. 28, 2006).

Also Published As

Publication number Publication date
US20110314187A1 (en) 2011-12-22

Similar Documents

Publication Publication Date Title
US9058217B2 (en) Preferential CPU utilization for tasks
KR102386174B1 (en) Computer system comprising key-value device and method for providing interrupt handling
US9632958B2 (en) System for migrating stash transactions
US10614004B2 (en) Memory transaction prioritization
US8560803B2 (en) Dynamic cache queue allocation based on destination availability
US20190303344A1 (en) Virtual channels for hardware acceleration
US11556391B2 (en) CPU utilization for service level I/O scheduling
US10873630B2 (en) Server architecture having dedicated compute resources for processing infrastructure-related workloads
JP6244949B2 (en) Information processing apparatus, control method, and control program
US8671232B1 (en) System and method for dynamically migrating stash transactions
CN109716305B (en) Method, computing device, and medium for implementing asynchronous cache maintenance operations
US20200310690A1 (en) Dynamic near-data processing control mechanism based on computer resource availability on solid-state disk platforms
US10372643B1 (en) Bit-mapped DMA transfer with dependency table configured to monitor status so that a processor is not rendered as a bottleneck in a system
US8219725B2 (en) Cache optimized balanced handling of initiatives in a non-uniform multiprocessor computing system
US9858188B2 (en) Adaptively enabling and disabling snooping fastpath commands
US10958588B2 (en) Reliability processing of remote direct memory access
US9280485B2 (en) Efficient cache volume sit scans
US11494236B2 (en) System and method for sharing central processing unit (CPU) resources with unbalanced applications
US20180329756A1 (en) Distributed processing system, distributed processing method, and storage medium
GB2509505A (en) Issuing a write to a secondary storage device, when a write to a primary storage device takes too long
US20240028518A1 (en) Performance and reliability of processor store operation data transfers
US11841808B2 (en) System and method for processing requests in a multithreaded system

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ALBRECHT, UDO;JUNG, MICHAEL;NASS, ELKE G.;REEL/FRAME:026440/0143

Effective date: 20110614

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20160710