US8527713B2 - Cache locking without interference from normal allocations - Google Patents

Cache locking without interference from normal allocations Download PDF

Info

Publication number
US8527713B2
US8527713B2 US11/343,765 US34376506A US8527713B2 US 8527713 B2 US8527713 B2 US 8527713B2 US 34376506 A US34376506 A US 34376506A US 8527713 B2 US8527713 B2 US 8527713B2
Authority
US
United States
Prior art keywords
cache
instructions
entries
bnca
allocation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US11/343,765
Other versions
US20070180199A1 (en
Inventor
Victor Roberts Augsburg
James Norris Dieffenderfer
Jeffrey Todd Bridges
Thomas Andrew Sartorius
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qualcomm Inc
Original Assignee
Qualcomm Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qualcomm Inc filed Critical Qualcomm Inc
Priority to US11/343,765 priority Critical patent/US8527713B2/en
Assigned to QUALCOMM INCORPORATED reassignment QUALCOMM INCORPORATED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: AUGSBURG, VICTOR ROBERTS, BRIDGES, JEFFREY TODD, SARTORIUS, THOMAS ANDREW, DIEFFENDERFER, JAMES NORRIS
Priority to KR1020117023069A priority patent/KR20110117267A/en
Priority to PCT/US2007/061415 priority patent/WO2007090175A1/en
Priority to EP07762667.9A priority patent/EP1979819B1/en
Priority to CN200780003978.1A priority patent/CN101375257B/en
Priority to JP2008553491A priority patent/JP5123215B2/en
Priority to KR1020087021051A priority patent/KR101083192B1/en
Publication of US20070180199A1 publication Critical patent/US20070180199A1/en
Priority to JP2012185359A priority patent/JP2013016187A/en
Publication of US8527713B2 publication Critical patent/US8527713B2/en
Application granted granted Critical
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0888Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches using selective caching, e.g. bypass
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/30003Arrangements for executing specific machine instructions
    • G06F9/3004Arrangements for executing specific machine instructions to perform operations on memory
    • G06F9/30047Prefetch instructions; cache control instructions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/30Arrangements for executing machine instructions, e.g. instruction decode
    • G06F9/38Concurrent instruction execution, e.g. pipeline, look ahead
    • G06F9/3824Operand accessing

Definitions

  • the present disclosure relates generally to the field of processors and in particular to a method of cache locking without interference from normal allocations.
  • Microprocessors perform computational tasks in a wide variety of applications, including portable electronic devices. Maximizing processor performance can be desirable, to permit additional functions and features to be implemented in portable electronic devices and other applications.
  • Memory access instructions present a portion of a (virtual or physical) address, known as a tag, to a CAM structure in the cache, where it is compared against the corresponding address bits of instructions and/or data stored in the cache's RAM (the CAM and RAM contents together are referred to herein as a cache entry).
  • a cache line is returned that contains the desired instructions and/or data (or, in the case of set-associative caches, a plurality of cache lines are returned, and one is selected by a different portion of the address known as the index). If the tag fails to match any CAM entry, it “misses,” and the memory access instruction proceeds to access the desired instruction or data from main memory.
  • Some instruction set architectures provide the means to “lock” a cache entry, so that the entry is not replaced during normal cache miss processing. In some cases, the selection of cache entries to lock is explicit and precise, so as to minimize the impact of locked entries on normal cache allocation algorithms.
  • FIG. 1 is a functional block diagram depicting one representative means of locking cache entries.
  • the cache 1 which may comprise an instruction cache, a data cache, or a unified cache, includes n entries, numbered from 0 to n ⁇ 1.
  • a FLOOR register 2 holds the entry number that represents the “floor” of the cache, or the lowest cache entry available for normal allocation. Cache entries below the floor are not available for replacement, and are hence “locked.” If no entries are locked, the FLOOR register contains a 0, and the cache replacement algorithm operates throughout the cache. If, as depicted in FIG. 1 , the bottom three entries are locked, the processor will increment the FLOOR register to three, the first cache entry available for reallocation.
  • the normal cache reallocation algorithm in this case operates in the portion of the cache from the “floor,” or three, to the top of the cache, n ⁇ 1.
  • Grouping the locked cache entries in one place simplifies the replacement algorithm. For example, if cache entries are replaced on a round-robin basis, only the “rollover” point is affected by the locked entries (i.e., when incrementing past n ⁇ 1, the next entry is that pointed to by the FLOOR register 2 rather than 0). There are no non-contiguous, locked entries scattered across the cache space that must be “skipped over” by a round-robin allocation. Note that the FLOOR method of grouping and locking cache entries is representative only, and is not limiting. Cache entries may be grouped together and locked against reallocation according to a broad variety of methods.
  • Interrupts are, generally generated by events outside of the processor, and may be non-determinate in nature. Thus, interrupts may occur during the execution of code that attempts to carefully arrange locked cache entries.
  • the interrupt-handling code may include memory access instructions, that are likely to miss in the cache, causing accesses to main memory. These memory accesses will normally generate allocations in the cache. That is, instructions and data fetched to service the interrupt will replace some cache lines. If the interrupts occur after locked cache entry processing has begun, but before the locked entries are established and arranged, cache entries that were meant to be locked may be reallocated. Additionally, non-locked cache entries may be allocated in an area intended for locked entries, such as below the FLOOR register. This may result in non-contiguous locked entries, imposing a significant burden on some cache replacement algorithms, such as round-robin.
  • a Block Normal Cache Allocation (BNCA) mode is defined.
  • cache entries may only be allocated by predetermined instructions. Normal memory access instructions (for example, as part of interrupt code) may execute and will retrieve data from main memory in the event of a cache miss; however, these instructions are not allowed to allocate entries in the cache. Only the predetermined instructions (for example, those used to establish locked cache entries) may allocate entries in the cache. When the locked entries are established, the processor exits BNCA mode, and any memory access instruction may allocate cache entries.
  • BNCA mode may be indicated by setting a bit in a configuration register.
  • One embodiment relates to a method of restricting allocations in a cache.
  • a restricted mode only a predetermined set of instructions is allowed to allocate cache entries.
  • a non-restricted mode any memory access instruction is allowed to allocate cache entries.
  • Another embodiment relates to a method of managing a cache.
  • a restricted mode is entered wherein only predetermined instructions may allocate cache entries. Cache entries are allocated using the predetermined instructions.
  • the restricted mode is exited and cache entries for any memory access instruction are allocated.
  • the processor includes a cache memory and a configuration register having a Block Normal Cache Allocation (BNCA) bit.
  • the processor additionally includes a cache controller operative to control the allocation of entries in the cache by memory access instructions, and further operative to disallow said allocation, except for a predetermined set of instructions, when the BNCA bit is set.
  • BNCA Block Normal Cache Allocation
  • FIG. 1 is a functional block diagram of a cache memory and a FLOOR register.
  • FIG. 2 is a functional block diagram of a processor.
  • FIG. 3 is a flow diagram of a method of locking cache entries.
  • FIG. 2 depicts a functional block diagram of a representative processor 10 , employing both a pipelined architecture and a hierarchical memory structure.
  • the processor 10 executes instructions in an instruction execution pipeline 12 according to control logic 14 .
  • control logic 14 includes a configuration register 15 , which includes a Block Normal Cache Allocation (BNCA) bit 16 , described in greater detail below.
  • the pipeline includes various registers or latches 17 , organized in pipe stages, and one or more Arithmetic Logic Units (ALU) 18 .
  • a General Purpose Register (GPR) file 20 provides registers comprising the top of the memory hierarchy.
  • GPR General Purpose Register
  • the pipeline fetches instructions from an Instruction Cache (I-cache) 22 , with memory addressing and permissions managed by an Instruction-side Translation Lookaside Buffer (ITLB) 24 .
  • Data is accessed from a Data Cache (D-cache) 26 , with memory addressing and permissions managed by a main Translation Lookaside Buffer (TLB) 28 .
  • the ITLB 24 may comprise a copy of part of the TLB 28 .
  • the ITLB 24 and TLB 28 may be integrated.
  • the I-cache 22 and D-cache 26 may be integrated, or unified. Misses in the I-cache 22 and/or the D-cache 26 cause an access to main (off-chip) memory 32 , under the control of a memory interface 30 .
  • the processor 10 may include an Input/Output (I/O) interface 34 , controlling access to various peripheral devices 36 .
  • I/O Input/Output
  • the processor 10 may include a second-level (L2) cache for either or both the I and D caches 22 , 26 .
  • L2 second-level cache for either or both the I and D caches 22 , 26 .
  • one or more of the functional blocks depicted in the processor 10 may be omitted from a particular embodiment.
  • a PLD (preload) instruction reads an address in main memory. If the PLD misses in the relevant cache, the data from memory is read and loaded into the cache, replacing an existing cache entry.
  • the PLD instruction may be used by a program to preload certain critical instructions and/or data into an I-cache 22 and/or D-cache 26 , respectfully, to ensure that subsequent code accessing the instructions and/or data will hit in the relevant cache 22 , 26 .
  • the program may use the PLD instruction in conjunction with writing to the FLOOR register (or other locking mechanism) to lock the preloaded entries in the cache 22 , 26 against replacement by other memory access operations. However, an interrupt received during this processing may upset the desired preloading and cache entry locking operation.
  • the processor 10 may enter a mode wherein only predetermined instructions may allocate cache entries.
  • the mode may be entered by, for example, writing a Block Normal Cache Allocation (BNCA) bit 16 , such as in a configuration register 15 .
  • BNCA Block Normal Cache Allocation
  • the BNCA indicator 16 is set or active, common memory access instructions are prevented from allocating (e.g., replacing) entries in the cache 22 , 26 .
  • Only select, predetermined instructions may allocate cache entries. These instructions may include, for example, the PLD instructions for allocating cache entries without reading data into the pipeline. Such instructions are unlikely to be executed by interrupt handling code, and hence interrupt code is unlikely to upset the preloading and cache entry locking operations performed in BNCA mode.
  • LPLD LPLD
  • an interrupt does occur during BNCA mode—that is, during a cache preloading and locking operation—any memory access instructions will executed properly; they are merely prevented from allocating cache entries. That is, if a memory access instruction misses in the cache 22 , 26 , a main memory access will be performed, and the instruction or data returned to the pipeline. However, the cache miss processing will be unable to allocate a cache entry. Consequently, if an access to the same or nearby instructions or data is subsequently attempted, it will again miss in the cache, and retrieve the instruction or data from main memory. This will degrade the execution performance of the interrupt routine by introducing repeated, long memory access latencies. However, the interrupt routine is assumed to be both short and rare. On the other hand, considerable performance gains and design simplification are achieved by ensuring that precise, locked cache entries may be established without upset by memory access instructions executing as part of interrupt code.
  • the process of establishing locked cache entries is described with reference to the flow diagram of FIG. 3 .
  • the BNCA bit 16 (or other BNCA mode indicator) is set (block 40 ). This indicates to the circuits controlling the I-cache 22 and D-cache 26 that only certain predetermined instructions (e.g., PLD) are allowed to allocate cache entries (block 42 ). This allows the processor 10 to establish precise, fixed cache entries, located so as to optimize the normal cache allocation procedure. If an interrupt occurs during BNCA mode, the interrupt code is executed, and all memory access instructions (other than the predetermined instructions) are disallowed from allocating cache entries (block 44 ). When the interrupt code completes, the processor 10 continues preloading and locking cache entries.
  • predetermined instructions e.g., PLD
  • the processor 10 exits the BNCA mode and clears the BNCA bit 16 (or other BNCA mode indicator) (block 48 ). The processor 10 then continues execution, allowing any memory access instruction to allocate cache entries per the normal allocation protocol (block 50 ). In this manner, the establishment of precise, locked cache entries is protected from upset by interrupt code memory access instructions.
  • a single BNCA bit 16 controls BNCA mode for all caches 22 , 26 .
  • a separate BNCA bit 16 controls BNCA mode separately for the I-cache 22 and the D-cache 26 .
  • one or both of the I-cache 22 and D-cache 26 may be divided into two or more banks, and each cache bank may independently enter BNCA mode, as indicated by corresponding BNCA bits 16 .
  • each set of an n-way set associative cache 22 , 26 may have an independently enter and exit a BNCA mode, as indicated by corresponding BNCA bits 16 .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Memory System Of A Hierarchy Structure (AREA)

Abstract

A Block Normal Cache Allocation (BNCA) mode is defined for a processor. In BNCA mode, cache entries may only be allocated by predetermined instructions. Normal memory access instructions (for example, as part of interrupt code) may execute and will retrieve data from main memory in the event of a cache miss; however, these instructions are not allowed to allocate entries in the cache. Only the predetermined instructions (for example, those used to establish locked cache entries) may allocate entries in the cache. When the locked entries are established, the processor exits BNCA mode, and any memory access instruction may allocate cache entries. BNCA mode may be indicated by setting a bit in a configuration register.

Description

FIELD
The present disclosure relates generally to the field of processors and in particular to a method of cache locking without interference from normal allocations.
BACKGROUND
Microprocessors perform computational tasks in a wide variety of applications, including portable electronic devices. Maximizing processor performance can be desirable, to permit additional functions and features to be implemented in portable electronic devices and other applications.
Due to the spatial and temporal locality characteristics of common in computer programs, the instructions and data being processed at any given time are statistically likely to be needed in the very near future, and may be retained in a high-speed, cache memory, where they are quickly available. Memory access instructions present a portion of a (virtual or physical) address, known as a tag, to a CAM structure in the cache, where it is compared against the corresponding address bits of instructions and/or data stored in the cache's RAM (the CAM and RAM contents together are referred to herein as a cache entry). If the tag matches, or “hits,” a cache line is returned that contains the desired instructions and/or data (or, in the case of set-associative caches, a plurality of cache lines are returned, and one is selected by a different portion of the address known as the index). If the tag fails to match any CAM entry, it “misses,” and the memory access instruction proceeds to access the desired instruction or data from main memory.
It is commonly desirable to maximize the overall cache hit rate, thus minimizing off-chip accesses to main memory, which can incur latency, stall the pipeline, and consume additional power. Additionally, in some applications, a few critical or often used instructions and/or data may be known, and this information may be maintained in the cache, regardless of the cache's overall performance. Some instruction set architectures provide the means to “lock” a cache entry, so that the entry is not replaced during normal cache miss processing. In some cases, the selection of cache entries to lock is explicit and precise, so as to minimize the impact of locked entries on normal cache allocation algorithms.
FIG. 1 is a functional block diagram depicting one representative means of locking cache entries. The cache 1, which may comprise an instruction cache, a data cache, or a unified cache, includes n entries, numbered from 0 to n−1. A FLOOR register 2 holds the entry number that represents the “floor” of the cache, or the lowest cache entry available for normal allocation. Cache entries below the floor are not available for replacement, and are hence “locked.” If no entries are locked, the FLOOR register contains a 0, and the cache replacement algorithm operates throughout the cache. If, as depicted in FIG. 1, the bottom three entries are locked, the processor will increment the FLOOR register to three, the first cache entry available for reallocation. The normal cache reallocation algorithm in this case operates in the portion of the cache from the “floor,” or three, to the top of the cache, n−1.
Grouping the locked cache entries in one place simplifies the replacement algorithm. For example, if cache entries are replaced on a round-robin basis, only the “rollover” point is affected by the locked entries (i.e., when incrementing past n−1, the next entry is that pointed to by the FLOOR register 2 rather than 0). There are no non-contiguous, locked entries scattered across the cache space that must be “skipped over” by a round-robin allocation. Note that the FLOOR method of grouping and locking cache entries is representative only, and is not limiting. Cache entries may be grouped together and locked against reallocation according to a broad variety of methods.
Interrupts are, generally generated by events outside of the processor, and may be non-determinate in nature. Thus, interrupts may occur during the execution of code that attempts to carefully arrange locked cache entries. The interrupt-handling code may include memory access instructions, that are likely to miss in the cache, causing accesses to main memory. These memory accesses will normally generate allocations in the cache. That is, instructions and data fetched to service the interrupt will replace some cache lines. If the interrupts occur after locked cache entry processing has begun, but before the locked entries are established and arranged, cache entries that were meant to be locked may be reallocated. Additionally, non-locked cache entries may be allocated in an area intended for locked entries, such as below the FLOOR register. This may result in non-contiguous locked entries, imposing a significant burden on some cache replacement algorithms, such as round-robin.
SUMMARY
In one or more embodiments, a Block Normal Cache Allocation (BNCA) mode is defined. When the processor enters BNCA mode, cache entries may only be allocated by predetermined instructions. Normal memory access instructions (for example, as part of interrupt code) may execute and will retrieve data from main memory in the event of a cache miss; however, these instructions are not allowed to allocate entries in the cache. Only the predetermined instructions (for example, those used to establish locked cache entries) may allocate entries in the cache. When the locked entries are established, the processor exits BNCA mode, and any memory access instruction may allocate cache entries. BNCA mode may be indicated by setting a bit in a configuration register.
One embodiment relates to a method of restricting allocations in a cache. In a restricted mode, only a predetermined set of instructions is allowed to allocate cache entries. In a non-restricted mode, any memory access instruction is allowed to allocate cache entries.
Another embodiment relates to a method of managing a cache. A restricted mode is entered wherein only predetermined instructions may allocate cache entries. Cache entries are allocated using the predetermined instructions. The restricted mode is exited and cache entries for any memory access instruction are allocated.
Another embodiment relates to a processor. The processor includes a cache memory and a configuration register having a Block Normal Cache Allocation (BNCA) bit. The processor additionally includes a cache controller operative to control the allocation of entries in the cache by memory access instructions, and further operative to disallow said allocation, except for a predetermined set of instructions, when the BNCA bit is set.
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is a functional block diagram of a cache memory and a FLOOR register.
FIG. 2 is a functional block diagram of a processor.
FIG. 3 is a flow diagram of a method of locking cache entries.
DETAILED DESCRIPTION
FIG. 2 depicts a functional block diagram of a representative processor 10, employing both a pipelined architecture and a hierarchical memory structure. The processor 10 executes instructions in an instruction execution pipeline 12 according to control logic 14. Included in control logic 14 is a configuration register 15, which includes a Block Normal Cache Allocation (BNCA) bit 16, described in greater detail below. The pipeline includes various registers or latches 17, organized in pipe stages, and one or more Arithmetic Logic Units (ALU) 18. A General Purpose Register (GPR) file 20 provides registers comprising the top of the memory hierarchy.
The pipeline fetches instructions from an Instruction Cache (I-cache) 22, with memory addressing and permissions managed by an Instruction-side Translation Lookaside Buffer (ITLB) 24. Data is accessed from a Data Cache (D-cache) 26, with memory addressing and permissions managed by a main Translation Lookaside Buffer (TLB) 28. In various embodiments, the ITLB 24 may comprise a copy of part of the TLB 28. Alternatively, the ITLB 24 and TLB 28 may be integrated. Similarly, in various embodiments of the processor 10, the I-cache 22 and D-cache 26 may be integrated, or unified. Misses in the I-cache 22 and/or the D-cache 26 cause an access to main (off-chip) memory 32, under the control of a memory interface 30.
The processor 10 may include an Input/Output (I/O) interface 34, controlling access to various peripheral devices 36. Those of skill in the art will recognize that numerous variations of the processor 10 are possible. For example, the processor 10 may include a second-level (L2) cache for either or both the I and D caches 22, 26. In addition, one or more of the functional blocks depicted in the processor 10 may be omitted from a particular embodiment.
Commonly instruction set architectures to include memory access instructions that only allocate cache entries, without reading instructions or data into the pipeline. For example, a PLD (preload) instruction reads an address in main memory. If the PLD misses in the relevant cache, the data from memory is read and loaded into the cache, replacing an existing cache entry. The PLD instruction may be used by a program to preload certain critical instructions and/or data into an I-cache 22 and/or D-cache 26, respectfully, to ensure that subsequent code accessing the instructions and/or data will hit in the relevant cache 22, 26. The program may use the PLD instruction in conjunction with writing to the FLOOR register (or other locking mechanism) to lock the preloaded entries in the cache 22, 26 against replacement by other memory access operations. However, an interrupt received during this processing may upset the desired preloading and cache entry locking operation.
According to one or more embodiments, the processor 10 may enter a mode wherein only predetermined instructions may allocate cache entries. The mode may be entered by, for example, writing a Block Normal Cache Allocation (BNCA) bit 16, such as in a configuration register 15. When the BNCA indicator 16 is set or active, common memory access instructions are prevented from allocating (e.g., replacing) entries in the cache 22, 26. Only select, predetermined instructions may allocate cache entries. These instructions may include, for example, the PLD instructions for allocating cache entries without reading data into the pipeline. Such instructions are unlikely to be executed by interrupt handling code, and hence interrupt code is unlikely to upset the preloading and cache entry locking operations performed in BNCA mode. Alternatively, in and instruction set architecture that provides for explicit cache locking, such as for example a LPLD (locking preload) instruction, cache allocations during BNCA mode may be restricted to such locking cache allocation instructions.
If an interrupt does occur during BNCA mode—that is, during a cache preloading and locking operation—any memory access instructions will executed properly; they are merely prevented from allocating cache entries. That is, if a memory access instruction misses in the cache 22, 26, a main memory access will be performed, and the instruction or data returned to the pipeline. However, the cache miss processing will be unable to allocate a cache entry. Consequently, if an access to the same or nearby instructions or data is subsequently attempted, it will again miss in the cache, and retrieve the instruction or data from main memory. This will degrade the execution performance of the interrupt routine by introducing repeated, long memory access latencies. However, the interrupt routine is assumed to be both short and rare. On the other hand, considerable performance gains and design simplification are achieved by ensuring that precise, locked cache entries may be established without upset by memory access instructions executing as part of interrupt code.
The process of establishing locked cache entries according to one embodiment is described with reference to the flow diagram of FIG. 3. The BNCA bit 16 (or other BNCA mode indicator) is set (block 40). This indicates to the circuits controlling the I-cache 22 and D-cache 26 that only certain predetermined instructions (e.g., PLD) are allowed to allocate cache entries (block 42). This allows the processor 10 to establish precise, fixed cache entries, located so as to optimize the normal cache allocation procedure. If an interrupt occurs during BNCA mode, the interrupt code is executed, and all memory access instructions (other than the predetermined instructions) are disallowed from allocating cache entries (block 44). When the interrupt code completes, the processor 10 continues preloading and locking cache entries. When all desired locked cache entries have been established (block 46), the processor 10 exits the BNCA mode and clears the BNCA bit 16 (or other BNCA mode indicator) (block 48). The processor 10 then continues execution, allowing any memory access instruction to allocate cache entries per the normal allocation protocol (block 50). In this manner, the establishment of precise, locked cache entries is protected from upset by interrupt code memory access instructions.
In one embodiment, a single BNCA bit 16 controls BNCA mode for all caches 22, 26. According to another embodiment, a separate BNCA bit 16 controls BNCA mode separately for the I-cache 22 and the D-cache 26. In yet another embodiment, one or both of the I-cache 22 and D-cache 26 may be divided into two or more banks, and each cache bank may independently enter BNCA mode, as indicated by corresponding BNCA bits 16. In still another embodiment, each set of an n-way set associative cache 22, 26 may have an independently enter and exit a BNCA mode, as indicated by corresponding BNCA bits 16. In general, those of skill in the art will readily recognize that the teachings of one or more embodiments discloses herein may be applied to numerous alternative implementations.
Although embodiments of the present disclosure has been described herein with respect to particular features, aspects and embodiments thereof, it will be apparent that numerous variations, modifications, and other embodiments are possible within the broad scope of the present teachings. The present embodiments are therefore to be construed in all aspects as illustrative and not restrictive and all changes coming within the meaning and equivalency range of the appended claims are intended to be embraced therein.

Claims (21)

What is claimed is:
1. A method of restricting allocations in a cache, comprising:
entering a restricted mode and allowing only a predetermined set of one or more types of instructions to allocate cache entries in the restricted mode; and
exiting the restricted mode to enter a non-restricted mode and allowing any memory access instruction to allocate cache entries in the non-restricted mode.
2. The method of claim 1 wherein the restricted mode is entered by setting a Block Normal Cache Allocation (BNCA) indicator, and exited by resetting the BNCA indicator.
3. The method of claim 2 wherein the BNCA indicator is a bit in a configuration register.
4. The method of claim 1 wherein the predetermined set of one or more types of instructions includes types of instructions operative to load data into the cache and lock the cache entry against replacement.
5. The method of claim 4 wherein separate types of instructions are operative to load and lock instruction cache and data cache entries.
6. The method of claim 1 wherein the restricted and non-restricted modes are separately applicable to a data cache and an instruction cache.
7. The method of claim 1 wherein the cache is divided into two or more banks, and wherein restricted and non-restricted modes are separately applicable to each bank.
8. The method of claim 1 wherein the cache is a set associative cache, and wherein restricted and non-restricted modes are separately applicable to each set.
9. A method of managing a cache, comprising:
entering a restricted mode wherein only one or more predetermined types of instructions may allocate cache entries;
allocating cache entries using the predetermined types of instructions; and
exiting the restricted mode and allocating cache entries for any memory access instruction.
10. The method of claim 9 further comprising, after allocating cache entries using the predetermined types of instructions, locking the allocated cache entries against replacement.
11. The method of claim 9 further comprising:
processing an interrupt received during the restricted mode; and
performing memory accesses for memory access instructions in the interrupt code, without allocating any cache entries for the memory access instructions in the interrupt code.
12. The method of claim 9 wherein entering the restricted mode comprises setting a Block Normal Cache Allocation (BNCA) bit in a configuration register, and wherein exiting the restricted mode comprises clearing the BNCA bit.
13. A processor, comprising:
a cache memory;
a configuration register having a Block Normal Cache Allocation (BNCA) bit; and
a cache controller operative to control the allocation of entries in the cache by memory access instructions, and further operative to disallow said allocation, except for a predetermined set of one or more types of instructions, when the BNCA bit is set.
14. The processor of claim 13 wherein said predetermined set of one or more types of instructions includes locking allocation instructions.
15. The method of claim 1, wherein each type of instruction is associated with a different type of operation to be performed.
16. The method of claim 15, wherein the type of operation to be performed is specified by an operation code of the associated instruction.
17. The method of claim 9, wherein each type of instruction is associated with a different type of operation to be performed.
18. The processor of claim 13, wherein each type of instruction is associated with a different type of operation to be performed.
19. A processor, comprising:
means for storing cache entries;
means for storing a Block Normal Cache Allocation (BNCA) bit; and
means for controlling the allocation of entries in the means for storing cache entries by memory access instructions; and
means for disallowing said allocation, except for a predetermined set of one or more types of instructions, when the BNCA bit is set.
20. The processor of claim 19 wherein said predetermined set of one or more types of instructions includes locking allocation instructions.
21. The processor of claim 19, wherein each type of instruction is associated with a different type of operation to be performed.
US11/343,765 2006-01-31 2006-01-31 Cache locking without interference from normal allocations Expired - Fee Related US8527713B2 (en)

Priority Applications (8)

Application Number Priority Date Filing Date Title
US11/343,765 US8527713B2 (en) 2006-01-31 2006-01-31 Cache locking without interference from normal allocations
CN200780003978.1A CN101375257B (en) 2006-01-31 2007-01-31 Cache locking without interference from normal allocation
PCT/US2007/061415 WO2007090175A1 (en) 2006-01-31 2007-01-31 Cache locking without interference from normal allocation
EP07762667.9A EP1979819B1 (en) 2006-01-31 2007-01-31 Cache locking without interference from normal allocation
KR1020117023069A KR20110117267A (en) 2006-01-31 2007-01-31 Cache locking without interference from normal allocation
JP2008553491A JP5123215B2 (en) 2006-01-31 2007-01-31 Cache locking without interference from normal allocation
KR1020087021051A KR101083192B1 (en) 2006-01-31 2007-01-31 Cache locking without interference from normal allocation
JP2012185359A JP2013016187A (en) 2006-01-31 2012-08-24 Cache locking without interference from normal allocations

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/343,765 US8527713B2 (en) 2006-01-31 2006-01-31 Cache locking without interference from normal allocations

Publications (2)

Publication Number Publication Date
US20070180199A1 US20070180199A1 (en) 2007-08-02
US8527713B2 true US8527713B2 (en) 2013-09-03

Family

ID=38123878

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/343,765 Expired - Fee Related US8527713B2 (en) 2006-01-31 2006-01-31 Cache locking without interference from normal allocations

Country Status (6)

Country Link
US (1) US8527713B2 (en)
EP (1) EP1979819B1 (en)
JP (2) JP5123215B2 (en)
KR (2) KR20110117267A (en)
CN (1) CN101375257B (en)
WO (1) WO2007090175A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9773476B2 (en) 2014-10-22 2017-09-26 Samsung Electronics Co., Ltd. Cache memory system for a pixel shader and method of operating the same

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2586589C2 (en) * 2011-12-30 2016-06-10 Интел Корпорейшн Module for coprocessor cache
CN102591799B (en) * 2011-12-30 2015-04-15 华为技术有限公司 Method and device for data storage
US9141544B2 (en) 2012-06-26 2015-09-22 Qualcomm Incorporated Cache memory with write through, no allocate mode
US20230418753A1 (en) * 2022-06-28 2023-12-28 Advanced Micro Devices, Inc. Allocation control for cache

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01237835A (en) 1988-01-20 1989-09-22 Advanced Micro Devicds Inc Single-chip integrated cash unit architecture and implementation of cash function thereinside
US5185878A (en) 1988-01-20 1993-02-09 Advanced Micro Device, Inc. Programmable cache memory as well as system incorporating same and method of operating programmable cache memory
JPH05216687A (en) 1991-10-28 1993-08-27 Internatl Business Mach Corp <Ibm> Method and apparatus for controlling interrupting cache operation
US5249286A (en) * 1990-05-29 1993-09-28 National Semiconductor Corporation Selectively locking memory locations within a microprocessor's on-chip cache
US5822764A (en) 1996-03-04 1998-10-13 Motorola, Inc. Method and circuit for efficiently replacing invalid locked portions of a cache with valid data
JP2001005658A (en) 1999-06-18 2001-01-12 Toshiba Corp Built-in type data processing system, and method for correcting system program
US20020007441A1 (en) * 1998-03-31 2002-01-17 Salvador Palanca Shared cache structure for temporal and non-temporal instructions
US6438655B1 (en) * 1999-04-20 2002-08-20 Lucent Technologies Inc. Method and memory cache for cache locking on bank-by-bank basis
US20050044320A1 (en) * 2003-08-19 2005-02-24 Sun Microsystems, Inc. Cache bank interface unit
US20050055506A1 (en) * 2003-09-04 2005-03-10 International Business Machines Corporation Pseudo-LRU for a locking cache
KR20050070012A (en) 2002-09-27 2005-07-05 어드밴스드 마이크로 디바이시즈, 인코포레이티드 Computer system with integrated directory and processor cache
US20050251630A1 (en) * 2004-05-04 2005-11-10 Matthews Jeanna N Preventing storage of streaming accesses in a cache

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4056768B2 (en) * 2002-03-04 2008-03-05 富士通株式会社 Microcomputer, cache memory control method, and clock control method

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5185878A (en) 1988-01-20 1993-02-09 Advanced Micro Device, Inc. Programmable cache memory as well as system incorporating same and method of operating programmable cache memory
JPH01237835A (en) 1988-01-20 1989-09-22 Advanced Micro Devicds Inc Single-chip integrated cash unit architecture and implementation of cash function thereinside
US5249286A (en) * 1990-05-29 1993-09-28 National Semiconductor Corporation Selectively locking memory locations within a microprocessor's on-chip cache
JPH05216687A (en) 1991-10-28 1993-08-27 Internatl Business Mach Corp <Ibm> Method and apparatus for controlling interrupting cache operation
US5371872A (en) * 1991-10-28 1994-12-06 International Business Machines Corporation Method and apparatus for controlling operation of a cache memory during an interrupt
US5822764A (en) 1996-03-04 1998-10-13 Motorola, Inc. Method and circuit for efficiently replacing invalid locked portions of a cache with valid data
US20020007441A1 (en) * 1998-03-31 2002-01-17 Salvador Palanca Shared cache structure for temporal and non-temporal instructions
US6438655B1 (en) * 1999-04-20 2002-08-20 Lucent Technologies Inc. Method and memory cache for cache locking on bank-by-bank basis
JP2001005658A (en) 1999-06-18 2001-01-12 Toshiba Corp Built-in type data processing system, and method for correcting system program
KR20050070012A (en) 2002-09-27 2005-07-05 어드밴스드 마이크로 디바이시즈, 인코포레이티드 Computer system with integrated directory and processor cache
US20050044320A1 (en) * 2003-08-19 2005-02-24 Sun Microsystems, Inc. Cache bank interface unit
US20050055506A1 (en) * 2003-09-04 2005-03-10 International Business Machines Corporation Pseudo-LRU for a locking cache
US20050251630A1 (en) * 2004-05-04 2005-11-10 Matthews Jeanna N Preventing storage of streaming accesses in a cache

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
International Search Report-PCT/US07/061415, International Search Authority-European Patent Office-Jun. 26, 2007.
International Search Report—PCT/US07/061415, International Search Authority—European Patent Office—Jun. 26, 2007.
Written Opinion-PCT/US07/061415, International Search Authority-European Patent Office-Jun. 26, 2007.
Written Opinion—PCT/US07/061415, International Search Authority—European Patent Office—Jun. 26, 2007.

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9773476B2 (en) 2014-10-22 2017-09-26 Samsung Electronics Co., Ltd. Cache memory system for a pixel shader and method of operating the same

Also Published As

Publication number Publication date
CN101375257B (en) 2014-06-18
CN101375257A (en) 2009-02-25
JP2013016187A (en) 2013-01-24
JP2009525552A (en) 2009-07-09
KR20110117267A (en) 2011-10-26
KR20080097206A (en) 2008-11-04
JP5123215B2 (en) 2013-01-23
EP1979819B1 (en) 2019-12-18
EP1979819A1 (en) 2008-10-15
WO2007090175A1 (en) 2007-08-09
KR101083192B1 (en) 2011-11-11
US20070180199A1 (en) 2007-08-02

Similar Documents

Publication Publication Date Title
US9747218B2 (en) CPU security mechanisms employing thread-specific protection domains
US7996644B2 (en) Fair sharing of a cache in a multi-core/multi-threaded processor by dynamically partitioning of the cache
US8606998B2 (en) System and method for instruction-based cache allocation policies
US6470422B2 (en) Buffer memory management in a system having multiple execution entities
US5809530A (en) Method and apparatus for processing multiple cache misses using reload folding and store merging
US7426626B2 (en) TLB lock indicator
US8250332B2 (en) Partitioned replacement for cache memory
US7493451B2 (en) Prefetch unit
EP3129886B1 (en) Dynamic cache replacement way selection based on address tag bits
EP3055774B1 (en) Multi-mode set associative cache memory dynamically configurable to selectively allocate into all or a subset of its ways depending on the mode
US20090106499A1 (en) Processor with prefetch function
US20060248280A1 (en) Prefetch address generation implementing multiple confidence levels
US20060026403A1 (en) Compare instruction
US20060248281A1 (en) Prefetching using hashed program counter
US20060248279A1 (en) Prefetching across a page boundary
US20080320228A1 (en) Method and apparatus for efficient replacement algorithm for pre-fetcher oriented data cache
US6487639B1 (en) Data cache miss lookaside buffer and method thereof
KR20120070584A (en) Store aware prefetching for a data stream
US10417134B2 (en) Cache memory architecture and policies for accelerating graph algorithms
US20100011165A1 (en) Cache management systems and methods
JP2019096309A (en) Execution of maintenance operation
US8527713B2 (en) Cache locking without interference from normal allocations
US20040243765A1 (en) Multithreaded processor with multiple caches
US6785797B2 (en) Address predicting apparatus and methods
US8423719B2 (en) Apparatus, processor and method of controlling cache memory

Legal Events

Date Code Title Description
AS Assignment

Owner name: QUALCOMM INCORPORATED, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AUGSBURG, VICTOR ROBERTS;SARTORIUS, THOMAS ANDREW;DIEFFENDERFER, JAMES NORRIS;AND OTHERS;REEL/FRAME:017335/0100;SIGNING DATES FROM 20060223 TO 20060224

Owner name: QUALCOMM INCORPORATED, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AUGSBURG, VICTOR ROBERTS;SARTORIUS, THOMAS ANDREW;DIEFFENDERFER, JAMES NORRIS;AND OTHERS;SIGNING DATES FROM 20060223 TO 20060224;REEL/FRAME:017335/0100

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20210903