US20080288725A1 - Method and apparatus for cache transactions in a data processing system - Google Patents

Method and apparatus for cache transactions in a data processing system Download PDF

Info

Publication number
US20080288725A1
US20080288725A1 US11/748,353 US74835307A US2008288725A1 US 20080288725 A1 US20080288725 A1 US 20080288725A1 US 74835307 A US74835307 A US 74835307A US 2008288725 A1 US2008288725 A1 US 2008288725A1
Authority
US
United States
Prior art keywords
cache
bus
cache entry
request
state
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/748,353
Inventor
William C. Moyer
Michael D. Snyder
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NXP USA Inc
Original Assignee
Freescale Semiconductor Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Freescale Semiconductor Inc filed Critical Freescale Semiconductor Inc
Priority to US11/748,353 priority Critical patent/US20080288725A1/en
Priority to US11/748,350 priority patent/US8972671B2/en
Assigned to FREESCALE SEMICONDUCTOR, INC. reassignment FREESCALE SEMICONDUCTOR, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MOYER, WILLIAM C., SNYDER, MICHAEL D.
Assigned to CITIBANK, N.A., AS COLLATERAL AGENT reassignment CITIBANK, N.A., AS COLLATERAL AGENT SECURITY AGREEMENT Assignors: FREESCALE SEMICONDUCTOR, INC.
Priority to JP2010508485A priority patent/JP2010527488A/en
Priority to KR1020097023703A priority patent/KR20100016516A/en
Priority to CN200880015903XA priority patent/CN101918923A/en
Priority to PCT/US2008/061887 priority patent/WO2008144179A1/en
Publication of US20080288725A1 publication Critical patent/US20080288725A1/en
Assigned to CITIBANK, N.A. reassignment CITIBANK, N.A. SECURITY AGREEMENT Assignors: FREESCALE SEMICONDUCTOR, INC.
Assigned to CITIBANK, N.A., AS COLLATERAL AGENT reassignment CITIBANK, N.A., AS COLLATERAL AGENT SECURITY AGREEMENT Assignors: FREESCALE SEMICONDUCTOR, INC.
Assigned to FREESCALE SEMICONDUCTOR, INC. reassignment FREESCALE SEMICONDUCTOR, INC. PATENT RELEASE Assignors: CITIBANK, N.A., AS COLLATERAL AGENT
Assigned to FREESCALE SEMICONDUCTOR, INC. reassignment FREESCALE SEMICONDUCTOR, INC. PATENT RELEASE Assignors: CITIBANK, N.A., AS COLLATERAL AGENT
Assigned to FREESCALE SEMICONDUCTOR, INC. reassignment FREESCALE SEMICONDUCTOR, INC. PATENT RELEASE Assignors: CITIBANK, N.A., AS COLLATERAL AGENT
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • G06F12/02Addressing or allocation; Relocation
    • G06F12/08Addressing or allocation; Relocation in hierarchically structured memory systems, e.g. virtual memory systems
    • G06F12/0802Addressing of a memory level in which the access to the desired data or data block requires associative addressing means, e.g. caches
    • G06F12/0806Multiuser, multiprocessor or multiprocessing cache systems
    • G06F12/0815Cache consistency protocols
    • G06F12/0831Cache consistency protocols using a bus scheme, e.g. with bus monitoring or watching means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F12/00Accessing, addressing or allocating within memory systems or architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/28Error detection; Error correction; Monitoring by checking the correct order of processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/36Preventing errors by testing or debugging software
    • G06F11/362Software debugging
    • G06F11/3648Software debugging using additional hardware
    • G06F11/3656Software debugging using additional hardware using a specific debug interface
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F13/38Information transfer, e.g. on bus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity

Definitions

  • This disclosure relates generally to a cache, and more specifically, to cache transactions in a data processing system.
  • FIG. 1 illustrates, in block diagram form, a data processing system in accordance with one embodiment of the present invention.
  • FIG. 2 illustrates, in block diagram form, a portion of processor 11 and/or processor 12 of FIG. 1 in accordance with one embodiment of the present invention.
  • FIG. 3 illustrates, in tabular form, a listing of cache state definitions in accordance with one embodiment of the present invention.
  • FIG. 4 illustrates, in state diagram form, a portion of a cache coherency state diagram in accordance with one embodiment of the present invention.
  • FIG. 5 illustrates, in tabular form, a plurality of snoop bus commands in accordance with one embodiment of the present invention.
  • FIG. 6 illustrates, in tabular form, information provided on a bus during snoop bus commands in accordance with one embodiment of the present invention.
  • a plurality of new snoop bus commands or new snoop transaction types, and the supporting logic and framework for using them is described herein.
  • “show line” and “show doubleword” snoop transactions are provided to allow debug circuitry (e.g. 14 of FIG. 1 ) to gain access to the latest values of variables, even if they are cached, or to extract cache contents.
  • these snoop transactions act like existing snoop transaction types in some respects (e.g. in that snoop lookups are performed by participating bus masters). However, for the new snoop bus commands, the result of the lookup will cause a snoop response transaction to be requested on any hit, not just on a hit to a modified line.
  • the response causes the contents of a cached doubleword or an entire cache line to be placed on a bus (e.g. 20 of FIG. 1 ), and thus available for capture by the debug circuitry (e.g. 14 of FIG. 1 ).
  • the new snoop transaction types include “show_line”, “show doubleword”, “show_target_line”, and “show_target_doubleword”. Alternate embodiments may have fewer, more, or different snoop transactions, may use different names for the snoop transactions, or may use modified versions of the snoop transactions.
  • bus is used to refer to a plurality of signals or conductors which may be used to transfer one or more various types of information, such as data, addresses, control, or status.
  • the conductors as discussed herein may be illustrated or described in reference to being a single conductor, a plurality of conductors, unidirectional conductors, or bidirectional conductors. However, different embodiments may vary the implementation of the conductors. For example, separate unidirectional conductors may be used rather than bidirectional conductors and vice versa.
  • plurality of conductors may be replaced with a single conductor that transfers multiple signals serially or in a time multiplexed manner. Likewise, single conductors carrying multiple signals may be separated out into various different conductors carrying subsets of these signals. Therefore, many options exist for transferring signals.
  • FIG. 1 illustrates, in block diagram form, a data processing system 10 in accordance with one embodiment of the present invention.
  • data processing system 10 includes a processor 12 , a debug circuit or debug circuitry 14 , an input/output (I/O) circuit 16 , and a memory 18 , all bi-directionally coupled to a bus 20 .
  • the memory 18 may be any type of memory, such as, for example, a read only memory (ROM), a random access memory (RAM), non-volatile memory (e.g. Flash), etc. Also, memory 18 may be a memory or other data storage located within another peripheral or slave or on a different integrated circuit.
  • ROM read only memory
  • RAM random access memory
  • Flash non-volatile memory
  • memory 18 may be a memory or other data storage located within another peripheral or slave or on a different integrated circuit.
  • processor 11 , processor 12 and debug circuitry 14 are bus masters.
  • I/O circuitry 16 may also be a bus master in some embodiments.
  • processor 12 is a processor capable of executing instructions, such as a microprocessor, digital signal processor, etc., or may be any other type of bus master, such as for example, a direct memory access (DMA) controller, a bus bridge, or the like.
  • DMA direct memory access
  • other slave devices may be included in data processing system 10 that are not illustrated in FIG. 1 , such as for example, another memory or cache memory coupled to bus 20 , as well as any type of peripheral circuit which resides on the system bus or is coupled to I/O circuit 16 .
  • debug circuitry 14 is also bi-directionally coupled to processors 11 and 12 by way of conductors 21 . In alternate embodiments, debug circuitry 14 may also be bi-directionally coupled to one or more of I/O circuitry 16 and memory 18 using one or more conductors (not shown) that are independent of bus 20 .
  • processor 11 includes a cache 15 which is bi-directionally coupled to bus 20 by way of bus interface unit (BIU) 17 .
  • Processor 12 includes a cache 13 and a cache 23 which are each bi-directionally coupled to bus 20 by way of bus interface unit (BIU) 19 .
  • cache 13 functions as a level one (L1) cache for storing data and/or instructions for use by processor 12
  • cache 23 functions as a level two (L2) cache for storing data and/or instructions for use by processor 12 .
  • L1 cache level one
  • L2 level two
  • data and/or instructions are loaded into caches 13 , 23 from memory 18 before being executed by the processor.
  • Caches 13 , 23 may comprise a separate data cache and a separate instruction cache or may comprise a unified instruction and data cache, or may comprise only data caches.
  • Caches 13 , 23 comprise one or more data arrays, one or more tag arrays, and one or more status bit arrays.
  • the tag and status information may be stored in a common array.
  • Each cache entry consists of a block or line of data, a tag value which associates the data with a memory address, and status information.
  • the status information may include whether the cache entry is valid, whether the cache entry is “dirty” (i.e. modified with respect to the data associated with the same address in memory 18 or other external memory blocks if present), and whether the cache entry is exclusive or shared with other bus masters. Alternate embodiments may include less, more, or different cache status information.
  • memory 18 , or a portion of memory 18 may be characterized as a level two (L2) cache.
  • I/O module 16 is bi-directionally coupled to the bus 20 and to external circuitry (not shown).
  • the I/O module 16 includes various interface circuits depending upon the type of external circuitry that is coupled to the I/O module 16 .
  • I/O module 16 may contain one or more alternate bus masters coupled to bus 20 , and may contain one or more caches.
  • I/O module 16 may serve as a bus bridge, and be connected to one or more devices incorporating one or more caches via conductors 27 .
  • Debug circuit 14 is bi-directionally coupled to one or more of bus processor 11 , processor 12 , I/O circuitry 16 , and/or memory 18 via bus 20 . Debug circuit 14 may also be bi-directionally coupled to external circuitry (not shown) such as a hardware tester or other debugging interface hardware. In one embodiment, debug circuitry 14 complies with the NEXUS debug protocol. For some embodiments, debug circuitry 14 also complies with the JTAG protocol. Alternate embodiments may use debug circuitry 14 that is complies with any desired debug protocol.
  • debug circuitry 14 functions as a diagnostic check for data processing operations related to an access to caches 15 , 13 , 23 , memory 18 , and I/O module 16 , and in other embodiments, other types of data retention circuits utilized by data processing system 10 whether internal to or external to data processing system 10 .
  • the processor 12 and alternate bus masters obtain mastership of the system bus 20 as needed and selectively access the system memory 18 to retrieve and store data and instructions.
  • Debug circuitry 14 may be configured by the user of data processing system 10 to capture the values of one or more memory locations by monitoring addresses presented on bus 20 , and capturing or sampling the related data values presented on bus 20 . During bus read operations, the data values are provided by a selected slave device such as memory 18 .
  • the data values are provided by a bus master of bus 20 such as processor 12 .
  • debug circuitry 14 monitors each transfer and selectively captures data values corresponding to data locations the user of debug circuitry 14 wishes to monitor. These values and conditions may be programmed into and transferred from debug circuitry 14 via one or more terminals 25 illustrated in FIG. 1 .
  • I/O circuitry 16 may also be coupled external to data processing system 10 by way of one or more terminals 27 .
  • terminals 25 and 27 may be integrated circuit pins.
  • terminals 25 and 27 may be anything that can be used to transfer signals external to data processing system 10 , such as, for example, conductive bumps, conductive pads, wires, etc.
  • the data value may be modified with respect to the corresponding value in memory 18 , i.e. the latest (most up to date) value is not present in memory 18 , but only within cache 13 . If debug circuitry 14 accesses memory 18 to obtain the desired variable's value, it will not get a correct copy, since the copy may be modified internally in cache 13 .
  • a new bus transaction type is provided in data processing system 10 to allow debug circuitry 14 to “see” a value which may be stored in a cache (e.g. cache 13 ).
  • the bus transaction type is a “show” transaction.
  • the bus transaction can be initiated by debug circuitry 14 acting as a bus master of bus 20 .
  • the bus transaction includes a desired address location to be accessed. It is determined (e.g. by cache control circuitry 54 of FIG. 2 ) if the address or cache entry results in a “hit” or a “miss” in the cache. When the address results in a hit, the value is provided to bus 20 associated with the address from the cache 13 .
  • Debug circuitry 14 then has access to the value from the bus 20 as it is provided by the processor (e.g. 12 ).
  • debug circuitry 14 monitors transactions on bus 20 , and is capable of sampling the address and associated data values as bus transfers occur. Desired values can then be transferred after sampling to a user of the system 10 performing a debugging operation by means of signals from debug circuitry 14 of FIG. 1 (e.g. via terminals 25 ).
  • a state of cache 13 is not modified when the value is placed on the bus 20 .
  • a “dirty bit” “D” is associated with a cache entry indicates whether or not data in the cache associated with the effective address is different (i.e. has been modified) from data stored at another memory location, such as memory 18 , corresponding to the data address.
  • a “dirty bit” associated with the address of the value is not modified when the value is retrieved from the cache and presented to the system bus 20 in response to executing a “show” bus transaction.
  • cache 13 is not modified if the value associated with the effective address in cache 13 is the same as the value in another memory location, i.e. the data is clean.
  • the data stored in the entry is provided to bus 20 and the entry is not modified.
  • a dirty bit of the hit cache entry indicates that data stored in the entry is dirty
  • the data stored in the entry is still provided to bus 20 , and the entry is not modified, and the dirty bit is not cleared. This is different than the normal operation of a “dirty bit” in a cache.
  • FIG. 2 illustrates, in block diagram form, a portion of processor 11 and/or processor 12 of FIG. 1 .
  • FIG. 2 illustrates one possible embodiment of cache 13 and BIU 19 , and/or cache 23 and BIU 19 , and/or cache 15 and BIU 17 , all of FIG. 1 , in more detail.
  • cache 13 uses the embodiment illustrated in FIG. 2 .
  • cache 15 and/or cache 23 may also use the cache embodiment illustrated in FIG. 2
  • cache 13 may use a cache embodiment different than that illustrated in FIG. 2 .
  • alternate embodiments of system 10 may use different architectures, structures, and/or circuitry to implement the various caches (e.g.
  • cache 13 includes cache control circuitry 54 and cache array 50 which are bi-directionally coupled to each other.
  • cache array 50 is an array of random access memory cells organized as a plurality of entries.
  • cache array 50 comprises SRAM (static random access memory) memory cells. In other embodiments, other memory types are suitable.
  • entry 52 is representative of the plurality of entries and includes a tag address bit field labeled “TAG ADDRESS FIELD”, a tag status bit field labeled “TAG STATUS BITS” and a data bit field labeled “DATA”.
  • the tag status bit field includes a valid bit labeled “V”, an exclusive/shared bit labeled “E”, and a dirty bit labeled “D”.
  • each entry may be referred to as a “cache line” for processor 12 .
  • Cache array 50 is bi-directionally coupled to bus 20 via BIU 19 .
  • cache control circuitry 54 selectively determines if an effective address pointed to by an instruction received from, for example, cache 13 , memory 18 , or debug circuitry 14 is a hit or a miss in cache 13 .
  • Alternate embodiments may use a translation look-up buffer (TLB) (not shown) coupled to cache 13 for performing address translation from an effective address to a physical address. This translation may be done in a standard manner that is known in the art.
  • TLB translation look-up buffer
  • Cache control circuitry 54 can determine if an effective address pointed to by an instruction results in a hit or miss by comparing the physical address for an access provided by the TLB to the stored value(s) contained in one or more tag entries in cache array 50 .
  • a TLB is not used and the effective address is provided directly to cache 13 without translation.
  • stored tag values correspond directly to effective addresses without address translation to a physical address being required.
  • cache 13 is bi-directionally coupled to BIU 19 .
  • BIU 19 is bi-directionally coupled to bus 20 .
  • bus 20 includes signals 30 and signals 32 .
  • signals 30 include a plurality of address signals 40 , a plurality of data signals 41 , one or more transfer_type signals 42 , one or more target_specifier signals 43 , and one or more cache_specifier signals 44 .
  • FIG. 3 illustrates, in tabular form, a listing of cache state definitions in accordance with one embodiment of the present invention. Alternate embodiments of the present invention may use fewer, more, or different cache states than those illustrated in FIG. 3 .
  • FIG. 3 illustrates a MESI (Modified Exclusive Shared and Invalid) cache protocol having four states, namely an invalid state (INV), a shared state (S), an exclusive unmodified state (EU), and an exclusive modified state (EM).
  • MESI Modified Exclusive Shared and Invalid
  • Other embodiments may use different cache protocols and their associated cache state definitions. Thus, the present invention is not limited in any way to the specific states illustrated in FIG. 3 .
  • FIG. 4 illustrates, in state diagram form, a portion of a cache coherency state diagram in accordance with one embodiment of the present invention.
  • the state diagram illustrated in FIG. 4 shows how a cache implementing the cache states of FIG. 3 can transition between the four cache states INV, S, EU, and EM.
  • New state transitions have been incorporated into the prior art state diagram for a MESI cache coherency protocol to support the new snoop transaction types (show_line, show_dw, show 13 target_line, show_target_dw) for one embodiment of the present invention.
  • FIG. 4 illustrates, in state diagram form, a portion of a cache coherency state diagram in accordance with one embodiment of the present invention.
  • the state diagram illustrated in FIG. 4 shows how a cache implementing the cache states of FIG. 3 can transition between the four cache states INV, S, EU, and EM.
  • New state transitions have been incorporated into the prior art state diagram for a MESI cache coherency protocol to support the new
  • the “show line” and “show doubleword” snoop transactions or bus commands are provided to allow debug circuitry 14 to gain access to the latest values of variables, even if they are cached, or to allow debug circuitry 14 to extract cache contents.
  • these snoop transactions may act like existing snoop transaction types in most respects, in that snoop lookups are performed by participating bus masters (e.g. 11 , 12 ), but the result of the lookup will optionally cause a snoop response transaction to be requested on any hit, not just on a hit to a modified or dirty cache line.
  • a specific cache target is identified as part of the transaction information, and for these specific command types, the result of the lookup will optionally cause a snoop response transaction to occur regardless of a hit or miss. Utilizing these snoop command types allows for visibility into the current state of a targeted cache without causing state changes of data, tag, or status information.
  • the cache control circuitry 54 In response to the request portion of the “show line”, “show doubleword”, “show target line”, and “show target doubleword” bus transactions, the cache control circuitry 54 (see FIG. 2 ) causes the contents of a cached doubleword or an entire line to be placed on bus 20 , and thus available for capture by the debug circuitry 14 .
  • This exposure occurs regardless of the presence of backing store for the variables, i.e. there may be no physical memory associated with the address of the variable other than in a cache itself.
  • By exposing the value of a variable in memory to the system bus 20 it can be captured by data trace logic in debug circuitry 14 and messaged out to the user via terminals 25 .
  • instructions are unobtrusive to the data cache, and no data cache state changes occur, regardless of the state of the variable (Modified, Exclusive, Shared, or Invalid).
  • options are provided to cause either an entire cache line to be placed on bus 20 , or to limit the exposure to a data bus-width element (e.g. a doubleword) containing the desired variable.
  • Alternate embodiments may instead use one snoop transaction type with a fixed width or size (e.g. line), or may use one bus command that has within it an encoding to select among a variable group of widths or sizes (e.g. line, doubleword, word, etc.)
  • cache control circuitry 54 allows a cache line or a portion of a cache line to be broadcast on a bus 20 when it is present in the cache (e.g. 13 ), regardless of whether it is clean (i.e. unmodified) or dirty (i.e. modified).
  • cache state logic see FIG. 4
  • cache control circuitry 54 see FIG. 2
  • memory e.g. 18
  • system 10 adds an additional bus transaction type “show” which allows debug circuitry 14 visibility into internal data cache state (e.g. of cache 13 ), which may be modified with respect to memory (e.g. 18 ).
  • “show” when “show” is presented as a snoop read burst transaction, a normal snoop lookup is performed. If the transaction address hits in the data cache (e.g. cache 13 ), then a snoop copyback is performed, regardless of the modified or clean state of the cache line. The copyback is marked as a “show line” write, and in one embodiment is a burst of four doublewords to the bus. The state of the cache line in all processors (e.g. 11 , 12 ) remains unchanged.
  • the providing of the cache line to the bus 20 allows the debug circuitry 14 to see the content of the cache line, and thus to transfer these values out to an external debugger (not shown) via terminals 25 .
  • updates to memory e.g. 18
  • doubleword snoop bus commands are supported. When presented as a snoop single-beat read transaction, a normal snoop lookup is also performed. These “doubleword” bus commands operate similarly to the “line” bus commands, except that only a single doubleword of data in a hitting cache line is provided, via a single-beat write bus transaction which is marked as “doubleword”.
  • the “show_target_line” and “show —target _doubleword” snoop bus commands are typically targeted to only a single master (e.g. processor 12 ) in the system 10 , and thus only a single response will occur for each of these transactions.
  • these commands are simultaneously broadcast to multiple masters (e.g. processors 11 and 12 ), but only a single master needs to respond for the data to become visible to the debug circuitry 14 .
  • masters e.g. processors 11 and 12
  • cache lines in the exclusive state this occurs naturally.
  • shared cache lines which are valid in multiple caches only a single “show” bus transaction is required to provide visibility of the data, even though multiple cached copies exist.
  • transitions 50 - 53 may be used to provide information from any cache (e.g. cache 13 ) to debug circuitry 14 (see FIG. 1 ).
  • the transfer-type signals 42 on bus 20 are used to indicate when one of the bus snoop transactions listed in FIG. 5 are currently taking place on bus 20 .
  • each snoop bus transaction includes a request portion 60 , a response portion 61 , and a data portion 62 that all are provided on bus 20 .
  • a bus master e.g.
  • debug circuitry 14 initiates a read to a cache (e.g. cache 13 ).
  • a cache e.g. cache 13
  • the cache responds with a write operation.
  • the data stored in the cache that corresponds to the address or cache entry specified in the request 60 portion is provided on bus 20 .
  • bus snoop transactions listed in FIG. 5 will now be described. Note that alternate embodiments may have other additional snoop transactions that are not listed in FIG. 5 .
  • a bus master e.g. debug circuitry 14
  • a bus master e.g. debug circuitry 14
  • provides a snoop transaction request 60 which includes a desired address by way of signals 40 of bus 20 (see FIG. 2 ), and provides a transaction type indication by way of one or more signals 42 .
  • the transaction is monitored by one or more caches in the system which participate in normal snooping of bus transactions for coherency purposes.
  • a selected cache in the system performs a “show copyback” (for responding to a “show cache line” request) or a “show doubleword” (for responding to a “show doubleword” request) (see FIG. 5 ) write transaction of the cache line data information corresponding to the address of the request transaction regardless of the state of the dirty bit D, and the cache remains in its current state, i.e. the dirty bit is unchanged. In addition, all other caches remain in their respective current states. In one embodiment, memory 18 is not updated with the data provided on the response, even though a write transaction is indicated on the bus 20 .
  • the response 61 includes status information from the cache and the cache entry that corresponds to the address in the request 60 . If there was a cache hit, data is provided via data conductors 41 during the data portion 62 . If there was a cache miss, data is not provided via data conductors 41 during the data portion 62 . In an alternate embodiment, no status, or only partial status information may be provided during the response. In one embodiment, if the requested address misses in all of the caches participating in the transaction, data may be provided by memory 18 instead of no data being provided during data portion 62 of the transaction. In one embodiment, if multiple caches “hit”, a single cache is selected to provide the response, and the remaining caches cancel any pending responses.
  • a bus master e.g. debug circuitry 14
  • a bus master provides a request 60 which includes a target specifier by way of signals 43 of bus 20 (see FIG. 2 ), provides a cache specifier by way of signals 44 , provides an address by way of signals 40 , and provides a width indicator DW (e.g. doubleword, line, etc.) by way of one or more signals 32 .
  • the target specifier specifies which block or portion of circuitry in system 10 has a cache and is the “target” of this bus transaction.
  • the cache specifier specifies which cache within the “target” is being accessed for this bus transaction.
  • the specified cache in the target performs a “show target copyback” (see FIG. 5 ) of the cache line or a “show target doubleword” of the requested doubleword, regardless of the state of the dirty bit D, and the cache remains in its current state.
  • the response 61 includes status information from the cache and the cache entry that corresponds to the address in the request 60 . If there was a cache hit, data is provided via data conductors 41 during the data portion 62 . If there was a cache miss, data is not provided via data conductors 41 during the data portion 62 .
  • the response may only include status information, and cache entry information may not be provided.
  • the status information may be indicated in a particular response type encoding signaled via one or more signals 32 of bus 20 , or may be provided directly in an unencoded format.
  • the bus master e.g. debug circuitry 14
  • the bus master e.g. debug circuitry 14
  • the specifier is used to specify a particular storage location within the targeted cache.
  • the response from the cache is different in that the response 61 does not include cache entry information, and may or may not include information or content from the tag address field of the cache entry (e.g. 52 in FIG. 2 ).
  • information or content from the tag address field of the cache entry e.g. 52 in FIG.
  • the “show doubleword” bus transaction for request 60 may function in the same manner as the “show cache line” bus translation for request 60 , except the width is a doubleword instead of a cache line.
  • the “show doubleword” bus transaction for response 61 may function in the same manner as the “show copyback” bus transaction for response 61 , except the width is a doubleword instead of a cache line.
  • the transaction types for “show target line” and “show target doubleword” may be similarly differentiated.
  • one or more caches in system 10 may be set associative or fully associative. If a cache (e.g. 13 ) is fully associative, the “cache entry” in FIG. 6 may be used to directly specify a desired entry in the cache. However, if the cache (e.g. 13 ) is set associative, the “cache entry” specifier may be replaced with “cache set” and “cache way” information for some embodiments.
  • transfer type transaction type
  • bus transaction type snoop transaction
  • snoop transaction type snoop bus command
  • bus command snoop bus command
  • a cache coherency protocol may be extended to incorporate debug visibility transactions in an advantageous manner, allowing for improved cache visibility by a debugger.
  • FIG. 1 and the discussion thereof describe an exemplary information processing architecture
  • this exemplary architecture is presented merely to provide a useful reference in discussing various aspects of the invention.
  • the description of the architecture has been simplified for purposes of discussion, and it is just one of many different types of appropriate architectures that may be used in accordance with the invention.
  • Those skilled in the art will recognize that the boundaries between logic blocks are merely illustrative and that alternative embodiments may merge logic blocks or circuit elements or impose an alternate decomposition of functionality upon various logic blocks or circuit elements.
  • any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components.
  • any two components so associated can also be viewed as being “operably connected,” or “operably coupled,” to each other to achieve the desired functionality.
  • system 10 are circuitry located on a single integrated circuit or within a same device.
  • system 10 may include any number of separate integrated circuits or separate devices interconnected with each other.
  • memory 18 may be located on a same integrated circuit as masters 11 and 12 or on a separate integrated circuit or located within another peripheral or slave discretely separate from other elements of system 10 .
  • Debug circuitry 14 and I/O circuitry 16 may also be located on separate integrated circuits or devices.
  • system 10 or portions thereof may be soft or code representations of physical circuitry or of logical representations convertible into physical circuitry. As such, system 10 may be embodied in a hardware description language of any appropriate type.
  • Coupled is not intended to be limited to a direct coupling or a mechanical coupling.

Abstract

A plurality of new snoop transaction types are described. Some include address information in the requests, and others include cache entry information in the requests. Some responses include tag address information, and some do not. Some provide tag address content on the data bus lines during the data portion of the transaction. These new snoop transaction types are very helpful during debug of a data processing system.

Description

    CROSS-REFERENCE TO RELATED APPLICATION(S)
  • This application is related to U.S. patent application Ser. No. 11/140,176, filed on May 27, 2005, entitled “Translation Information Retrieval”, first named inventor being William C. Moyer, Attorney Docket No. SC13901TH, and assigned to the current assignee hereof.
  • This application is related to U.S. patent application Ser. No. 11/536,085, filed on Sep. 28, 2006, entitled “Data Processing System Having Cache Memory Debugging Support and Method Therefor”, first named inventor being William C. Moyer, Attorney Docket No. NC10117TH, and assigned to the current assignee hereof.
  • This application is related to U.S. Pat. No. 6,954,826, issued Oct. 11, 2005, entitled “READ ACCESS AND STORAGE CIRCUITRY READ ALLOCATION APPLICABLE TO A CACHE”, first named inventor being William C. Moyer, Attorney Docket No. SC12820TH, and assigned to the current assignee hereof.
  • This application is related to U.S. Pat. No. 7,185,148, issued Feb. 27, 2007, entitled “READ ACCESS AND STORAGE CIRCUITRY READ ALLOCATION APPLICABLE TO A CACHE”, first named inventor being William C. Moyer, Attorney Docket No. SC12820TH-D01, and assigned to the current assignee hereof.
  • This application is related to U.S. patent application Ser. No. ______, filed concurrently herewith, entitled “Method and Apparatus for Cache Transactions in a Data Processing System”, first named inventor being William C. Moyer, Attorney Docket No. NC45282TH, and assigned to the current assignee hereof.
  • BACKGROUND
  • 1. Field
  • This disclosure relates generally to a cache, and more specifically, to cache transactions in a data processing system.
  • 2. Related Art
  • In current data processing systems formed on an integrated circuit, it is often difficult to allow real-time debuggers to be able to view the current value of one or more variables in memory that may be hidden from access due to a cache or a cache hierarchy. Currently, in real-time debugging, if a variable is cached, the debugger has no access to the latest value of the variable without a very intrusive set of operations to halt the processor and extract the value from the cache. The problem is even more acute if a cache hierarchy exists between the processor and the debugger. Also, current methods may affect the state of the cache, which is usually detrimental in a debugging context. For example, if the state of the cache is affected, then certain debug issues may not be exposed in the same manner as a result of differences in cache state.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention is illustrated by way of example and is not limited by the accompanying figures, in which like references indicate similar elements. Elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale.
  • FIG. 1 illustrates, in block diagram form, a data processing system in accordance with one embodiment of the present invention.
  • FIG. 2 illustrates, in block diagram form, a portion of processor 11 and/or processor 12 of FIG. 1 in accordance with one embodiment of the present invention.
  • FIG. 3 illustrates, in tabular form, a listing of cache state definitions in accordance with one embodiment of the present invention.
  • FIG. 4 illustrates, in state diagram form, a portion of a cache coherency state diagram in accordance with one embodiment of the present invention.
  • FIG. 5 illustrates, in tabular form, a plurality of snoop bus commands in accordance with one embodiment of the present invention.
  • FIG. 6 illustrates, in tabular form, information provided on a bus during snoop bus commands in accordance with one embodiment of the present invention.
  • DETAILED DESCRIPTION
  • A plurality of new snoop bus commands or new snoop transaction types, and the supporting logic and framework for using them is described herein. In one embodiment, “show line” and “show doubleword” snoop transactions are provided to allow debug circuitry (e.g. 14 of FIG. 1) to gain access to the latest values of variables, even if they are cached, or to extract cache contents. In one embodiment, these snoop transactions act like existing snoop transaction types in some respects (e.g. in that snoop lookups are performed by participating bus masters). However, for the new snoop bus commands, the result of the lookup will cause a snoop response transaction to be requested on any hit, not just on a hit to a modified line. The response causes the contents of a cached doubleword or an entire cache line to be placed on a bus (e.g. 20 of FIG. 1), and thus available for capture by the debug circuitry (e.g. 14 of FIG. 1). In the illustrated embodiment, the new snoop transaction types include “show_line”, “show doubleword”, “show_target_line”, and “show_target_doubleword”. Alternate embodiments may have fewer, more, or different snoop transactions, may use different names for the snoop transactions, or may use modified versions of the snoop transactions.
  • As used herein, the term “bus” is used to refer to a plurality of signals or conductors which may be used to transfer one or more various types of information, such as data, addresses, control, or status. The conductors as discussed herein may be illustrated or described in reference to being a single conductor, a plurality of conductors, unidirectional conductors, or bidirectional conductors. However, different embodiments may vary the implementation of the conductors. For example, separate unidirectional conductors may be used rather than bidirectional conductors and vice versa. Also, plurality of conductors may be replaced with a single conductor that transfers multiple signals serially or in a time multiplexed manner. Likewise, single conductors carrying multiple signals may be separated out into various different conductors carrying subsets of these signals. Therefore, many options exist for transferring signals.
  • FIG. 1 illustrates, in block diagram form, a data processing system 10 in accordance with one embodiment of the present invention. In one embodiment, data processing system 10 includes a processor 12, a debug circuit or debug circuitry 14, an input/output (I/O) circuit 16, and a memory 18, all bi-directionally coupled to a bus 20.
  • The memory 18 may be any type of memory, such as, for example, a read only memory (ROM), a random access memory (RAM), non-volatile memory (e.g. Flash), etc. Also, memory 18 may be a memory or other data storage located within another peripheral or slave or on a different integrated circuit.
  • In one embodiment processor 11, processor 12 and debug circuitry 14 are bus masters. I/O circuitry 16 may also be a bus master in some embodiments. In one embodiment, processor 12 is a processor capable of executing instructions, such as a microprocessor, digital signal processor, etc., or may be any other type of bus master, such as for example, a direct memory access (DMA) controller, a bus bridge, or the like. Note that other slave devices may be included in data processing system 10 that are not illustrated in FIG. 1, such as for example, another memory or cache memory coupled to bus 20, as well as any type of peripheral circuit which resides on the system bus or is coupled to I/O circuit 16. In the illustrated embodiment, debug circuitry 14 is also bi-directionally coupled to processors 11 and 12 by way of conductors 21. In alternate embodiments, debug circuitry 14 may also be bi-directionally coupled to one or more of I/O circuitry 16 and memory 18 using one or more conductors (not shown) that are independent of bus 20.
  • In the illustrated embodiment, processor 11 includes a cache 15 which is bi-directionally coupled to bus 20 by way of bus interface unit (BIU) 17. Processor 12 includes a cache 13 and a cache 23 which are each bi-directionally coupled to bus 20 by way of bus interface unit (BIU) 19. In one embodiment, cache 13 functions as a level one (L1) cache for storing data and/or instructions for use by processor 12 and cache 23 functions as a level two (L2) cache for storing data and/or instructions for use by processor 12. Typically, data and/or instructions are loaded into caches 13, 23 from memory 18 before being executed by the processor. Caches 13, 23 may comprise a separate data cache and a separate instruction cache or may comprise a unified instruction and data cache, or may comprise only data caches. Caches 13, 23 comprise one or more data arrays, one or more tag arrays, and one or more status bit arrays. In some embodiments the tag and status information may be stored in a common array. Each cache entry consists of a block or line of data, a tag value which associates the data with a memory address, and status information. For some embodiments, the status information may include whether the cache entry is valid, whether the cache entry is “dirty” (i.e. modified with respect to the data associated with the same address in memory 18 or other external memory blocks if present), and whether the cache entry is exclusive or shared with other bus masters. Alternate embodiments may include less, more, or different cache status information. In an alternate embodiment, memory 18, or a portion of memory 18 may be characterized as a level two (L2) cache.
  • An input/output (I/O) module 16 is bi-directionally coupled to the bus 20 and to external circuitry (not shown). The I/O module 16 includes various interface circuits depending upon the type of external circuitry that is coupled to the I/O module 16. I/O module 16 may contain one or more alternate bus masters coupled to bus 20, and may contain one or more caches. In one embodiment, I/O module 16 may serve as a bus bridge, and be connected to one or more devices incorporating one or more caches via conductors 27.
  • Debug circuit 14 is bi-directionally coupled to one or more of bus processor 11, processor 12, I/O circuitry 16, and/or memory 18 via bus 20. Debug circuit 14 may also be bi-directionally coupled to external circuitry (not shown) such as a hardware tester or other debugging interface hardware. In one embodiment, debug circuitry 14 complies with the NEXUS debug protocol. For some embodiments, debug circuitry 14 also complies with the JTAG protocol. Alternate embodiments may use debug circuitry 14 that is complies with any desired debug protocol.
  • In general, debug circuitry 14 functions as a diagnostic check for data processing operations related to an access to caches 15, 13, 23, memory 18, and I/O module 16, and in other embodiments, other types of data retention circuits utilized by data processing system 10 whether internal to or external to data processing system 10. The processor 12 and alternate bus masters obtain mastership of the system bus 20 as needed and selectively access the system memory 18 to retrieve and store data and instructions. Debug circuitry 14 may be configured by the user of data processing system 10 to capture the values of one or more memory locations by monitoring addresses presented on bus 20, and capturing or sampling the related data values presented on bus 20. During bus read operations, the data values are provided by a selected slave device such as memory 18. During bus write operations, the data values are provided by a bus master of bus 20 such as processor 12. As the bus transfers occur, debug circuitry 14 monitors each transfer and selectively captures data values corresponding to data locations the user of debug circuitry 14 wishes to monitor. These values and conditions may be programmed into and transferred from debug circuitry 14 via one or more terminals 25 illustrated in FIG. 1. I/O circuitry 16 may also be coupled external to data processing system 10 by way of one or more terminals 27. In one embodiment, terminals 25 and 27 may be integrated circuit pins. In alternate embodiments, terminals 25 and 27 may be anything that can be used to transfer signals external to data processing system 10, such as, for example, conductive bumps, conductive pads, wires, etc.
  • Note that if a data variable or value is located within cache 13, the data value may be modified with respect to the corresponding value in memory 18, i.e. the latest (most up to date) value is not present in memory 18, but only within cache 13. If debug circuitry 14 accesses memory 18 to obtain the desired variable's value, it will not get a correct copy, since the copy may be modified internally in cache 13.
  • A new bus transaction type is provided in data processing system 10 to allow debug circuitry 14 to “see” a value which may be stored in a cache (e.g. cache 13). The bus transaction type is a “show” transaction. The bus transaction can be initiated by debug circuitry 14 acting as a bus master of bus 20. The bus transaction includes a desired address location to be accessed. It is determined (e.g. by cache control circuitry 54 of FIG. 2) if the address or cache entry results in a “hit” or a “miss” in the cache. When the address results in a hit, the value is provided to bus 20 associated with the address from the cache 13. Debug circuitry 14 then has access to the value from the bus 20 as it is provided by the processor (e.g. 12). In one embodiment, debug circuitry 14 monitors transactions on bus 20, and is capable of sampling the address and associated data values as bus transfers occur. Desired values can then be transferred after sampling to a user of the system 10 performing a debugging operation by means of signals from debug circuitry 14 of FIG. 1 (e.g. via terminals 25). Note that in one embodiment, a state of cache 13 is not modified when the value is placed on the bus 20. In most prior art systems, it is not possible to place the value of the data variable on bus 20 from cache 13 when a normal cache hit occurs, since the cache 13 satisfies the prior art load or store request generated by a normal load or store instruction. This causes an issue for debug circuitry 14, since the value is not made visible on bus 20 for capture. When the access address results in a miss, there is no information to be provided from the cache (e.g. 13) to debug circuitry 14. On a miss, data may be retrieved via bus 20, and thus be visible to debug circuitry 14 for capture. In these prior art systems, providing visibility only on cache misses results in limitations on the debugging operations that can be performed.
  • Typically, some data values are subject to change and are updated in response to executing instructions. In some debugging situations it may be desirable to retrieve the latest version of the value even when resident in the cache. A “dirty bit” “D” is associated with a cache entry indicates whether or not data in the cache associated with the effective address is different (i.e. has been modified) from data stored at another memory location, such as memory 18, corresponding to the data address. In accordance with the disclosed embodiment, a “dirty bit” associated with the address of the value is not modified when the value is retrieved from the cache and presented to the system bus 20 in response to executing a “show” bus transaction. Likewise, cache 13 is not modified if the value associated with the effective address in cache 13 is the same as the value in another memory location, i.e. the data is clean. In other words, even when a dirty bit of the hit cache entry indicates that data stored in the entry is clean, the data stored in the entry is provided to bus 20 and the entry is not modified. When a dirty bit of the hit cache entry indicates that data stored in the entry is dirty, the data stored in the entry is still provided to bus 20, and the entry is not modified, and the dirty bit is not cleared. This is different than the normal operation of a “dirty bit” in a cache.
  • FIG. 2 illustrates, in block diagram form, a portion of processor 11 and/or processor 12 of FIG. 1. FIG. 2 illustrates one possible embodiment of cache 13 and BIU 19, and/or cache 23 and BIU 19, and/or cache 15 and BIU 17, all of FIG. 1, in more detail. For ease of discussion, it will be assumed herein below that cache 13 uses the embodiment illustrated in FIG. 2. However, in alternate embodiments, cache 15 and/or cache 23 may also use the cache embodiment illustrated in FIG. 2, and/or cache 13 may use a cache embodiment different than that illustrated in FIG. 2. Thus, alternate embodiments of system 10 may use different architectures, structures, and/or circuitry to implement the various caches (e.g. 13, 23, and 15). In the illustrated embodiment, cache 13 includes cache control circuitry 54 and cache array 50 which are bi-directionally coupled to each other. In one embodiment, cache array 50 is an array of random access memory cells organized as a plurality of entries. In one embodiment, cache array 50 comprises SRAM (static random access memory) memory cells. In other embodiments, other memory types are suitable. In one embodiment of cache 13, entry 52 is representative of the plurality of entries and includes a tag address bit field labeled “TAG ADDRESS FIELD”, a tag status bit field labeled “TAG STATUS BITS” and a data bit field labeled “DATA”. The tag status bit field includes a valid bit labeled “V”, an exclusive/shared bit labeled “E”, and a dirty bit labeled “D”.
  • In one embodiment of cache 13, each entry may be referred to as a “cache line” for processor 12. Cache array 50 is bi-directionally coupled to bus 20 via BIU 19. In one embodiment, cache control circuitry 54 selectively determines if an effective address pointed to by an instruction received from, for example, cache 13, memory 18, or debug circuitry 14 is a hit or a miss in cache 13. Alternate embodiments may use a translation look-up buffer (TLB) (not shown) coupled to cache 13 for performing address translation from an effective address to a physical address. This translation may be done in a standard manner that is known in the art. Cache control circuitry 54 can determine if an effective address pointed to by an instruction results in a hit or miss by comparing the physical address for an access provided by the TLB to the stored value(s) contained in one or more tag entries in cache array 50. In some embodiments, a TLB is not used and the effective address is provided directly to cache 13 without translation. In such embodiments, stored tag values correspond directly to effective addresses without address translation to a physical address being required.
  • Still referring to FIG. 2, cache 13 is bi-directionally coupled to BIU 19. BIU 19 is bi-directionally coupled to bus 20. In the illustrated embodiment, bus 20 includes signals 30 and signals 32. In one embodiments, signals 30 include a plurality of address signals 40, a plurality of data signals 41, one or more transfer_type signals 42, one or more target_specifier signals 43, and one or more cache_specifier signals 44.
  • FIG. 3 illustrates, in tabular form, a listing of cache state definitions in accordance with one embodiment of the present invention. Alternate embodiments of the present invention may use fewer, more, or different cache states than those illustrated in FIG. 3. FIG. 3 illustrates a MESI (Modified Exclusive Shared and Invalid) cache protocol having four states, namely an invalid state (INV), a shared state (S), an exclusive unmodified state (EU), and an exclusive modified state (EM). Other embodiments may use different cache protocols and their associated cache state definitions. Thus, the present invention is not limited in any way to the specific states illustrated in FIG. 3.
  • FIG. 4 illustrates, in state diagram form, a portion of a cache coherency state diagram in accordance with one embodiment of the present invention. The state diagram illustrated in FIG. 4 shows how a cache implementing the cache states of FIG. 3 can transition between the four cache states INV, S, EU, and EM. New state transitions have been incorporated into the prior art state diagram for a MESI cache coherency protocol to support the new snoop transaction types (show_line, show_dw, show13 target_line, show_target_dw) for one embodiment of the present invention. For clarity purposes, note that some of the prior art bus transactions that are not relevant to the present invention have been omitted from FIG. 4 (e.g. bus commands that affect a whole block of cache, such as, for example, a bus command that flushes an entire cache block).
  • Still referring to FIG. 4, note that the “show line” (show_line hit) and “show doubleword” (show_dw hit) bus commands do not cause the cache 13 to change state. Similarly, note that the “show target line” (show—l target_line) and “show target doubleword” (show_target_dw) bus commands do not cause the cache 13 to change state. These two new pairs of snoop transaction types or transfer type bus commands, and the supporting logic and framework for using them, allow debug circuitry 14 of FIG. 1 more visibility into the caches (e.g. 15, 13, and 23) used in system 10. The “show line” and “show doubleword” snoop transactions or bus commands are provided to allow debug circuitry 14 to gain access to the latest values of variables, even if they are cached, or to allow debug circuitry 14 to extract cache contents. In one embodiment, these snoop transactions may act like existing snoop transaction types in most respects, in that snoop lookups are performed by participating bus masters (e.g. 11, 12), but the result of the lookup will optionally cause a snoop response transaction to be requested on any hit, not just on a hit to a modified or dirty cache line. For the “show_target_line” and “show_target_dw” snoop commands, a specific cache target is identified as part of the transaction information, and for these specific command types, the result of the lookup will optionally cause a snoop response transaction to occur regardless of a hit or miss. Utilizing these snoop command types allows for visibility into the current state of a targeted cache without causing state changes of data, tag, or status information.
  • In response to the request portion of the “show line”, “show doubleword”, “show target line”, and “show target doubleword” bus transactions, the cache control circuitry 54 (see FIG. 2) causes the contents of a cached doubleword or an entire line to be placed on bus 20, and thus available for capture by the debug circuitry 14. This exposure occurs regardless of the presence of backing store for the variables, i.e. there may be no physical memory associated with the address of the variable other than in a cache itself. By exposing the value of a variable in memory to the system bus 20, it can be captured by data trace logic in debug circuitry 14 and messaged out to the user via terminals 25. Note that the instructions are unobtrusive to the data cache, and no data cache state changes occur, regardless of the state of the variable (Modified, Exclusive, Shared, or Invalid). In the illustrated embodiment, options are provided to cause either an entire cache line to be placed on bus 20, or to limit the exposure to a data bus-width element (e.g. a doubleword) containing the desired variable. Alternate embodiments may instead use one snoop transaction type with a fixed width or size (e.g. line), or may use one bus command that has within it an encoding to select among a variable group of widths or sizes (e.g. line, doubleword, word, etc.)
  • Note that in one embodiment, snoop transaction types are provided that allow the value of a memory variable to be provided to debug circuitry 14 (see FIG. 1). In some embodiments, cache control circuitry 54 allows a cache line or a portion of a cache line to be broadcast on a bus 20 when it is present in the cache (e.g. 13), regardless of whether it is clean (i.e. unmodified) or dirty (i.e. modified). In one embodiment there is provided cache state logic (see FIG. 4) in cache control circuitry 54 (see FIG. 2) which does not modify the cache state (see FIG. 3) or replace cache lines to obtain the variable. In addition, for some embodiments, memory (e.g. 18) is not updated with the provision of the cache data to debug circuitry 14 in order to ensure that queries by debug circuitry 14 are minimally intrusive to the state of system 10.
  • In one embodiment, system 10 adds an additional bus transaction type “show” which allows debug circuitry 14 visibility into internal data cache state (e.g. of cache 13), which may be modified with respect to memory (e.g. 18). In one embodiment, when “show” is presented as a snoop read burst transaction, a normal snoop lookup is performed. If the transaction address hits in the data cache (e.g. cache 13), then a snoop copyback is performed, regardless of the modified or clean state of the cache line. The copyback is marked as a “show line” write, and in one embodiment is a burst of four doublewords to the bus. The state of the cache line in all processors (e.g. 11, 12) remains unchanged. The providing of the cache line to the bus 20 allows the debug circuitry 14 to see the content of the cache line, and thus to transfer these values out to an external debugger (not shown) via terminals 25. Note that in one embodiment, updates to memory (e.g. 18) should typically be blocked, if possible, during these “show” transfers to allow the state of the system 10 to remain minimally perturbed during debugging.
  • Additionally, in some embodiments, “doubleword” snoop bus commands are supported. When presented as a snoop single-beat read transaction, a normal snoop lookup is also performed. These “doubleword” bus commands operate similarly to the “line” bus commands, except that only a single doubleword of data in a hitting cache line is provided, via a single-beat write bus transaction which is marked as “doubleword”.
  • The “show_target_line” and “show—target_doubleword” snoop bus commands are typically targeted to only a single master (e.g. processor 12) in the system 10, and thus only a single response will occur for each of these transactions. For the “show_line” and “show_doubleword” snoop transaction requests, these commands are simultaneously broadcast to multiple masters (e.g. processors 11 and 12), but only a single master needs to respond for the data to become visible to the debug circuitry 14. For cache lines in the exclusive state, this occurs naturally. For shared cache lines which are valid in multiple caches, only a single “show” bus transaction is required to provide visibility of the data, even though multiple cached copies exist. When the plurality of bus masters attempt to simultaneously “show” the cache line, only a single master will win arbitration for bus 20 and then respond with the requested information. A pending response in any other bus master can be killed or terminated when the arbitration winner performs the “show” response on bus 20, and the addresses of the pending response and the performed response match. The participating masters will monitor the bus to determine if another cache has provided the “show” information, and if so, will terminate their own requests to perform a “show” response for the indicated data. This may be done by monitoring the address and the transaction type information for transactions on bus 20, and determining that a “show” type response transaction is performed by another master. Alternatively, if multiple caches respond to a “show” command, all responses could be sent individually to the external debugger along with information on which cache responded, allowing the external debugger to see all copies of the line.
  • Referring to FIG. 4, in one embodiment, transitions 50-53 may be used to provide information from any cache (e.g. cache 13) to debug circuitry 14 (see FIG. 1). Referring now to FIGS. 2, 5 and 6, in one embodiment, the transfer-type signals 42 on bus 20 (see FIG. 2) are used to indicate when one of the bus snoop transactions listed in FIG. 5 are currently taking place on bus 20. Referring to FIGS. 5 and 6, note that each snoop bus transaction includes a request portion 60, a response portion 61, and a data portion 62 that all are provided on bus 20. In the request portion 60, a bus master (e.g. debug circuitry 14) initiates a read to a cache (e.g. cache 13). In the response portion 61, the cache responds with a write operation. And in the data portion 62, the data stored in the cache that corresponds to the address or cache entry specified in the request 60 portion is provided on bus 20.
  • One embodiment of the bus snoop transactions listed in FIG. 5 will now be described. Note that alternate embodiments may have other additional snoop transactions that are not listed in FIG. 5.
  • One embodiment of the “show cache line” and “show doubleword” bus transactions of FIG. 5 will now be described. For the “show cache line” bus transaction, a bus master (e.g. debug circuitry 14) provides a snoop transaction request 60 which includes a desired address by way of signals 40 of bus 20 (see FIG. 2), and provides a transaction type indication by way of one or more signals 42. The transaction is monitored by one or more caches in the system which participate in normal snooping of bus transactions for coherency purposes. In response to a hit occurring on the lookup associated with the snoop transaction, a selected cache in the system performs a “show copyback” (for responding to a “show cache line” request) or a “show doubleword” (for responding to a “show doubleword” request) (see FIG. 5) write transaction of the cache line data information corresponding to the address of the request transaction regardless of the state of the dirty bit D, and the cache remains in its current state, i.e. the dirty bit is unchanged. In addition, all other caches remain in their respective current states. In one embodiment, memory 18 is not updated with the data provided on the response, even though a write transaction is indicated on the bus 20. In one embodiment, the response 61 includes status information from the cache and the cache entry that corresponds to the address in the request 60. If there was a cache hit, data is provided via data conductors 41 during the data portion 62. If there was a cache miss, data is not provided via data conductors 41 during the data portion 62. In an alternate embodiment, no status, or only partial status information may be provided during the response. In one embodiment, if the requested address misses in all of the caches participating in the transaction, data may be provided by memory 18 instead of no data being provided during data portion 62 of the transaction. In one embodiment, if multiple caches “hit”, a single cache is selected to provide the response, and the remaining caches cancel any pending responses.
  • One embodiment of the “show target line” and “show target doubleword” bus transactions of FIG. 5 corresponding to the type-1 transactions illustrated in FIG. 6 will now be described. For these “show target” bus transactions, a bus master (e.g. debug circuitry 14) provides a request 60 which includes a target specifier by way of signals 43 of bus 20 (see FIG. 2), provides a cache specifier by way of signals 44, provides an address by way of signals 40, and provides a width indicator DW (e.g. doubleword, line, etc.) by way of one or more signals 32. The target specifier specifies which block or portion of circuitry in system 10 has a cache and is the “target” of this bus transaction. The cache specifier specifies which cache within the “target” is being accessed for this bus transaction. In response (see column 61 in FIG. 6), the specified cache in the target performs a “show target copyback” (see FIG. 5) of the cache line or a “show target doubleword” of the requested doubleword, regardless of the state of the dirty bit D, and the cache remains in its current state. Referring to FIG. 6, in the illustrated embodiment, the response 61 includes status information from the cache and the cache entry that corresponds to the address in the request 60. If there was a cache hit, data is provided via data conductors 41 during the data portion 62. If there was a cache miss, data is not provided via data conductors 41 during the data portion 62. In one embodiment, the response may only include status information, and cache entry information may not be provided. The status information may be indicated in a particular response type encoding signaled via one or more signals 32 of bus 20, or may be provided directly in an unencoded format.
  • One embodiment of the “show target line” and “show target doubleword” bus transactions of FIG. 5 corresponding to the type-2 transactions illustrated in FIG. 6 will now be described. For the type-2 “show cache line” bus transaction, what is different from type-1 is that instead of providing an address in the request portion 60, the bus master (e.g. debug circuitry 14) provides a cache entry specifier. The specifier is used to specify a particular storage location within the targeted cache. The response from the cache is different in that the response 61 does not include cache entry information, and may or may not include information or content from the tag address field of the cache entry (e.g. 52 in FIG. 2). For one embodiment of type-2 transactions, information or content from the tag address field of the cache entry (e.g. 52 in FIG. 2) may be provided on data conductors 41 of bus 20 (see FIG. 2) during the data portion 62 if it was not provided during the response portion 61. Alternate embodiments may use one or more of the type-1 transactions, may use one or more of the type-2 transactions, or may use different transactions. Note that the “show doubleword” bus transaction for request 60 (see FIG. 5) may function in the same manner as the “show cache line” bus translation for request 60, except the width is a doubleword instead of a cache line. Similarly, note that the “show doubleword” bus transaction for response 61 may function in the same manner as the “show copyback” bus transaction for response 61, except the width is a doubleword instead of a cache line. The transaction types for “show target line” and “show target doubleword” may be similarly differentiated.
  • Note that one or more caches in system 10 may be set associative or fully associative. If a cache (e.g. 13) is fully associative, the “cache entry” in FIG. 6 may be used to directly specify a desired entry in the cache. However, if the cache (e.g. 13) is set associative, the “cache entry” specifier may be replaced with “cache set” and “cache way” information for some embodiments.
  • Note that the terms transfer type, transaction type, bus transaction type, snoop transaction, snoop transaction type, snoop bus command, and bus command have been used interchangeably herein.
  • As can be seen from the description of the present invention, a cache coherency protocol may be extended to incorporate debug visibility transactions in an advantageous manner, allowing for improved cache visibility by a debugger.
  • Because the apparatus implementing the present invention is, for the most part, composed of electronic components and circuits known to those skilled in the art, circuit details will not be explained in any greater extent than that considered necessary as illustrated above, for the understanding and appreciation of the underlying concepts of the present invention and in order not to obfuscate or distract from the teachings of the present invention.
  • Some of the above embodiments, as applicable, may be implemented using a variety of different information processing systems. For example, although FIG. 1 and the discussion thereof describe an exemplary information processing architecture, this exemplary architecture is presented merely to provide a useful reference in discussing various aspects of the invention. Of course, the description of the architecture has been simplified for purposes of discussion, and it is just one of many different types of appropriate architectures that may be used in accordance with the invention. Those skilled in the art will recognize that the boundaries between logic blocks are merely illustrative and that alternative embodiments may merge logic blocks or circuit elements or impose an alternate decomposition of functionality upon various logic blocks or circuit elements.
  • Thus, it is to be understood that the architectures depicted herein are merely exemplary, and that in fact many other architectures can be implemented which achieve the same functionality. In an abstract, but still definite sense, any arrangement of components to achieve the same functionality is effectively “associated” such that the desired functionality is achieved. Hence, any two components herein combined to achieve a particular functionality can be seen as “associated with” each other such that the desired functionality is achieved, irrespective of architectures or intermedial components. Likewise, any two components so associated can also be viewed as being “operably connected,” or “operably coupled,” to each other to achieve the desired functionality.
  • Also for example, in one embodiment, the illustrated elements of system 10 are circuitry located on a single integrated circuit or within a same device. Alternatively, system 10 may include any number of separate integrated circuits or separate devices interconnected with each other. For example, memory 18 may be located on a same integrated circuit as masters 11 and 12 or on a separate integrated circuit or located within another peripheral or slave discretely separate from other elements of system 10. Debug circuitry 14 and I/O circuitry 16 may also be located on separate integrated circuits or devices. Also for example, system 10 or portions thereof may be soft or code representations of physical circuitry or of logical representations convertible into physical circuitry. As such, system 10 may be embodied in a hardware description language of any appropriate type.
  • Furthermore, those skilled in the art will recognize that boundaries between the functionality of the above described operations merely illustrative. The functionality of multiple operations may be combined into a single operation, and/or the functionality of a single operation may be distributed in additional operations. Moreover, alternative embodiments may include multiple instances of a particular operation, and the order of operations may be altered in various other embodiments.
  • Although the invention is described herein with reference to specific embodiments, various modifications and changes can be made without departing from the scope of the present invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of the present invention. Any benefits, advantages, or solutions to problems that are described herein with regard to specific embodiments are not intended to be construed as a critical, required, or essential feature or element of any or all the claims.
  • The term “coupled,” as used herein, is not intended to be limited to a direct coupling or a mechanical coupling.
  • Furthermore, the terms “a” or “an,” as used herein, are defined as one or more than one. Also, the use of introductory phrases such as “at least one” and “one or more” in the claims should not be construed to imply that the introduction of another claim element by the indefinite articles “a” or “an” limits any particular claim containing such introduced claim element to inventions containing only one such element, even when the same claim includes the introductory phrases “one or more” or “at least one” and indefinite articles such as “a” or “an.” The same holds true for the use of definite articles.
  • Unless stated otherwise, terms such as “first” and “second” are used to arbitrarily distinguish between the elements such terms describe. Thus, these terms are not necessarily intended to indicate temporal or other prioritization of such elements.
  • Additional Text
    • 1. A method for a cache, the method comprising:
      • in response to receiving a request to provide at least one portion of a cache entry of the cache, providing the at least one portion of the cache entry regardless of whether the cache entry is dirty without modifying a state of the cache entry.
    • 2. The method of statement 1, wherein the at least one portion of the cache entry comprises a byte, a word, or a doubleword.
    • 3. The method of statement 1, wherein the state of the cache entry comprises at least one of an invalid state, a shared state, an exclusive unmodified state, and an exclusive modified state.
    • 4. The method of statement 1, wherein the request to provide the at least one portion has a corresponding address.
    • 5. The method of statement 1, wherein the request to provide the at least one portion has a corresponding target specifier and a corresponding cache specifier.
    • 6. The method of statement 5, wherein the target specifier is used to select a target device from a plurality of target devices coupled to a device issuing the request and wherein the cache specifier is used to select a cache from within the selected target device.
    • 7. The method of statement 1 further comprising receiving the request to provide at least one portion of the cache entry from a debug circuitry.
    • 8. The method of statement 1 further comprising receiving the request to provide at least one portion of the cache entry from at least one of a bus bridge, a direct memory access controller, and a processor.
    • 9. The method of statement 1 further comprising providing the at least one portion of the cache entry without modifying a memory associated with the cache.
    • 10. The method of statement 1, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches and wherein the step of providing the at least one portion of the cache entry is performed by only one of the plurality of caches.
    • 11. The method of statement 1, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches, wherein the state of the cache entry is shared, and wherein the step of providing the at least one cache entry is performed by only one of the plurality of caches.
    • 12. A method for a cache, the method comprising:
      • in response to receiving a request to provide at least one portion of a cache entry of the cache, providing the at least one portion of the cache entry regardless of a state of the cache entry without modifying the state of the cache entry.
    • 13. The method of statement 12, wherein the at least one portion of the cache entry comprises a byte, a word, or a doubleword.
    • 14. The method of statement 12, wherein the state of the cache entry comprises at least one of an invalid state, a shared state, an exclusive unmodified state, and an exclusive modified state.
    • 15. The method of statement 12, wherein the request to provide the at least one portion has a corresponding address.
    • 16. The method of statement 12, wherein the request to provide the at least one portion has a corresponding target specifier and a corresponding cache specifier.
    • 17. The method of statement 16, wherein the target specifier is used to select a target device from a plurality of target devices coupled to a device issuing the request and wherein the cache specifier is used to select a cache from within the selected target device.
    • 18. The method of statement 12 further comprising receiving the request to provide at least one portion of the cache entry from a debug circuitry.
    • 19. The method of statement 12 further comprising receiving the request to provide at least one portion of the cache entry from at least one of a bus bridge, a direct memory access controller, and a processor.
    • 20. The method of statement 12 further comprising providing the at least one portion of the cache entry without modifying a memory associated with the cache.
    • 21. The method of statement 12, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches and wherein the step of providing the at least one portion of the cache entry is performed by only one of the plurality of caches.
    • 22. The method of statement 12, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches, wherein the state of the cache entry is shared, and wherein the step of providing the at least one cache entry is performed by only one of the plurality of caches.

Claims (22)

1. A method for a cache, the method comprising:
in response to receiving a request to provide at least one portion of a cache entry of the cache, providing the at least one portion of the cache entry regardless of whether the cache entry is dirty without modifying a state of the cache entry.
2. The method of claim 1, wherein the at least one portion of the cache entry comprises a byte, a word, or a doubleword.
3. The method of claim 1, wherein the state of the cache entry comprises at least one of an invalid state, a shared state, an exclusive unmodified state, and an exclusive modified state.
4. The method of claim 1, wherein the request to provide the at least one portion has a corresponding address.
5. The method of claim 1, wherein the request to provide the at least one portion has a corresponding target specifier and a corresponding cache specifier.
6. The method of claim 5, wherein the target specifier is used to select a target device from a plurality of target devices coupled to a device issuing the request and wherein the cache specifier is used to select a cache from within the selected target device.
7. The method of claim 1 further comprising receiving the request to provide at least one portion of the cache entry from a debug circuitry.
8. The method of claim 1 further comprising receiving the request to provide at least one portion of the cache entry from at least one of a bus bridge, a direct memory access controller, and a processor.
9. The method of claim 1 further comprising providing the at least one portion of the cache entry without modifying a memory associated with the cache.
10. The method of claim 1, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches and wherein the step of providing the at least one portion of the cache entry is performed by only one of the plurality of caches.
11. The method of claim 1, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches, wherein the state of the cache entry is shared, and wherein the step of providing the at least one cache entry is performed by only one of the plurality of caches.
12. A method for a cache, the method comprising:
in response to receiving a request to provide at least one portion of a cache entry of the cache, providing the at least one portion of the cache entry regardless of a state of the cache entry without modifying the state of the cache entry.
13. The method of claim 12, wherein the at least one portion of the cache entry comprises a byte, a word, or a doubleword.
14. The method of claim 12, wherein the state of the cache entry comprises at least one of an invalid state, a shared state, an exclusive unmodified state, and an exclusive modified state.
15. The method of claim 12, wherein the request to provide the at least one portion has a corresponding address.
16. The method of claim 12, wherein the request to provide the at least one portion has a corresponding target specifier and a corresponding cache specifier.
17. The method of claim 16, wherein the target specifier is used to select a target device from a plurality of target devices coupled to a device issuing the request and wherein the cache specifier is used to select a cache from within the selected target device.
18. The method of claim 12 further comprising receiving the request to provide at least one portion of the cache entry from a debug circuitry.
19. The method of claim 12 further comprising receiving the request to provide at least one portion of the cache entry from at least one of a bus bridge, a direct memory access controller, and a processor.
20. The method of claim 12 further comprising providing the at least one portion of the cache entry without modifying a memory associated with the cache.
21. The method of claim 12, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches and wherein the step of providing the at least one portion of the cache entry is performed by only one of the plurality of caches.
22. The method of claim 12, wherein the request to provide at least one portion of the cache entry is received by a plurality of caches, wherein the state of the cache entry is shared, and wherein the step of providing the at least one cache entry is performed by only one of the plurality of caches.
US11/748,353 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system Abandoned US20080288725A1 (en)

Priority Applications (6)

Application Number Priority Date Filing Date Title
US11/748,353 US20080288725A1 (en) 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system
US11/748,350 US8972671B2 (en) 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system
JP2010508485A JP2010527488A (en) 2007-05-14 2008-04-29 Method and apparatus for cache transactions in a data processing system
KR1020097023703A KR20100016516A (en) 2007-05-14 2008-04-29 Method and apparatus for cache transactions in a data processing system
CN200880015903XA CN101918923A (en) 2007-05-14 2008-04-29 Method and apparatus for cache transactions in a data processing system
PCT/US2008/061887 WO2008144179A1 (en) 2007-05-14 2008-04-29 Method and apparatus for cache transactions in a data processing system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/748,353 US20080288725A1 (en) 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system

Publications (1)

Publication Number Publication Date
US20080288725A1 true US20080288725A1 (en) 2008-11-20

Family

ID=40028700

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/748,353 Abandoned US20080288725A1 (en) 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system
US11/748,350 Active 2031-02-13 US8972671B2 (en) 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system

Family Applications After (1)

Application Number Title Priority Date Filing Date
US11/748,350 Active 2031-02-13 US8972671B2 (en) 2007-05-14 2007-05-14 Method and apparatus for cache transactions in a data processing system

Country Status (5)

Country Link
US (2) US20080288725A1 (en)
JP (1) JP2010527488A (en)
KR (1) KR20100016516A (en)
CN (1) CN101918923A (en)
WO (1) WO2008144179A1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090024892A1 (en) * 2007-07-18 2009-01-22 Vinod Bussa System and Method of Testing using Test Pattern Re-Execution in Varying Timing Scenarios for Processor Design Verification and Validation
US20090024876A1 (en) * 2007-07-18 2009-01-22 Sampan Arora System and Method for Verification of Cache Snoop Logic and Coherency Between Instruction & Data Caches for Processor Design Verification and Validation
US20090024886A1 (en) * 2007-07-18 2009-01-22 Sampan Arora System and Method for Predicting lwarx and stwcx Instructions in Test Pattern Generation and Simulation for Processor Design Verification and Validation
US20090024894A1 (en) * 2007-07-18 2009-01-22 International Business Machines Corporation System and method for predicting iwarx and stwcx instructions in test pattern generation and simulation for processor design verification/validation in interrupt mode
US20090024873A1 (en) * 2007-07-18 2009-01-22 Sandip Bag System and Method for Increasing Error Checking Performance by Calculating CRC Calculations After Multiple Test Patterns for Processor Design Verification and Validation
US20090024877A1 (en) * 2007-07-18 2009-01-22 Shubhodeep Roy Choudhury System and Method for Creating Different Start Cache and Bus States Using Multiple Test Patterns for Processor Design Verification and Validation
US20100023697A1 (en) * 2008-07-25 2010-01-28 International Business Machines Corporation Testing Real Page Number Bits in a Cache Directory
US20100205377A1 (en) * 2009-02-06 2010-08-12 Moyer William C Debug control for snoop operations in a multiprocessor system and method thereof
US20100205378A1 (en) * 2009-02-06 2010-08-12 Moyer William C Method for debugger initiated coherency transactions using a shared coherency manager
US8935485B2 (en) 2011-08-08 2015-01-13 Arm Limited Snoop filter and non-inclusive shared cache memory
EP2838024A3 (en) * 2013-07-25 2015-07-01 Altera Corporation Cache debug system for programmable circuits
US9477600B2 (en) 2011-08-08 2016-10-25 Arm Limited Apparatus and method for shared cache control including cache lines selectively operable in inclusive or non-inclusive mode
US10108568B2 (en) 2015-03-30 2018-10-23 Samsung Electronics Co., Ltd. Master capable of communicating with slave and system including the master
US10261827B2 (en) 2015-10-29 2019-04-16 International Business Machines Corporation Interprocessor memory status communication
US10346305B2 (en) * 2015-10-29 2019-07-09 International Business Machines Corporation Interprocessor memory status communication

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5118562B2 (en) * 2008-06-20 2013-01-16 株式会社東芝 Debugging support device

Citations (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4843542A (en) * 1986-11-12 1989-06-27 Xerox Corporation Virtual memory cache for use in multi-processing systems
US5420999A (en) * 1992-03-27 1995-05-30 International Business Machines Corporation Free memory cell management
US5471598A (en) * 1993-10-18 1995-11-28 Cyrix Corporation Data dependency detection and handling in a microprocessor with write buffer
US5561779A (en) * 1994-05-04 1996-10-01 Compaq Computer Corporation Processor board having a second level writeback cache system and a third level writethrough cache system which stores exclusive state information for use in a multiprocessor computer system
US5561786A (en) * 1992-07-24 1996-10-01 Microsoft Corporation Computer method and system for allocating and freeing memory utilizing segmenting and free block lists
US5685011A (en) * 1995-05-15 1997-11-04 Nvidia Corporation Apparatus for handling failures to provide a safe address translation in an improved input/output architecture for a computer system
US5689707A (en) * 1995-12-04 1997-11-18 Ncr Corporation Method and apparatus for detecting memory leaks using expiration events and dependent pointers to indicate when a memory allocation should be de-allocated
US5689679A (en) * 1993-04-28 1997-11-18 Digital Equipment Corporation Memory system and method for selective multi-level caching using a cache level code
US5765199A (en) * 1994-01-31 1998-06-09 Motorola, Inc. Data processor with alocate bit and method of operation
US5819304A (en) * 1996-01-29 1998-10-06 Iowa State University Research Foundation, Inc. Random access memory assembly
US5829027A (en) * 1994-05-04 1998-10-27 Compaq Computer Corporation Removable processor board having first, second and third level cache system for use in a multiprocessor computer system
US5913923A (en) * 1996-12-06 1999-06-22 National Semiconductor Corporation Multiple bus master computer system employing a shared address translation unit
US5990914A (en) * 1997-09-09 1999-11-23 Compaq Computer Corporation Generating an error signal when accessing an invalid memory page
US6145122A (en) * 1998-04-27 2000-11-07 Motorola, Inc. Development interface for a data processor
US20010032305A1 (en) * 2000-02-24 2001-10-18 Barry Edwin F. Methods and apparatus for dual-use coprocessing/debug interface
US6353829B1 (en) * 1998-12-23 2002-03-05 Cray Inc. Method and system for memory allocation in a multiprocessing environment
US20020174316A1 (en) * 2001-05-18 2002-11-21 Telgen Corporation Dynamic resource management and allocation in a distributed processing device
US6496902B1 (en) * 1998-12-31 2002-12-17 Cray Inc. Vector and scalar data cache for a vector multiprocessor
US6560688B1 (en) * 1998-10-01 2003-05-06 Advanced Micro Devices, Inc. System and method for improving accelerated graphics port systems
US6684296B2 (en) * 2001-05-18 2004-01-27 Broadcom Corporation Source controlled cache allocation
US6741254B1 (en) * 1997-12-30 2004-05-25 Micron Technology, Inc. Method of implementing an accelerated graphics port for a multiple memory controller computer system
US20040193831A1 (en) * 2003-03-25 2004-09-30 Moyer William C. Memory management in a data processing system
US20040205420A1 (en) * 2002-06-26 2004-10-14 Charles Seeley Error reporting network in multiprocessor computer
US6934937B1 (en) * 2000-03-30 2005-08-23 Broadcom Corporation Multi-channel, multi-service debug on a pipelined CPU architecture
US6954826B2 (en) * 2003-05-21 2005-10-11 Freescale Semiconductor, Inc. Read access and storage circuitry read allocation applicable to a cache
US20060085489A1 (en) * 2004-10-15 2006-04-20 Microsoft Corporation Memory cache management in XML/relational data mapping
US20060212426A1 (en) * 2004-12-21 2006-09-21 Udaya Shakara Efficient CAM-based techniques to perform string searches in packet payloads
US7117300B1 (en) * 2001-12-27 2006-10-03 James David V Method and apparatus for restricted search operation in content addressable memory (CAM) devices
US20080010409A1 (en) * 2006-07-05 2008-01-10 Cisco Technology, Inc. Dynamic, on-demand storage area network (SAN) cache
US7506105B2 (en) * 2005-05-02 2009-03-17 Freescale Semiconductor, Inc. Prefetching using hashed program counter

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5636363A (en) 1991-06-14 1997-06-03 Integrated Device Technology, Inc. Hardware control structure and method for off-chip monitoring entries of an on-chip cache
US6044478A (en) 1997-05-30 2000-03-28 National Semiconductor Corporation Cache with finely granular locked-down regions
US6260131B1 (en) 1997-11-18 2001-07-10 Intrinsity, Inc. Method and apparatus for TLB memory ordering
US6016555A (en) 1997-11-19 2000-01-18 Texas Instruments Incorporated Non-intrusive software breakpoints in a processor instruction execution pipeline
US6532553B1 (en) 1998-12-08 2003-03-11 Arm Limited Debugging data processing systems
US6321329B1 (en) 1999-05-19 2001-11-20 Arm Limited Executing debug instructions
US6681346B2 (en) 2000-05-11 2004-01-20 Goodrich Corporation Digital processing system including a DMA controller operating in the virtual address domain and a method for operating the same
EP1213650A3 (en) 2000-08-21 2006-08-30 Texas Instruments France Priority arbitration based on current task and MMU
US6968446B1 (en) 2001-08-09 2005-11-22 Advanced Micro Devices, Inc. Flags handling for system call instructions
US7346744B1 (en) 2002-11-04 2008-03-18 Newisys, Inc. Methods and apparatus for maintaining remote cluster state information
US7010672B2 (en) 2002-12-11 2006-03-07 Infineon Technologies Ag Digital processor with programmable breakpoint/watchpoint trigger generation circuit
JP4182801B2 (en) * 2003-04-24 2008-11-19 日本電気株式会社 Multi-processor system
US7296137B2 (en) 2005-05-27 2007-11-13 Freescale Semiconductor, Inc. Memory management circuitry translation information retrieval during debugging
US7299335B2 (en) 2005-05-27 2007-11-20 Freescale Semiconductor, Inc. Translation information retrieval transparent to processor core
US7555605B2 (en) 2006-09-28 2009-06-30 Freescale Semiconductor, Inc. Data processing system having cache memory debugging support and method therefor

Patent Citations (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4843542A (en) * 1986-11-12 1989-06-27 Xerox Corporation Virtual memory cache for use in multi-processing systems
US5420999A (en) * 1992-03-27 1995-05-30 International Business Machines Corporation Free memory cell management
US5561786A (en) * 1992-07-24 1996-10-01 Microsoft Corporation Computer method and system for allocating and freeing memory utilizing segmenting and free block lists
US5689679A (en) * 1993-04-28 1997-11-18 Digital Equipment Corporation Memory system and method for selective multi-level caching using a cache level code
US5471598A (en) * 1993-10-18 1995-11-28 Cyrix Corporation Data dependency detection and handling in a microprocessor with write buffer
US5765199A (en) * 1994-01-31 1998-06-09 Motorola, Inc. Data processor with alocate bit and method of operation
US5561779A (en) * 1994-05-04 1996-10-01 Compaq Computer Corporation Processor board having a second level writeback cache system and a third level writethrough cache system which stores exclusive state information for use in a multiprocessor computer system
US5829027A (en) * 1994-05-04 1998-10-27 Compaq Computer Corporation Removable processor board having first, second and third level cache system for use in a multiprocessor computer system
US5685011A (en) * 1995-05-15 1997-11-04 Nvidia Corporation Apparatus for handling failures to provide a safe address translation in an improved input/output architecture for a computer system
US5689707A (en) * 1995-12-04 1997-11-18 Ncr Corporation Method and apparatus for detecting memory leaks using expiration events and dependent pointers to indicate when a memory allocation should be de-allocated
US5819304A (en) * 1996-01-29 1998-10-06 Iowa State University Research Foundation, Inc. Random access memory assembly
US5913923A (en) * 1996-12-06 1999-06-22 National Semiconductor Corporation Multiple bus master computer system employing a shared address translation unit
US5990914A (en) * 1997-09-09 1999-11-23 Compaq Computer Corporation Generating an error signal when accessing an invalid memory page
US6741254B1 (en) * 1997-12-30 2004-05-25 Micron Technology, Inc. Method of implementing an accelerated graphics port for a multiple memory controller computer system
US6145122A (en) * 1998-04-27 2000-11-07 Motorola, Inc. Development interface for a data processor
US6560688B1 (en) * 1998-10-01 2003-05-06 Advanced Micro Devices, Inc. System and method for improving accelerated graphics port systems
US6353829B1 (en) * 1998-12-23 2002-03-05 Cray Inc. Method and system for memory allocation in a multiprocessing environment
US6496902B1 (en) * 1998-12-31 2002-12-17 Cray Inc. Vector and scalar data cache for a vector multiprocessor
US20010032305A1 (en) * 2000-02-24 2001-10-18 Barry Edwin F. Methods and apparatus for dual-use coprocessing/debug interface
US6934937B1 (en) * 2000-03-30 2005-08-23 Broadcom Corporation Multi-channel, multi-service debug on a pipelined CPU architecture
US20020174316A1 (en) * 2001-05-18 2002-11-21 Telgen Corporation Dynamic resource management and allocation in a distributed processing device
US6684296B2 (en) * 2001-05-18 2004-01-27 Broadcom Corporation Source controlled cache allocation
US7117300B1 (en) * 2001-12-27 2006-10-03 James David V Method and apparatus for restricted search operation in content addressable memory (CAM) devices
US20040205420A1 (en) * 2002-06-26 2004-10-14 Charles Seeley Error reporting network in multiprocessor computer
US20040193831A1 (en) * 2003-03-25 2004-09-30 Moyer William C. Memory management in a data processing system
US6954826B2 (en) * 2003-05-21 2005-10-11 Freescale Semiconductor, Inc. Read access and storage circuitry read allocation applicable to a cache
US20050273562A1 (en) * 2003-05-21 2005-12-08 Moyer William C Read access and storage circuitry read allocation applicable to a cache
US7185148B2 (en) * 2003-05-21 2007-02-27 Freescale Semiconductor, Inc. Read access and storage circuitry read allocation applicable to a cache
US20060085489A1 (en) * 2004-10-15 2006-04-20 Microsoft Corporation Memory cache management in XML/relational data mapping
US20060212426A1 (en) * 2004-12-21 2006-09-21 Udaya Shakara Efficient CAM-based techniques to perform string searches in packet payloads
US7506105B2 (en) * 2005-05-02 2009-03-17 Freescale Semiconductor, Inc. Prefetching using hashed program counter
US20080010409A1 (en) * 2006-07-05 2008-01-10 Cisco Technology, Inc. Dynamic, on-demand storage area network (SAN) cache

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7647539B2 (en) 2007-07-18 2010-01-12 International Business Machines Corporation System and method of testing using test pattern re-execution in varying timing scenarios for processor design verification and validation
US8127192B2 (en) 2007-07-18 2012-02-28 International Business Machines Corporation Predicting lwarx and stwcx instructions in test pattern generation and simulation for processor design verification/validation in interrupt mode
US20090024886A1 (en) * 2007-07-18 2009-01-22 Sampan Arora System and Method for Predicting lwarx and stwcx Instructions in Test Pattern Generation and Simulation for Processor Design Verification and Validation
US20090024894A1 (en) * 2007-07-18 2009-01-22 International Business Machines Corporation System and method for predicting iwarx and stwcx instructions in test pattern generation and simulation for processor design verification/validation in interrupt mode
US20090024873A1 (en) * 2007-07-18 2009-01-22 Sandip Bag System and Method for Increasing Error Checking Performance by Calculating CRC Calculations After Multiple Test Patterns for Processor Design Verification and Validation
US20090024877A1 (en) * 2007-07-18 2009-01-22 Shubhodeep Roy Choudhury System and Method for Creating Different Start Cache and Bus States Using Multiple Test Patterns for Processor Design Verification and Validation
US20090024876A1 (en) * 2007-07-18 2009-01-22 Sampan Arora System and Method for Verification of Cache Snoop Logic and Coherency Between Instruction & Data Caches for Processor Design Verification and Validation
US20090024892A1 (en) * 2007-07-18 2009-01-22 Vinod Bussa System and Method of Testing using Test Pattern Re-Execution in Varying Timing Scenarios for Processor Design Verification and Validation
US7739570B2 (en) 2007-07-18 2010-06-15 International Business Machines Corporation System and method for increasing error checking performance by calculating CRC calculations after multiple test patterns for processor design verification and validation
US7689886B2 (en) 2007-07-18 2010-03-30 International Business Machines Corporation System and method for predicting lwarx and stwcx instructions in test pattern generation and simulation for processor design verification and validation
US7661023B2 (en) * 2007-07-18 2010-02-09 International Business Machines Corporation System and method for verification of cache snoop logic and coherency between instruction & data caches for processor design verification and validation
US7747908B2 (en) 2007-07-18 2010-06-29 International Business Machines Corporation System and method for creating different start cache and bus states using multiple test patterns for processor design verification and validation
US20100023697A1 (en) * 2008-07-25 2010-01-28 International Business Machines Corporation Testing Real Page Number Bits in a Cache Directory
US8185694B2 (en) 2008-07-25 2012-05-22 International Business Machines Corporation Testing real page number bits in a cache directory
US20100205377A1 (en) * 2009-02-06 2010-08-12 Moyer William C Debug control for snoop operations in a multiprocessor system and method thereof
US20100205378A1 (en) * 2009-02-06 2010-08-12 Moyer William C Method for debugger initiated coherency transactions using a shared coherency manager
US8200908B2 (en) 2009-02-06 2012-06-12 Freescale Semiconductor, Inc. Method for debugger initiated coherency transactions using a shared coherency manager
US8688910B2 (en) * 2009-02-06 2014-04-01 Freescale Semiconductor, Inc. Debug control for snoop operations in a multiprocessor system and method thereof
US8935485B2 (en) 2011-08-08 2015-01-13 Arm Limited Snoop filter and non-inclusive shared cache memory
US9477600B2 (en) 2011-08-08 2016-10-25 Arm Limited Apparatus and method for shared cache control including cache lines selectively operable in inclusive or non-inclusive mode
EP2838024A3 (en) * 2013-07-25 2015-07-01 Altera Corporation Cache debug system for programmable circuits
US9594655B2 (en) 2013-07-25 2017-03-14 Altera Corporation Cache debug system for programmable circuits
US10108568B2 (en) 2015-03-30 2018-10-23 Samsung Electronics Co., Ltd. Master capable of communicating with slave and system including the master
US10261827B2 (en) 2015-10-29 2019-04-16 International Business Machines Corporation Interprocessor memory status communication
US10261828B2 (en) 2015-10-29 2019-04-16 International Business Machines Corporation Interprocessor memory status communication
US10346305B2 (en) * 2015-10-29 2019-07-09 International Business Machines Corporation Interprocessor memory status communication
US10884931B2 (en) 2015-10-29 2021-01-05 International Business Machines Corporation Interprocessor memory status communication

Also Published As

Publication number Publication date
US8972671B2 (en) 2015-03-03
KR20100016516A (en) 2010-02-12
CN101918923A (en) 2010-12-15
JP2010527488A (en) 2010-08-12
US20080288724A1 (en) 2008-11-20
WO2008144179A8 (en) 2010-01-21
WO2008144179A1 (en) 2008-11-27

Similar Documents

Publication Publication Date Title
US8972671B2 (en) Method and apparatus for cache transactions in a data processing system
USRE49305E1 (en) Data processing system having cache memory debugging support and method therefor
US7055003B2 (en) Data cache scrub mechanism for large L2/L3 data cache structures
US5996048A (en) Inclusion vector architecture for a level two cache
US6564306B2 (en) Apparatus and method for performing speculative cache directory tag updates
US5940856A (en) Cache intervention from only one of many cache lines sharing an unmodified value
US7032074B2 (en) Method and mechanism to use a cache to translate from a virtual bus to a physical bus
US5524208A (en) Method and apparatus for performing cache snoop testing using DMA cycles in a computer system
US5963974A (en) Cache intervention from a cache line exclusively holding an unmodified value
US5946709A (en) Shared intervention protocol for SMP bus using caches, snooping, tags and prioritizing
US5940864A (en) Shared memory-access priorization method for multiprocessors using caches and snoop responses
US20020087809A1 (en) Multiprocessor computer system with sectored cache line mechanism for cache intervention
JPH05204756A (en) Digital processor
US8918591B2 (en) Data processing system having selective invalidation of snoop requests and method therefor
JPH05216756A (en) Processor and cache memory
US8200908B2 (en) Method for debugger initiated coherency transactions using a shared coherency manager
US5943685A (en) Method of shared intervention via a single data provider among shared caches for SMP bus
JPH05128000A (en) Microprocessor having built-in cache memory and method for testing built-in cache memory
US20220391283A1 (en) Handling non-correctable errors
JPH05210583A (en) Processor and method of executing correct sequencing of write data
EP0976047B1 (en) Read operations in multiprocessor computer system
US20060143333A1 (en) I/O hub resident cache line monitor and device register update
US7055002B2 (en) Integrated purge store mechanism to flush L2/L3 cache structure for improved reliabity and serviceability
JP2780555B2 (en) Microprocessor with built-in cache memory

Legal Events

Date Code Title Description
AS Assignment

Owner name: FREESCALE SEMICONDUCTOR, INC., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MOYER, WILLIAM C.;SNYDER, MICHAEL D.;REEL/FRAME:019300/0591

Effective date: 20070510

AS Assignment

Owner name: CITIBANK, N.A., AS COLLATERAL AGENT, NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:020045/0448

Effective date: 20070718

Owner name: CITIBANK, N.A., AS COLLATERAL AGENT,NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:020045/0448

Effective date: 20070718

AS Assignment

Owner name: CITIBANK, N.A.,NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:024085/0001

Effective date: 20100219

Owner name: CITIBANK, N.A., NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:024085/0001

Effective date: 20100219

AS Assignment

Owner name: CITIBANK, N.A., AS COLLATERAL AGENT,NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:024397/0001

Effective date: 20100413

Owner name: CITIBANK, N.A., AS COLLATERAL AGENT, NEW YORK

Free format text: SECURITY AGREEMENT;ASSIGNOR:FREESCALE SEMICONDUCTOR, INC.;REEL/FRAME:024397/0001

Effective date: 20100413

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: FREESCALE SEMICONDUCTOR, INC., TEXAS

Free format text: PATENT RELEASE;ASSIGNOR:CITIBANK, N.A., AS COLLATERAL AGENT;REEL/FRAME:037356/0143

Effective date: 20151207

Owner name: FREESCALE SEMICONDUCTOR, INC., TEXAS

Free format text: PATENT RELEASE;ASSIGNOR:CITIBANK, N.A., AS COLLATERAL AGENT;REEL/FRAME:037354/0655

Effective date: 20151207

Owner name: FREESCALE SEMICONDUCTOR, INC., TEXAS

Free format text: PATENT RELEASE;ASSIGNOR:CITIBANK, N.A., AS COLLATERAL AGENT;REEL/FRAME:037356/0553

Effective date: 20151207