会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 33. 发明授权
    • Cache index based system address bus
    • 基于缓存索引的系统地址总线
    • US06477613B1
    • 2002-11-05
    • US09345302
    • 1999-06-30
    • Ravi Kumar ArimilliJohn Steven DodsonGuy Lynn GuthrieJody B. JoynerJerry Don Lewis
    • Ravi Kumar ArimilliJohn Steven DodsonGuy Lynn GuthrieJody B. JoynerJerry Don Lewis
    • G06F1200
    • G06F12/0811G06F12/0895G06F12/0897
    • Following a cache miss by an operation, the address for the operation is transmitted on the bus coupling the cache to lower levels of the storage hierarchy. A portion of the address including the index field is transmitted during a first bus cycle, and may be employed to begin directory lookups in lower level storage devices before the address tag is received. The remainder of the address is transmitted during subsequent bus cycles, which should be in time for address tag comparisons with the congruence class elements. To allow multiple directory lookups to be occurring concurrently in a pipelined directory, a portion of multiple addresses for several data access operations, each portion including the index field for the respective address, may be transmitted during the first bus cycle or staged in consecutive bus cycles, with the remainders of each address—including the cache tags—transmitted during the subsequent bus cycles. This allows directory lookups utilizing the index fields to be processed concurrently within a lower level storage device for multiple operations, with the address tags being provided later, but still timely for tag comparisons at the end of the directory lookup. Where the lower level storage device operates at a higher frequency than the bus, overall latency is reduced and directory bandwidth is more efficiently utilized.
    • 在操作的高速缓存未命中之后,操作的地址在将高速缓存耦合到存储层级的较低级别的总线上传输。 包括索引字段的地址的一部分在第一总线周期期间被发送,并且可以用于在接收到地址标签之前开始下级存储设备中的目录查找。 在随后的总线周期期间传送地址的其余部分,这些时间应与地址标签与同余类元素进行比较。 为了允许在流水线目录中同时发生多个目录查找,可以在第一个总线周期期间发送多个数据访问操作的多个地址的一部分,每个部分包括相应地址的索引字段,或者在连续的总线周期中分段 ,每个地址的剩余部分,包括在后续总线周期期间发送的缓存标签。 这允许使用索引字段的目录查找在较低级存储设备中同时处理以用于多个操作,其中地址标签稍后提供,但是在目录查找结束时仍然适合于标签比较。 在较低级存储设备以比总线更高的频率工作的地方,总体延迟降低,目录带宽更有效地利用。
    • 35. 发明授权
    • Layered local cache with imprecise reload mechanism
    • 分层本地缓存与不精确的重载机制
    • US06434667B1
    • 2002-08-13
    • US09340075
    • 1999-06-25
    • Ravi Kumar ArimilliLeo James ClarkJohn Steven DodsonGuy Lynn Guthrie
    • Ravi Kumar ArimilliLeo James ClarkJohn Steven DodsonGuy Lynn Guthrie
    • G06F1200
    • G06F12/0811G06F12/0859G06F12/0888
    • A method of improving memory access for a computer system, by sending load requests to a lower level storage subsystem along with associated information pertaining to intended use of the requested information by the requesting processor, without using a high level load queue. Returning the requested information to the processor along with the associated use information allows the information to be placed immediately without using reload buffers. A register load bus separate from the cache load bus (and having a smaller granularity) is used to return the information. An upper level (L1) cache may then be imprecisely reloaded (the upper level cache can also be imprecisely reloaded with store instructions). The lower level (L2) cache can monitor L1 and L2 cache activity, which can be used to select a victim cache block in the L1 cache (based on the additional L2 information), or to select a victim cache block in the L2 cache (based on the additional L1 information). L2 control of the L1 directory also allows certain snoop requests to be resolved without waiting for L1 acknowledgement. The invention can be applied to, e.g., instruction, operand data and translation caches.
    • 一种改进计算机系统的存储器访问的方法,通过将请求发送到较低级别的存储子系统以及由请求处理器对与请求的信息的预期用途有关的关联信息而不使用高级别的负载队列来进行发送。 将所请求的信息与相关联的使用信息一起返回到处理器允许立即放置信息而不使用重新加载缓冲器。 使用与缓存负载总线分离(并具有较小粒度)的寄存器负载总线返回信息。 然后可能不精确地重新加载上级(L1)高速缓存(高级缓存也可以不精确地用存储指令重新加载)。 低级(L​​2)缓存可以监视L1和L2高速缓存活动,其可用于在L1高速缓存中选择受害者缓存块(基于附加的L2信息),或者选择L2缓存中的受害缓存块( 基于附加的L1信息)。 L1目录的L2控制也允许解决某些侦听请求,而无需等待L1确认。 本发明可以应用于例如指令,操作数数据和翻译高速缓存。
    • 36. 发明授权
    • Queue-less and state-less layered local data cache mechanism
    • 无队列和无状态的分层本地数据缓存机制
    • US06418513B1
    • 2002-07-09
    • US09340077
    • 1999-06-25
    • Ravi Kumar ArimilliLeo James ClarkJohn Steven DodsonGuy Lynn Guthrie
    • Ravi Kumar ArimilliLeo James ClarkJohn Steven DodsonGuy Lynn Guthrie
    • G06F1200
    • G06F12/0897
    • A method of improving memory access for a computer system, by sending load requests to a lower level storage subsystem along with associated information pertaining to intended use of the requested information by the requesting processor, without using a high level load queue. Returning the requested information to the processor along with the associated use information allows the information to be placed immediately without using reload buffers. A register load bus separate from the cache load bus (and having a smaller granularity) is used to return the information. An upper level (L1) cache may then be imprecisely reloaded (the upper level cache can also be imprecisely reloaded with store instructions). The lower level (L2) cache can monitor L1 and L2 cache activity, which can be used to select a victim cache block in the L1 cache (based on the additional L2 information), or to select a victim cache block in the L2 cache (based on the additional L1 information). L2 control of the L1 directory also allows certain snoop requests to be resolved without waiting for L1 acknowledgement. The invention can be applied to, e.g., instruction, operand data and translation caches.
    • 一种改进计算机系统的存储器访问的方法,通过将请求发送到较低级别的存储子系统以及由请求处理器对与请求的信息的预期用途有关的关联信息而不使用高级别的负载队列来进行发送。 将所请求的信息与相关联的使用信息一起返回到处理器允许立即放置信息而不使用重新加载缓冲器。 使用与缓存负载总线分离(并具有较小粒度)的寄存器负载总线返回信息。 然后可能不精确地重新加载上级(L1)高速缓存(高级缓存也可以不精确地用存储指令重新加载)。 低级(L​​2)缓存可以监视L1和L2高速缓存活动,其可用于在L1高速缓存中选择受害者缓存块(基于附加的L2信息),或者选择L2缓存中的受害缓存块( 基于附加的L1信息)。 L1目录的L2控制也允许解决某些侦听请求,而无需等待L1确认。 本发明可以应用于例如指令,操作数数据和翻译高速缓存。
    • 38. 发明授权
    • High performance multiprocessor system with exclusive-deallocate cache state
    • 具有独占解除缓存状态的高性能多处理器系统
    • US06385702B1
    • 2002-05-07
    • US09437198
    • 1999-11-09
    • Ravi Kumar ArimilliLakshminarayana Baba ArimilliJohn Steven DodsonGuy Lynn GuthrieWilliam John Starke
    • Ravi Kumar ArimilliLakshminarayana Baba ArimilliJohn Steven DodsonGuy Lynn GuthrieWilliam John Starke
    • G06F1200
    • G06F12/0831G06F12/0811
    • A cache coherency protocol uses a “Exclusive-Deallocate” (ED) coherency state to indicate that a particular value is currently held in an upper level cache in an exclusive, unmodified form (not shared with any other caches of the computer system, including caches associated with the same processing unit), so that the value can conveniently be modified without any lower level bus transactions since no lower level caches have allocated a line for the value. If the value is subsequently modified in the upper level cache, its coherency state is simply switched to “modified” without the need for any bus transactions. Conversely, if the value is evicted from the upper level cache without ever having been modified, it can be loaded into the lower level cache with a coherency state indicating that the lower level cache contains the unmodified value exclusive of all other caches in other processing units of the computer system. If the value is initially loaded into the upper level cache from a cache of another processing unit, or from a lower level cache of the same processing unit, then the upper level cache may be selectively programmed to mark the cache line with the ED state.
    • 高速缓存一致性协议使用“独占解除分配”(ED)一致性状态来指示特定值当前以独占未修改的形式(不与计算机系统的任何其他高速缓存共享,包括高速缓存)保持在高级缓存中 与相同的处理单元关联),使得该值可以方便地被修改而没有任何较低级别的总线事务,因为没有较低级别的高速缓存已经为该值分配了一行。 如果该值随后在高级缓存中被修改,则其一致性状态被简单地切换到“修改”,而不需要任何总线事务。 相反,如果该值从上级缓存中被逐出而没有被修改,则可以将其加载到具有一致性状态的相关性状态中,该相关性状态指示低级缓存包含其他处理单元中所有其他高速缓存的排他性的未修改值 的计算机系统。 如果该值最初从另一处理单元的高速缓存或相同处理单元的较低级高速缓存加载到高级缓存中,则可以选择性地编程高级缓存以用ED状态标记高速缓存行。