会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 22. 发明申请
    • METHOD, APPARATUS, AND COMPUTER PROGRAM PRODUCT FOR A CACHE COHERENCY PROTOCOL STATE THAT PREDICTS LOCATIONS OF SHARED MEMORY BLOCKS
    • 用于预测共享存储块位置的高速缓存协议状态的方法,设备和计算机程序产品
    • US20080215819A1
    • 2008-09-04
    • US12107350
    • 2008-04-22
    • Jason Frederick CantinSteven R. Kunkel
    • Jason Frederick CantinSteven R. Kunkel
    • G06F12/08
    • G06F12/0831G06F12/0813G06F2212/507
    • A method, apparatus, and computer program product are disclosed for reducing the number of unnecessarily broadcast local requests to reduce the latency to access data from remote nodes in an SMP computer system. A shared invalid cache coherency protocol state is defined that predicts whether a memory read request to read data in a shared cache line can be satisfied within a local node. When a cache line is in the shared invalid state, a valid copy of the data is predicted to be located in the local node. When a cache line is in the invalid state and not in the shared invalid state, a valid copy of the data is predicted to be located in one of the remote nodes.Memory read requests to read data in a cache line that is not currently in the shared invalid state are broadcast first to remote nodes. Memory read requests to read data in a cache line that is currently in the shared invalid state are broadcast first to a local node, and in response to being unable to satisfy the memory read requests within the local node, the memory read requests are broadcast to the remote nodes.
    • 公开了用于减少不必要地广播的本地请求的数量以减少从SMP计算机系统中的远程节点访问数据的等待时间的方法,装置和计算机程序产品。 定义共享的无效高速缓存一致性协议状态,其预测在本地节点内是否可以满足在共享高速缓存行中读取数据的存储器读取请求。 当高速缓存行处于共享无效状态时,预测数据的有效副本位于本地节点中。 当高速缓存行处于无效状态而不处于共享无效状态时,预测数据的有效副本位于远程节点之一中。 在当前处于共享无效状态的缓存行中读取数据的内存读取请求首先被广播到远程节点。 在当前处于共享无效状态的高速缓存行中读取数据的存储器读取请求首先被广播到本地节点,并且响应于不能满足本地节点内的存储器读取请求,存储器读取请求被广播到 远程节点。
    • 24. 发明授权
    • Cache updating in multiprocessor systems
    • 多处理器系统中的缓存更新
    • US06728842B2
    • 2004-04-27
    • US10061859
    • 2002-02-01
    • Jeffrey D. BrownSteven R. KunkelDavid A. Luick
    • Jeffrey D. BrownSteven R. KunkelDavid A. Luick
    • G06F1200
    • G06F12/0831
    • Embodiments are provided in which cache update is implemented by using a counter table having a plurality of entries to keep track of different modified cache lines of a cache of a processor. If a cache line of the cache is modified by the processor and the original content of the cache line came from a cache of another processor, a counter in the counter table restarts and reaches a predetermined value (e.g., overflows) triggering the broadcast of the modified cache line so that the cache of the other processor can snarf a copy of the modified cache line. As a result, when the other processor reads from a memory address matching that of the cache line, the cache of the other processor already has the most current copy for the matching memory address to feed the processor. Therefore, a cache read miss is avoided and system performance is improved.
    • 提供了实施例,其中通过使用具有多个条目的计数器表来实现高速缓存更新,以跟踪处理器的高速缓存的不同修改的高速缓存行。 如果高速缓存的高速缓存行被处理器修改,并且高速缓存行的原始内容来自另一个处理器的高速缓存,则计数器表中的计数器重新启动并达到触发广播的预定值(例如,溢出) 修改的高速缓存行,使得其他处理器的缓存可以绕过修改的高速缓存行的副本。 结果,当另一个处理器从与高速缓存行的存储器地址匹配的存储器地址读取时,另一个处理器的高速缓存器已经具有用于匹配存储器地址的最新的副本来馈送处理器。 因此,避免了缓存读取缺失,提高了系统性能。
    • 26. 发明授权
    • Local and global memory request predictor
    • 本地和全局内存请求预测器
    • US08874853B2
    • 2014-10-28
    • US12793795
    • 2010-06-04
    • Jason F. CantinSteven R. Kunkel
    • Jason F. CantinSteven R. Kunkel
    • G06F12/00G06F12/08
    • G06F12/0813
    • A method, circuit arrangement, and design structure utilize broadcast prediction data to determine whether to globally broadcast a memory request in a computing system of the type that includes a plurality of nodes, each node including a plurality of processing units. The method includes updating broadcast prediction data for a cache line associated with a first memory request within a hardware-based broadcast prediction data structure in turn associated with a first processing unit in response to the first memory request, the broadcast prediction data for the cache line including data associated with a history of ownership of the cache line. The method further comprises accessing the broadcast prediction data structure and determining whether to perform an early broadcast of a second memory request to a second node based on broadcast prediction data within the broadcast prediction data structure in response to that second memory request associated with the cache line.
    • 一种方法,电路装置和设计结构利用广播预测数据来确定是否在包括多个节点的类型的计算系统中全局广播存储器请求,每个节点包括多个处理单元。 所述方法包括响应于第一存储器请求,在基于硬件的广播预测数据结构中更新与第一存储器请求相关联的高速缓存行的广播预测数据,该第一存储器请求与第一处理单元相对应,用于高速缓存行的广播预测数据 包括与高速缓存行的所有权历史相关联的数据。 该方法还包括响应于与高速缓存行相关联的第二存储器请求,基于广播预测数据结构内的广播预测数据,来访问广播预测数据结构并确定是否对第二节点执行第二存储器请求的早期广播 。
    • 27. 发明授权
    • Predictive ownership control of shared memory computing system data
    • 共享内存计算系统数据的预测所有权控制
    • US08244988B2
    • 2012-08-14
    • US12432835
    • 2009-04-30
    • Jason F. CantinSteven R. Kunkel
    • Jason F. CantinSteven R. Kunkel
    • G06F12/00
    • G06F12/084G06F9/528G06F12/0842G06F12/126G06F2212/1016
    • A method, circuit arrangement, and design structure utilize a lock prediction data structure to control ownership of a cache line in a shared memory computing system. In a first node among the plurality of nodes, lock prediction data in a hardware-based lock prediction data structure for a cache line associated with a first memory request is updated in response to that first memory request, wherein at least a portion of the lock prediction data is predictive of whether the cache line is associated with a release operation. The lock prediction data is then accessed in response to a second memory request associated with the cache line and issued by a second node and a determination is made as to whether to transfer ownership of the cache line from the first node to the second node based at least in part on the accessed lock prediction data.
    • 方法,电路布置和设计结构利用锁预测数据结构来控制共享存储器计算系统中的高速缓存行的所有权。 在所述多个节点中的第一节点中,响应于所述第一存储器请求来更新与用于与第一存储器请求相关联的高速缓存行的基于硬件的锁预测数据结构中的预测数据锁定,其中所述锁的至少一部分 预测数据预测高速缓存行是否与释放操作相关联。 然后,锁定预测数据被响应于与高速缓存线相关联并由第二节点发出的第二存储器请求而被访问,并且基于在第二节点处是否将高速缓存行的所有权从第一节点转移到第二节点进行确定 至少部分地基于所访问的锁定预测数据。
    • 28. 发明授权
    • Shared data prefetching with memory region cache line monitoring
    • 共享数据预取与内存区域缓存行监视
    • US08112587B2
    • 2012-02-07
    • US12432823
    • 2009-04-30
    • Jason F. CantinSteven R. Kunkel
    • Jason F. CantinSteven R. Kunkel
    • G06F13/28
    • G06F12/0862G06F12/0813G06F12/0815G06F2212/6022
    • A method, circuit arrangement, and design structure for prefetching data for responding to a memory request, in a shared memory computing system of the type that includes a plurality of nodes, is provided. Prefetching data comprises, receiving, in response to a first memory request by a first node, presence data for a memory region associated with the first memory request from a second node that sources data requested by the first memory request, and selectively prefetching at least one cache line from the memory region based on the received presence data. Responding to a memory request comprises tracking presence data associated with memory regions associated with cached cache lines in the first node, and, in response to a memory request by a second node, forwarding the tracked presence data for a memory region associated with the memory request to the second node.
    • 提供了在包括多个节点的类型的共享存储器计算系统中的用于预取数据以响应存储器请求的方法,电路布置和设计结构。 预取数据包括:响应于第一节点的第一存储器请求,接收来自第二节点的与第一存储器请求相关联的存储器区域的存在数据,该第二节点发送由第一存储器请求请求的数据,并且选择性地预取至少一个 基于接收的存在数据从存储器区域的高速缓存行。 响应于存储器请求包括跟踪与第一节点中与高速缓存行相关联的存储器区域的存在数据,并且响应于第二节点的存储器请求,转发与存储器请求相关联的存储器区域的跟踪存在数据 到第二个节点。
    • 30. 发明申请
    • METHOD, APPARATUS, AND COMPUTER PROGRAM PRODUCT FOR A CACHE COHERENCY PROTOCOL STATE THAT PREDICTS LOCATIONS OF MODIFIED MEMORY BLOCKS
    • 用于预测改进的存储块位置的高速缓存协议状态的方法,设备和计算机程序产品
    • US20080147987A1
    • 2008-06-19
    • US11954742
    • 2007-12-12
    • JASON FREDERICK CANTINSteven R. Kunkel
    • JASON FREDERICK CANTINSteven R. Kunkel
    • G06F12/00
    • G06F12/0831G06F12/0813G06F2212/507
    • A method, apparatus, and computer program product are disclosed for reducing the number of unnecessarily broadcast remote requests to reduce the latency to access data from local nodes and to reduce global traffic in an SMP computer system. A modified invalid cache coherency protocol state is defined that predicts whether a memory access request to read or write data in a cache line can be satisfied within a local node. When a cache line is in the modified invalid state, the only valid copies of the data are predicted to be located in the local node. When a cache line is in the invalid state and not in the modified invalid state, a valid copy of the data is predicted to be located in one of the remote nodes.Memory access requests to read exclusive or write data in a cache line that is not currently in the modified invalid state are broadcast first to all nodes. Memory access requests to read exclusive or write data in a cache line that is currently in the modified invalid state are broadcast first to a local node, and in response to being unable to satisfy the memory access requests within the local node, the memory access requests are broadcast to the remote nodes.
    • 公开了一种方法,装置和计算机程序产品,用于减少不必要地广播的远程请求的数量,以减少从本地节点访问数据的等待时间并减少SMP计算机系统中的全局流量。 定义了修改的无效高速缓存一致性协议状态,其预测在本地节点内是否可以满足在高速缓存行中读取或写入数据的存储器访问请求。 当缓存行处于修改的无效状态时,数据的唯一有效副本被预测位于本地节点中。 当高速缓存行处于无效状态而不处于修改的无效状态时,预测数据的有效副本位于远程节点之一中。 在当前处于修改的无效状态的高速缓存行中读取独占或写入数据的存储器访问请求首先被广播到所有节点。 在当前处于修改的无效状态的高速缓存行中读取独占或写入数据的存储器访问请求首先被广播到本地节点,并且响应于不能满足本地节点内的存储器访问请求,存储器访问请求 广播到远程节点。