会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 31. 发明授权
    • Virtual barrier synchronization cache castout election
    • 虚拟屏障同步缓存突发选举
    • US08095733B2
    • 2012-01-10
    • US12419343
    • 2009-04-07
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • G06F13/00G06F13/28
    • G06F12/0811G06F9/30101G06F9/3851G06F9/522
    • A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region. Each of the plurality of processing units includes a processor core and a cache memory including a cache controller and a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory. The cache controller of a first processing unit, responsive to a memory access request from its processor core that targets a first VBSR line, transfers responsibility for writing back to the virtual barrier synchronization region a second VBSR line contemporaneously held in the cache arrays of first, second and third processing units. The responsibility is transferred via an election held over the interconnect fabric.
    • 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括高速缓存控制器和从系统存储器的虚拟屏障同步区域缓存VBSR行的高速缓存阵列。 响应于来自其处理器核心的第一VBSR线路的存储器访问请求的第一处理单元的高速缓存控制器将负责向第一虚拟屏障同步区域写回同时保存在第一VBSR线路的高速缓存阵列中的第二VBSR线路, 第二和第三处理单元。 通过互连结构上的选举来转移责任。
    • 32. 发明申请
    • Virtual Barrier Synchronization Cache Castout Election
    • 虚拟障碍同步缓存铸造选举
    • US20100257316A1
    • 2010-10-07
    • US12419343
    • 2009-04-07
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • G06F12/08G06F12/00
    • G06F12/0811G06F9/30101G06F9/3851G06F9/522
    • A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region. Each of the plurality of processing units includes a processor core and a cache memory including a cache controller and a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory. The cache controller of a first processing unit, responsive to a memory access request from its processor core that targets a first VBSR line, transfers responsibility for writing back to the virtual barrier synchronization region a second VBSR line contemporaneously held in the cache arrays of first, second and third processing units. The responsibility is transferred via an election held over the interconnect fabric.
    • 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括高速缓存控制器和从系统存储器的虚拟屏障同步区域缓存VBSR行的高速缓存阵列。 响应于来自其处理器核心的第一VBSR线路的存储器访问请求的第一处理单元的高速缓存控制器将负责向第一虚拟屏障同步区域写回同时保存在第一VBSR线路的高速缓存阵列中的第二VBSR线路, 第二和第三处理单元。 通过互连结构上的选举来转移责任。
    • 35. 发明授权
    • Partial cache line storage-modifying operation based upon a hint
    • 基于提示的部分缓存行存储修改操作
    • US08140771B2
    • 2012-03-20
    • US12024424
    • 2008-02-01
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/04G06F9/312
    • G06F12/0822
    • In at least one embodiment, a method of data processing in a data processing system having a memory hierarchy includes a processor core executing a storage-modifying memory access instruction to determine a memory address. The processor core transmits to a cache memory within the memory hierarchy a storage-modifying memory access request including the memory address, an indication of a memory access type, and, if present, a partial cache line hint signaling access to less than all granules of a target cache line of data associated with the memory address. In response to the storage-modifying memory access request, the cache memory performs a storage-modifying access to all granules of the target cache line of data if the partial cache line hint is not present and performs a storage-modifying access to less than all granules of the target cache line of data if the partial cache line hint is present.
    • 在至少一个实施例中,具有存储器层次的数据处理系统中的数据处理方法包括执行存储修改存储器访问指令以确定存储器地址的处理器核心。 处理器核心向存储器层级内的高速缓冲存储器传送存储修改存储器访问请求,该存储修改存储器访问请求包括存储器地址,存储器访问类型的指示,以及如果存在的话,部分高速缓存行提示信令访问少于所有颗粒的 与存储器地址相关联的数据的目标高速缓存行。 响应于存储修改存储器访问请求,如果不存在部分高速缓存行提示,则高速缓存存储器对目标高速缓存行数据行的所有颗粒进行存储修改访问,并执行对小于全部的存储修改访问 如果存在部分高速缓存线提示,则目标高速缓存行数据的颗粒。
    • 36. 发明授权
    • Data processing system, method and interconnect fabric supporting multiple planes of processing nodes
    • 支持多个处理节点平面的数据处理系统,方法和互连结构
    • US07818388B2
    • 2010-10-19
    • US11245887
    • 2005-10-07
    • Ravi K. ArimilliBenjiman L. GoodmanGuy L. GuthriePraveen S. ReddyWilliam J. Starke
    • Ravi K. ArimilliBenjiman L. GoodmanGuy L. GuthriePraveen S. ReddyWilliam J. Starke
    • G06F15/16
    • G06F15/16
    • A data processing system includes a first plane including a first plurality of processing nodes, each including multiple processing units, and a second plane including a second plurality of processing nodes, each including multiple processing units. The data processing system also includes a plurality of point-to-point first tier links. Each of the first plurality and second plurality of processing nodes includes one or more first tier links among the plurality of first tier links, where the first tier link(s) within each processing node connect a pair of processing units in the same processing node for communication. The data processing system further includes a plurality of point-to-point second tier links. At least a first of the plurality of second tier links connects processing units in different ones of the first plurality of processing nodes, at least a second of the plurality of second tier links connects processing units in different ones of the second plurality of processing nodes, and at least a third of the plurality of second tier links connects a processing unit in the first plane to a processing unit in the second plane.
    • 数据处理系统包括包括第一多个处理节点的第一平面,每个处理节点包括多个处理单元,以及包括第二多个处理节点的第二平面,每个处理节点包括多个处理单元。 数据处理系统还包括多个点对点第一层链路。 第一多个处理节点和第二多个处理节点中的每一个包括多个第一层链路之中的一个或多个第一层链路,其中每个处理节点内的第一层链路连接相同处理节点中的一对处理单元,用于 通讯。 数据处理系统还包括多个点到点第二层链路。 所述多个第二层链路中的至少第一层连接所述第一多个处理节点中的不同处理节点中的处理单元,所述多个第二层链路中的至少一个链接连接所述第二多个处理节点中的不同处理节点中的处理单元, 并且所述多个第二层链路中的至少三分之一链路将所述第一平面中的处理单元连接到所述第二平面中的处理单元。
    • 37. 发明申请
    • PROCESSOR PERFORMANCE IMPROVEMENT FOR INSTRUCTION SEQUENCES THAT INCLUDE BARRIER INSTRUCTIONS
    • 包括障碍指示的指令序列的处理器性能改进
    • US20130205120A1
    • 2013-08-08
    • US13369029
    • 2012-02-08
    • Guy L GuthrieWilliam J. StarkeDerek E Williams
    • Guy L GuthrieWilliam J. StarkeDerek E Williams
    • G06F9/312
    • G06F9/52G06F9/30087G06F9/30145G06F9/3834G06F12/0831
    • A technique for processing an instruction sequence that includes a barrier instruction, a load instruction preceding the barrier instruction, and a subsequent memory access instruction following the barrier instruction includes determining that the load instruction is resolved based upon receipt of an earliest of a good combined response for a read operation corresponding to the load instruction and data for the load instruction. The technique also includes if execution of the subsequent memory access instruction is not initiated prior to completion of the barrier instruction, initiating in response to determining the barrier instruction completed, execution of the subsequent memory access instruction. The technique further includes if execution of the subsequent memory access instruction is initiated prior to completion of the barrier instruction, discontinuing in response to determining the barrier instruction completed, tracking of the subsequent memory access instruction with respect to invalidation.
    • 一种用于处理指示序列的技术,该指令序列包括屏障指令,屏障指令之前的加载指令,以及跟随障碍指令之后的随后存储器访问指令,包括:基于接收到最早的良好组合响应来确定加载指令是否被解决 用于与加载指令相对应的读取操作和用于加载指令的数据。 该技术还包括如果在完成屏障指令之前没有启动后续存储器访问指令的执行,则响应于确定完成的屏障指令启动后续存储器访问指令的执行。 该技术还包括如果在完成屏障指令之前启动后续存储器访问指令的执行,则响应于确定所完成的屏障指令而中断,跟踪关于无效的后续存储器访问指令。
    • 39. 发明授权
    • Mode-based castout destination selection
    • 基于模式的castout目的地选择
    • US08312220B2
    • 2012-11-13
    • US12420933
    • 2009-04-09
    • Guy L. GuthrieHarmony L. HelterhoffWilliam J. StarkePhillip G. WilliamsJeffrey A. Stuecheli
    • Guy L. GuthrieHarmony L. HelterhoffWilliam J. StarkePhillip G. WilliamsJeffrey A. Stuecheli
    • G06F12/08
    • G06F12/0811G06F12/12
    • In response to a data request of a first of a plurality of processing units, the first processing unit selects a victim cache line to be castout from the lower level cache of the first processing unit and determines whether a mode is set. If not, the first processing unit issues on the interconnect fabric an LCO command identifying the victim cache line and indicating that a lower level cache is the intended destination. If the mode is set, the first processing unit issues a castout command with an alternative intended destination. In response to a coherence response to the LCO command indicating success of the LCO command, the first processing unit removes the victim cache line from its lower level cache, and the victim cache line is held elsewhere in the data processing system. The mode can be set to inhibit castouts to system memory, for example, for testing.
    • 响应于多个处理单元中的第一处理单元的数据请求,第一处理单元从第一处理单元的较低级高速缓存中选择要丢弃的牺牲高速缓存行,并且确定是否设置了模式。 如果不是,则第一处理单元在互连结构上发出识别受害者高速缓存行的LCO命令,并指示较低级别的高速缓存是预期的目的地。 如果模式被设置,则第一处理单元发出具有替代预定目的地的停顿命令。 响应于指示LCO命令成功的LCO命令的一致性响应,第一处理单元从其较低级高速缓存中去除受害者高速缓存行,并且将受害者高速缓存行保持在数据处理系统的其他地方。 该模式可以设置为抑制系统内存的丢弃,例如进行测试。
    • 40. 发明授权
    • Victim cache prefetching
    • 受害者缓存预取
    • US08209489B2
    • 2012-06-26
    • US12256064
    • 2008-10-22
    • Guy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliPhillip G. Williams
    • Guy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliPhillip G. Williams
    • G06F12/08
    • G06F12/0862G06F12/0897Y02D10/13
    • A processing unit for a multiprocessor data processing system includes a processor core and a cache hierarchy coupled to the processor core to provide low latency data access. The cache hierarchy includes an upper level cache coupled to the processor core and a lower level victim cache coupled to the upper level cache. In response to a prefetch request of the processor core that misses in the upper level cache, the lower level victim cache determines whether the prefetch request misses in the directory of the lower level victim cache and, if so, allocates a state machine in the lower level victim cache that services the prefetch request by issuing the prefetch request to at least one other processing unit of the multiprocessor data processing system.
    • 用于多处理器数据处理系统的处理单元包括处理器核心和耦合到处理器核心的高速缓存层级以提供低延迟数据访问。 高速缓存层级包括耦合到处理器核心的高级缓存和耦合到高级缓存的较低级别的牺牲缓存。 响应于在高级缓存中丢失的处理器核心的预取请求,较低级别的受害者缓存确定预取请求是否丢失在较低级别的受害者缓存的目录中,并且如果是,则在下级缓存中分配状态机 通过向多处理器数据处理系统的至少一个其他处理单元发出预取请求来服务于预取请求。