会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 51. 发明授权
    • Selective cache-to-cache lateral castouts
    • 选择性高速缓存到缓存横向转义
    • US09189403B2
    • 2015-11-17
    • US12650018
    • 2009-12-30
    • Guy L. GuthrieWilliam J. StarkeJeffrey StuecheliDerek E. WilliamsThomas R. Puzak
    • Guy L. GuthrieWilliam J. StarkeJeffrey StuecheliDerek E. WilliamsThomas R. Puzak
    • G06F12/00G06F12/08G06F12/12
    • G06F12/0811G06F12/12
    • A data processing system includes first and second processing units and a system memory. The first processing unit has first upper and first lower level caches, and the second processing unit has second upper and lower level caches. In response to a data request, a victim cache line to be castout from the first lower level cache is selected, and the first lower level cache selects between performing a lateral castout (LCO) of the victim cache line to the second lower level cache and a castout of the victim cache line to the system memory based upon a confidence indicator associated with the victim cache line. In response to selecting an LCO, the first processing unit issues an LCO command on the interconnect fabric and removes the victim cache line from the first lower level cache, and the second lower level cache holds the victim cache line.
    • 数据处理系统包括第一和第二处理单元和系统存储器。 第一处理单元具有第一上层和第一下层高速缓存,第二处理单元具有第二上层和下层高速缓存。 响应于数据请求,选择要从第一较低级高速缓存丢弃的受害者高速缓存行,并且第一较低级高速缓存选择在执行到第二低级高速缓存的受害者高速缓存行的横向流出(LCO) 基于与受害者高速缓存行相关联的置信指示,将受害者缓存行的丢弃发送到系统存储器。 响应于选择LCO,第一处理单元在互连结构上发布LCO命令,并从第一低级缓存中移除受害者高速缓存行,并且第二下级缓存保存受害缓存行。
    • 53. 发明授权
    • Aggregate data processing system having multiple overlapping synthetic computers
    • 具有多个重叠合成计算机的综合数据处理系统
    • US08656128B2
    • 2014-02-18
    • US13599856
    • 2012-08-30
    • Guy L GuthrieCharles F. MarinoWilliam J. StarkeDerek E. Williams
    • Guy L GuthrieCharles F. MarinoWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0813G06F12/0284G06F15/167
    • A first SMP computer has first and second processing units and a first system memory pool, a second SMP computer has third and fourth processing units and a second system memory pool, and a third SMP computer has at least fifth and sixth processing units and third, fourth and fifth system memory pools. The fourth system memory pool is inaccessible to the third, fourth and sixth processing units and accessible to at least the second and fifth processing units, and the fifth system memory pool is inaccessible to the first, second and sixth processing units and accessible to at least the fourth and fifth processing units. A first interconnect couples the second processing unit for load-store coherent, ordered access to the fourth system memory pool, and a second interconnect couples the fourth processing unit for load-store coherent, ordered access to the fifth system memory pool.
    • 第一SMP计算机具有第一和第二处理单元和第一系统存储器池,第二SMP计算机具有第三和第四处理单元和第二系统存储器池,并且第三SMP计算机具有至少第五和第六处理单元,第三SMP计算机具有至少第五和第六处理单元, 第四和第五系统内存池。 第四系统存储器池对于第三,第四和第六处理单元是不可访问的,并且可访问至少第二和第五处理单元,并且第五系统存储器池对于第一,第二和第六处理单元是不可访问的,并且至少可访问 第四和第五处理单元。 第一互连耦合第二处理单元,用于对第四系统存储池进行加载存储相关的有序访问,并且第二互连耦合第四处理单元,用于加载存储相关的有序访问到第五系统存储池。
    • 57. 发明授权
    • Synchronizing access to data in shared memory via upper level cache queuing
    • 通过高级缓存排队同步访问共享内存中的数据
    • US08327074B2
    • 2012-12-04
    • US13445080
    • 2012-04-12
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0811G06F9/3004G06F9/30072G06F9/30087G06F9/3834
    • A processing unit includes a store-in lower level cache having reservation logic that determines presence or absence of a reservation and a processor core including a store-through upper level cache, an instruction execution unit, a load unit that, responsive to a hit in the upper level cache on a load-reserve operation generated through execution of a load-reserve instruction by the instruction execution unit, temporarily buffers a load target address of the load-reserve operation, and a flag indicating that the load-reserve operation bound to a value in the upper level cache. If a storage-modifying operation is received that conflicts with the load target address of the load-reserve operation, the processor core sets the flag to a particular state, and, responsive to execution of a store-conditional instruction, transmits an associated store-conditional operation to the lower level cache with a fail indication if the flag is set to the particular state.
    • 处理单元包括具有确定存在或不存在预留的预约逻辑的存储下位缓存和包括存储通过上级缓存,指令执行单元,负载单元的处理器核心,该负载单元响应于 由指令执行单元通过执行装载预约指令而产生的加载备用操作的上级缓存暂时缓冲加载备用操作的加载目标地址,以及指示载入预约操作被绑定到 上级缓存中的值。 如果接收到与加载保留操作的加载目标地址冲突的存储修改操作,则处理器核心将该标志设置为特定状态,并且响应于执行存储条件指令,发送关联的存储 - 如果该标志被设置为特定状态,则向低级缓存进行条件操作,并显示故障指示。
    • 58. 发明申请
    • SYNCHRONIZING ACCESS TO DATA IN SHARED MEMORY VIA UPPER LEVEL CACHE QUEUING
    • 通过上级缓存队列同步访问共享存储器中的数据
    • US20120198167A1
    • 2012-08-02
    • US13445080
    • 2012-04-12
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/08
    • G06F12/0811G06F9/3004G06F9/30072G06F9/30087G06F9/3834
    • A processing unit includes a store-in lower level cache having reservation logic that determines presence or absence of a reservation and a processor core including a store-through upper level cache, an instruction execution unit, a load unit that, responsive to a hit in the upper level cache on a load-reserve operation generated through execution of a load-reserve instruction by the instruction execution unit, temporarily buffers a load target address of the load-reserve operation, and a flag indicating that the load-reserve operation bound to a value in the upper level cache. If a storage-modifying operation is received that conflicts with the load target address of the load-reserve operation, the processor core sets the flag to a particular state, and, responsive to execution of a store-conditional instruction, transmits an associated store-conditional operation to the lower level cache with a fail indication if the flag is set to the particular state.
    • 处理单元包括具有确定存在或不存在预留的预约逻辑的存储下位缓存和包括存储通过上级缓存,指令执行单元,负载单元的处理器核心,该负载单元响应于 由指令执行单元通过执行装载预约指令而产生的加载备用操作的上级缓存暂时缓冲加载备用操作的加载目标地址,以及指示载入预约操作被绑定到 上级缓存中的值。 如果接收到与加载保留操作的加载目标地址冲突的存储修改操作,则处理器核心将该标志设置为特定状态,并且响应于执行存储条件指令,发送关联的存储 - 如果该标志被设置为特定状态,则向低级缓存进行条件操作,并显示故障指示。
    • 60. 发明授权
    • Virtual barrier synchronization cache
    • 虚拟障碍同步缓存
    • US08131935B2
    • 2012-03-06
    • US12419364
    • 2009-04-07
    • Ravi K. ArimilliGuy L. GuthrieRobert A. CargnoniWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieRobert A. CargnoniWilliam J. StarkeDerek E. Williams
    • G06F13/00G06F13/28
    • G06F12/0811G06F9/522
    • A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region of the system memory. Each of the plurality of processing units includes a processor core and a cache memory including a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory and a cache controller. The cache controller, responsive to a store request from the processor core to update a particular VBSR line, performs a non-blocking update of the cache array in each other of the plurality of processing units contemporaneously holding a copy of the particular VBSR line by transmitting a VBSR update command on the interconnect fabric.
    • 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问系统内存的虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括从系统存储器的虚拟屏障同步区域缓存VBSR行的缓存阵列和高速缓存控制器。 高速缓存控制器响应于来自处理器核心的存储请求来更新特定VBSR线路,通过发送来同时保存特定VBSR线路的副本的多个处理单元中的彼此之间的高速缓存阵列的非阻塞更新 互连结构上的VBSR更新命令。