会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 4. 发明授权
    • Selective cache-to-cache lateral castouts
    • 选择性高速缓存到缓存横向转义
    • US09189403B2
    • 2015-11-17
    • US12650018
    • 2009-12-30
    • Guy L. GuthrieWilliam J. StarkeJeffrey StuecheliDerek E. WilliamsThomas R. Puzak
    • Guy L. GuthrieWilliam J. StarkeJeffrey StuecheliDerek E. WilliamsThomas R. Puzak
    • G06F12/00G06F12/08G06F12/12
    • G06F12/0811G06F12/12
    • A data processing system includes first and second processing units and a system memory. The first processing unit has first upper and first lower level caches, and the second processing unit has second upper and lower level caches. In response to a data request, a victim cache line to be castout from the first lower level cache is selected, and the first lower level cache selects between performing a lateral castout (LCO) of the victim cache line to the second lower level cache and a castout of the victim cache line to the system memory based upon a confidence indicator associated with the victim cache line. In response to selecting an LCO, the first processing unit issues an LCO command on the interconnect fabric and removes the victim cache line from the first lower level cache, and the second lower level cache holds the victim cache line.
    • 数据处理系统包括第一和第二处理单元和系统存储器。 第一处理单元具有第一上层和第一下层高速缓存,第二处理单元具有第二上层和下层高速缓存。 响应于数据请求,选择要从第一较低级高速缓存丢弃的受害者高速缓存行,并且第一较低级高速缓存选择在执行到第二低级高速缓存的受害者高速缓存行的横向流出(LCO) 基于与受害者高速缓存行相关联的置信指示,将受害者缓存行的丢弃发送到系统存储器。 响应于选择LCO,第一处理单元在互连结构上发布LCO命令,并从第一低级缓存中移除受害者高速缓存行,并且第二下级缓存保存受害缓存行。
    • 5. 发明申请
    • SELECTIVE CACHE-TO-CACHE LATERAL CASTOUTS
    • 选择性高速缓存行驶路线
    • US20120203973A1
    • 2012-08-09
    • US13445646
    • 2012-04-12
    • Guy L. GuthrieWilliam J. StarkeJeffrey StuecheliDerek E. WilliamsThomas R. Puzak
    • Guy L. GuthrieWilliam J. StarkeJeffrey StuecheliDerek E. WilliamsThomas R. Puzak
    • G06F12/12
    • G06F12/0811G06F12/12
    • A data processing system includes first and second processing units and a system memory. The first processing unit has first upper and first lower level caches, and the second processing unit has second upper and lower level caches. In response to a data request, a victim cache line to be castout from the first lower level cache is selected, and the first lower level cache selects between performing a lateral castout (LCO) of the victim cache line to the second lower level cache and a castout of the victim cache line to the system memory based upon a confidence indicator associated with the victim cache line. In response to selecting an LCO, the first processing unit issues an LCO command on the interconnect fabric and removes the victim cache line from the first lower level cache, and the second lower level cache holds the victim cache line.
    • 数据处理系统包括第一和第二处理单元和系统存储器。 第一处理单元具有第一上层和第一下层高速缓存,第二处理单元具有第二上层和下层高速缓存。 响应于数据请求,选择要从第一较低级高速缓存丢弃的受害者高速缓存行,并且第一较低级高速缓存选择在执行到第二低级高速缓存的受害者高速缓存行的横向流出(LCO) 基于与受害者高速缓存行相关联的置信指示,将受害者缓存行的丢弃发送到系统存储器。 响应于选择LCO,第一处理单元在互连结构上发布LCO命令,并从第一低级缓存中移除受害者高速缓存行,并且第二下级缓存保存受害缓存行。
    • 7. 发明申请
    • PROCESSOR PERFORMANCE IMPROVEMENT FOR INSTRUCTION SEQUENCES THAT INCLUDE BARRIER INSTRUCTIONS
    • 包括障碍指示的指令序列的处理器性能改进
    • US20130205120A1
    • 2013-08-08
    • US13369029
    • 2012-02-08
    • Guy L GuthrieWilliam J. StarkeDerek E Williams
    • Guy L GuthrieWilliam J. StarkeDerek E Williams
    • G06F9/312
    • G06F9/52G06F9/30087G06F9/30145G06F9/3834G06F12/0831
    • A technique for processing an instruction sequence that includes a barrier instruction, a load instruction preceding the barrier instruction, and a subsequent memory access instruction following the barrier instruction includes determining that the load instruction is resolved based upon receipt of an earliest of a good combined response for a read operation corresponding to the load instruction and data for the load instruction. The technique also includes if execution of the subsequent memory access instruction is not initiated prior to completion of the barrier instruction, initiating in response to determining the barrier instruction completed, execution of the subsequent memory access instruction. The technique further includes if execution of the subsequent memory access instruction is initiated prior to completion of the barrier instruction, discontinuing in response to determining the barrier instruction completed, tracking of the subsequent memory access instruction with respect to invalidation.
    • 一种用于处理指示序列的技术,该指令序列包括屏障指令,屏障指令之前的加载指令,以及跟随障碍指令之后的随后存储器访问指令,包括:基于接收到最早的良好组合响应来确定加载指令是否被解决 用于与加载指令相对应的读取操作和用于加载指令的数据。 该技术还包括如果在完成屏障指令之前没有启动后续存储器访问指令的执行,则响应于确定完成的屏障指令启动后续存储器访问指令的执行。 该技术还包括如果在完成屏障指令之前启动后续存储器访问指令的执行,则响应于确定所完成的屏障指令而中断,跟踪关于无效的后续存储器访问指令。
    • 10. 发明申请
    • Virtual Barrier Synchronization Cache
    • 虚拟障碍同步缓存
    • US20100257317A1
    • 2010-10-07
    • US12419364
    • 2009-04-07
    • Ravi K. ArimilliGuy L. GuthrieRobert A. CargnoniWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieRobert A. CargnoniWilliam J. StarkeDerek E. Williams
    • G06F12/08G06F12/00
    • G06F12/0811G06F9/522
    • A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region of the system memory. Each of the plurality of processing units includes a processor core and a cache memory including a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory and a cache controller. The cache controller, responsive to a store request from the processor core to update a particular VBSR line, performs a non-blocking update of the cache array in each other of the plurality of processing units contemporaneously holding a copy of the particular VBSR line by transmitting a VBSR update command on the interconnect fabric.
    • 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问系统内存的虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括从系统存储器的虚拟屏障同步区域缓存VBSR行的缓存阵列和高速缓存控制器。 高速缓存控制器响应于来自处理器核心的存储请求来更新特定VBSR线路,通过发送来同时保存特定VBSR线路的副本的多个处理单元中的彼此之间的高速缓存阵列的非阻塞更新 互连结构上的VBSR更新命令。