会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 21. 发明申请
    • REDUCING NUMBER OF REJECTED SNOOP REQUESTS BY EXTENDING TIME TO RESPOND TO SNOOP REQUEST
    • 通过延长时间减少违反SNOOP要求的数量以应对SNOOP要求
    • US20080201533A1
    • 2008-08-21
    • US12114790
    • 2008-05-04
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • G06F12/08
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. An incoming snoop request is entered in the first available latch in a pipeline of latches in a stall/reorder unit if the stall/reorder unit is not full. The entered snoop request is dispatched to a selector upon entering a bottom latch in the pipeline. The stall/reorder unit is not informed as to whether the dispatched snoop request is accepted by an arbitration mechanism for several clock cycles after the dispatch occurred. A copy of the dispatched snoop request is stored in a top latch in an overrun pipeline of latches in the first unit upon dispatching the snoop request. By maintaining information about the snoop request, the snoop request may be dispatched again to the selector in case the dispatched snoop request was rejected thereby increasing the chance that the snoop request will ultimately be accepted.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 如果失速/重新排序单元未满,则在失速/重新排序单元中的锁存器管线中的第一可用锁存器中输入进入的窥探请求。 输入的窥探请求在进入管道中的底部闩锁时被调度到选择器。 失速/重新排序单元不知道发送后发送的几个时钟周期是否由仲裁机制接受发送的窥探请求。 调度窥探请求的副本在调度窥探请求时被存储在第一单元中的锁存器的溢出管道中的顶部锁存器中。 通过维护关于窥探请求的信息,可以在被发送的窥探请求被拒绝的情况下再次发送到选择器的窥探请求,从而增加窥探请求将最终被接受的机会。
    • 22. 发明授权
    • Reducing number of rejected snoop requests by extending time to respond to snoop request
    • 通过延长响应窥探请求的时间来减少被拒绝的窥探请求数
    • US07386682B2
    • 2008-06-10
    • US11056764
    • 2005-02-11
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. An incoming snoop request is entered in the first available latch in a pipeline of latches in a stall/reorder unit if the stall/reorder unit is not full. The entered snoop request is dispatched to a selector upon entering a bottom latch in the pipeline. The stall/reorder unit is not informed as to whether the dispatched snoop request is accepted by an arbitration mechanism for several clock cycles after the dispatch occurred. A copy of the dispatched snoop request is stored in a top latch in an overrun pipeline of latches in the first unit upon dispatching the snoop request. By maintaining information about the snoop request, the snoop request may be dispatched again to the selector in case the dispatched snoop request was rejected thereby increasing the chance that the snoop request will ultimately be accepted.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 如果失速/重新排序单元未满,则在失速/重新排序单元中的锁存器管线中的第一可用锁存器中输入进入的窥探请求。 输入的窥探请求在进入管道中的底部闩锁时被调度到选择器。 失速/重新排序单元不知道发送后发送的几个时钟周期是否由仲裁机制接受发送的窥探请求。 调度窥探请求的副本在调度窥探请求时被存储在第一单元中的锁存器的溢出管道中的顶部锁存器中。 通过维护关于窥探请求的信息,可以在被发送的窥探请求被拒绝的情况下再次发送到选择器的窥探请求,从而增加窥探请求将最终被接受的机会。
    • 26. 发明授权
    • Victim cache lateral castout targeting
    • 受害者缓存横向倾斜定位
    • US08489819B2
    • 2013-07-16
    • US12340511
    • 2008-12-19
    • Guy L. GuthrieMichael S. SiegelWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieMichael S. SiegelWilliam J. StarkeDerek E. Williams
    • G06F12/00G06F13/00G06F13/28
    • G06F12/126G06F12/0811G06F12/0862
    • A data processing system includes a plurality of processing units coupled by an interconnect fabric. In response to a data request, a victim cache line is selected for castout from a first lower level cache of a first processing unit, and a target lower level cache of one of the plurality of processing units is selected based upon architectural proximity of the target lower level cache to a home system memory to which the address of the victim cache line is assigned. The first processing unit issues on the interconnect fabric a lateral castout (LCO) command that identifies the victim cache line to be castout from the first lower level cache and indicates that the target lower level cache is an intended destination. In response to a coherence response indicating success of the LCO command, the victim cache line is removed from the first lower level cache and held in the second lower level cache.
    • 数据处理系统包括通过互连结构耦合的多个处理单元。 响应于数据请求,从第一处理单元的第一较低级别高速缓存中选择牺牲缓存行来进行舍弃,并且基于目标的体系结构接近来选择多个处理单元之一的目标下级高速缓存 低级缓存到分配了受害者缓存行的地址的归属系统存储器。 所述第一处理单元在所述互连结构上发出侧向锁定(LCO)命令,所述侧向锁定(LCO)命令标识要从所述第一低级缓存中抛出的所述牺牲缓存线,并且指示所述目标低级缓存是预期目的地。 响应于指示LCO命令的成功的一致性响应,从第一低级缓存中删除受害者高速缓存行并保存在第二较低级高速缓存中。
    • 28. 发明授权
    • Aggregate data processing system having multiple overlapping synthetic computers
    • 具有多个重叠合成计算机的综合数据处理系统
    • US08370595B2
    • 2013-02-05
    • US12643800
    • 2009-12-21
    • Guy L. GuthrieCharles F. MarinoWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieCharles F. MarinoWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0813G06F12/0284G06F15/167
    • A first SMP computer has first and second processing units and a first system memory pool, a second SMP computer has third and fourth processing units and a second system memory pool, and a third SMP computer has at least fifth and sixth processing units and third, fourth and fifth system memory pools. The fourth system memory pool is inaccessible to the third, fourth and sixth processing units and accessible to at least the second and fifth processing units, and the fifth system memory pool is inaccessible to the first, second and sixth processing units and accessible to at least the fourth and fifth processing units. A first interconnect couples the second processing unit for load-store coherent, ordered access to the fourth system memory pool, and a second interconnect couples the fourth processing unit for load-store coherent, ordered access to the fifth system memory pool.
    • 第一SMP计算机具有第一和第二处理单元和第一系统存储器池,第二SMP计算机具有第三和第四处理单元和第二系统存储器池,并且第三SMP计算机具有至少第五和第六处理单元,第三SMP计算机具有至少第五和第六处理单元, 第四和第五系统内存池。 第四系统存储器池对于第三,第四和第六处理单元是不可访问的,并且可访问至少第二和第五处理单元,并且第五系统存储器池对于第一,第二和第六处理单元是不可访问的,并且至少可访问 第四和第五处理单元。 第一互连耦合第二处理单元,用于对第四系统存储池进行加载存储相关的有序访问,并且第二互连耦合第四处理单元,用于加载存储相关的有序访问到第五系统存储池。
    • 29. 发明授权
    • Partial cache line storage-modifying operation based upon a hint
    • 基于提示的部分缓存行存储修改操作
    • US08140771B2
    • 2012-03-20
    • US12024424
    • 2008-02-01
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/04G06F9/312
    • G06F12/0822
    • In at least one embodiment, a method of data processing in a data processing system having a memory hierarchy includes a processor core executing a storage-modifying memory access instruction to determine a memory address. The processor core transmits to a cache memory within the memory hierarchy a storage-modifying memory access request including the memory address, an indication of a memory access type, and, if present, a partial cache line hint signaling access to less than all granules of a target cache line of data associated with the memory address. In response to the storage-modifying memory access request, the cache memory performs a storage-modifying access to all granules of the target cache line of data if the partial cache line hint is not present and performs a storage-modifying access to less than all granules of the target cache line of data if the partial cache line hint is present.
    • 在至少一个实施例中,具有存储器层次的数据处理系统中的数据处理方法包括执行存储修改存储器访问指令以确定存储器地址的处理器核心。 处理器核心向存储器层级内的高速缓冲存储器传送存储修改存储器访问请求,该存储修改存储器访问请求包括存储器地址,存储器访问类型的指示,以及如果存在的话,部分高速缓存行提示信令访问少于所有颗粒的 与存储器地址相关联的数据的目标高速缓存行。 响应于存储修改存储器访问请求,如果不存在部分高速缓存行提示,则高速缓存存储器对目标高速缓存行数据行的所有颗粒进行存储修改访问,并执行对小于全部的存储修改访问 如果存在部分高速缓存线提示,则目标高速缓存行数据的颗粒。
    • 30. 发明授权
    • Bandwidth of a cache directory by slicing the cache directory into two smaller cache directories and replicating snooping logic for each sliced cache directory
    • 缓存目录的带宽通过将缓存目录分成两个较小的缓存目录,并为每个切片缓存目录复制侦听逻辑
    • US08135910B2
    • 2012-03-13
    • US11056721
    • 2005-02-11
    • Guy L. GuthrieWilliam J. StarkeDerek E. WilliamsPhillip G. Williams
    • Guy L. GuthrieWilliam J. StarkeDerek E. WilliamsPhillip G. Williams
    • G06F12/00
    • G06F12/0831G06F12/0851
    • A cache, system and method for improving the snoop bandwidth of a cache directory. A cache directory may be sliced into two smaller cache directories each with its own snooping logic. By having two cache directories that can be accessed simultaneously, the bandwidth can be essentially doubled. Furthermore, a “frequency matcher” may shift the cycle speed to a lower speed upon receiving snoop addresses from the interconnect thereby slowing down the rate at which requests are transmitted to the dispatch pipelines. Each dispatch pipeline is coupled to a sliced cache directory and is configured to search the cache directory to determine if data at the received addresses is stored in the cache memory. As a result of slowing down the rate at which requests are transmitted to the dispatch pipelines and accessing the two sliced cache directories simultaneously, the bandwidth or throughput of the cache directory may be improved.
    • 用于提高缓存目录的窥探带宽的缓存,系统和方法。 高速缓存目录可以分成两个较小的缓存目录,每个具有自己的侦听逻辑。 通过具有可以同时访问的两个缓存目录,带宽可以基本上加倍。 此外,当从互连接收到窥探地址时,“频率匹配器”可以将周期速度转移到较低速度,从而将请求发送到调度管线的速率减慢。 每个调度流水线被耦合到一个切片缓存目录,并被配置为搜索该高速缓存目录以确定该接收到的地址上的数据是否被存储在该高速缓冲存储器中。 作为将请求发送到调度管线的速度变慢并且同时访问两个分片缓存目录的结果,可以提高缓存目录的带宽或吞吐量。