会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 11. 发明申请
    • VICTIM CACHE LATERAL CASTOUT TARGETING
    • US20100235577A1
    • 2010-09-16
    • US12340511
    • 2008-12-19
    • Guy L. GuthrieMichael S. SiegelWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieMichael S. SiegelWilliam J. StarkeDerek E. Williams
    • G06F12/08G06F12/00
    • G06F12/126G06F12/0811G06F12/0862
    • A data processing system includes a plurality of processing units coupled by an interconnect fabric. In response to a data request, a victim cache line is selected for castout from a first lower level cache of a first processing unit, and a target lower level cache of one of the plurality of processing units is selected based upon architectural proximity of the target lower level cache to a home system memory to which the address of the victim cache line is assigned. The first processing unit issues on the interconnect fabric a lateral castout (LCO) command that identifies the victim cache line to be castout from the first lower level cache and indicates that the target lower level cache is an intended destination. In response to a coherence response indicating success of the LCO command, the victim cache line is removed from the first lower level cache and held in the second lower level cache.
    • 数据处理系统包括通过互连结构耦合的多个处理单元。 响应于数据请求,从第一处理单元的第一较低级别高速缓存中选择牺牲缓存行来进行舍弃,并且基于目标的体系结构接近来选择多个处理单元之一的目标下级高速缓存 低级缓存到分配了受害者缓存行的地址的归属系统存储器。 所述第一处理单元在所述互连结构上发出侧向锁定(LCO)命令,所述侧向锁定(LCO)命令标识要从所述第一低级缓存中抛出的所述牺牲缓存线,并且指示所述目标低级缓存是预期目的地。 响应于指示LCO命令的成功的一致性响应,从第一低级缓存中删除受害者高速缓存行并保存在第二较低级高速缓存中。
    • 13. 发明申请
    • DATA PROCESSING SYSTEM, CACHE SYSTEM AND METHOD FOR UPDATING AN INVALID COHERENCY STATE IN RESPONSE TO SNOOPING AN OPERATION
    • 数据处理系统,缓存系统和用于更新无效级别的方法以响应于操作
    • US20080301377A1
    • 2008-12-04
    • US12190766
    • 2008-08-13
    • GUY L. GUTHRIEWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • GUY L. GUTHRIEWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • G06F12/08
    • G06F12/0831G06F2212/507
    • A cache coherent data processing system includes at least first and second coherency domains. In a first cache memory within the first coherency domain of the data processing system, a coherency state field associated with a storage location and an address tag is set to a first data-invalid coherency state that indicates that the address tag is valid and that the storage location does not contain valid data. In response to snooping an exclusive access operation, the exclusive access request specifying a target address matching the address tag and indicating a relative domain location of a requestor that initiated the exclusive access operation, the first cache memory updates the coherency state field from the first data-invalid coherency state to a second data-invalid coherency state that indicates that the address tag is valid, that the storage location does not contain valid data, and whether a target memory block associated with the address tag is cached within the first coherency domain upon successful completion of the exclusive access operation based upon the relative location of the requestor.
    • 缓存相干数据处理系统至少包括第一和第二相干域。 在数据处理系统的第一相关域内的第一高速缓冲存储器中,将与存储位置和地址标签相关联的一致性状态字段设置为指示地址标签有效的第一数据无效一致性状态, 存储位置不包含有效数据。 响应于窥探独占访问操作,所述专用访问请求指定与所述地址标签相匹配的目标地址并且指示发起所述独占访问操作的请求者的相对域位置,所述第一高速缓存存储器从所述第一数据更新所述一致性状态字段 - 无效的一致性状态到指示地址标签有效的第二数据无效一致性状态,存储位置不包含有效数据,以及与地址标签相关联的目标存储器块是否被缓存在第一相关域内 基于请求者的相对位置成功完成独占访问操作。
    • 14. 发明申请
    • REDUCING NUMBER OF REJECTED SNOOP REQUESTS BY EXTENDING TIME TO RESPOND TO SNOOP REQUEST
    • 通过延长时间减少违反SNOOP要求的数量以应对SNOOP要求
    • US20080201533A1
    • 2008-08-21
    • US12114790
    • 2008-05-04
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • G06F12/08
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. An incoming snoop request is entered in the first available latch in a pipeline of latches in a stall/reorder unit if the stall/reorder unit is not full. The entered snoop request is dispatched to a selector upon entering a bottom latch in the pipeline. The stall/reorder unit is not informed as to whether the dispatched snoop request is accepted by an arbitration mechanism for several clock cycles after the dispatch occurred. A copy of the dispatched snoop request is stored in a top latch in an overrun pipeline of latches in the first unit upon dispatching the snoop request. By maintaining information about the snoop request, the snoop request may be dispatched again to the selector in case the dispatched snoop request was rejected thereby increasing the chance that the snoop request will ultimately be accepted.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 如果失速/重新排序单元未满,则在失速/重新排序单元中的锁存器管线中的第一可用锁存器中输入进入的窥探请求。 输入的窥探请求在进入管道中的底部闩锁时被调度到选择器。 失速/重新排序单元不知道发送后发送的几个时钟周期是否由仲裁机制接受发送的窥探请求。 调度窥探请求的副本在调度窥探请求时被存储在第一单元中的锁存器的溢出管道中的顶部锁存器中。 通过维护关于窥探请求的信息,可以在被发送的窥探请求被拒绝的情况下再次发送到选择器的窥探请求,从而增加窥探请求将最终被接受的机会。
    • 15. 发明授权
    • Reducing number of rejected snoop requests by extending time to respond to snoop request
    • 通过延长响应窥探请求的时间来减少被拒绝的窥探请求数
    • US07386682B2
    • 2008-06-10
    • US11056764
    • 2005-02-11
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieHugh ShenWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. An incoming snoop request is entered in the first available latch in a pipeline of latches in a stall/reorder unit if the stall/reorder unit is not full. The entered snoop request is dispatched to a selector upon entering a bottom latch in the pipeline. The stall/reorder unit is not informed as to whether the dispatched snoop request is accepted by an arbitration mechanism for several clock cycles after the dispatch occurred. A copy of the dispatched snoop request is stored in a top latch in an overrun pipeline of latches in the first unit upon dispatching the snoop request. By maintaining information about the snoop request, the snoop request may be dispatched again to the selector in case the dispatched snoop request was rejected thereby increasing the chance that the snoop request will ultimately be accepted.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 如果失速/重新排序单元未满,则在失速/重新排序单元中的锁存器管线中的第一可用锁存器中输入进入的窥探请求。 输入的窥探请求在进入管道中的底部闩锁时被调度到选择器。 失速/重新排序单元不知道发送后发送的几个时钟周期是否由仲裁机制接受发送的窥探请求。 调度窥探请求的副本在调度窥探请求时被存储在第一单元中的锁存器的溢出管道中的顶部锁存器中。 通过维护关于窥探请求的信息,可以在被发送的窥探请求被拒绝的情况下再次发送到选择器的窥探请求,从而增加窥探请求将最终被接受的机会。
    • 17. 发明授权
    • Reducing number of rejected snoop requests by extending time to respond to snoop request
    • 通过延长响应窥探请求的时间来减少被拒绝的窥探请求数
    • US07340568B2
    • 2008-03-04
    • US11056740
    • 2005-02-11
    • Benjiman L. GoodmanGuy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • Benjiman L. GoodmanGuy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • G06F12/00
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. A “stall/reorder unit” in a cache receives a snoop request from an interconnect. Information, such as the address, of the snoop request is stored in a queue of the stall/reorder unit. The stall/reorder unit forwards the snoop request to a selector which also receives a request from a processor. An arbitration mechanism selects either the snoop request or the request from the processor. If the snoop request is denied by the arbitration mechanism, information, e.g., address, about the snoop request may be maintained in the stall/reorder unit. The request may be later resent to the selector. This process may be repeated up to “n” clock cycles. By providing the snoop request additional opportunities (n clock cycles) to be accepted by the arbitration mechanism, fewer snoop requests may ultimately be denied.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 缓存中的“停止/重新排序单元”从互连中接收窥探请求。 窥探请求的诸如地址的信息被存储在失速/重新排序单元的队列中。 停止/重新排序单元将窥探请求转发到也从处理器接收请求的选择器。 仲裁机制选择来自处理器的窥探请求或请求。 如果侦听请求被仲裁机制拒绝,关于窥探请求的信息(例如地址)可以被保留在停止/重新排序单元中。 请求可能会稍后重新发送到选择器。 该过程可以重复直到“n”个时钟周期。 通过提供窥探请求仲裁机制接受的额外机会(n个时钟周期),最终可能会拒绝更少的侦听请求。
    • 19. 发明授权
    • Cache-based speculation of stores following synchronizing operations
    • 同步操作后,存储器中基于缓存的推测
    • US08683140B2
    • 2014-03-25
    • US13456420
    • 2012-04-26
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0837G06F12/0895
    • A method of processing store requests in a data processing system includes enqueuing a store request in a store queue of a cache memory of the data processing system. The store request identifies a target memory block by a target address and specifies store data. While the store request and a barrier request older than the store request are enqueued in the store queue, a read-claim machine of the cache memory is dispatched to acquire coherence ownership of target memory block of the store request. After coherence ownership of the target memory block is acquired and the barrier request has been retired from the store queue, a cache array of the cache memory is updated with the store data.
    • 在数据处理系统中处理存储请求的方法包括在数据处理系统的高速缓存存储器的存储队列中引入存储请求。 存储请求通过目标地址识别目标存储器块并指定存储数据。 当存储请求和存储请求之前的屏障请求在存储队列中排队时,调度高速缓冲存储器的读取机器以获取存储请求的目标存储器块的一致性所有权。 在获取目标存储器块的一致性所有权并且屏障请求已经从存储队列中退出之后,用存储数据更新高速缓冲存储器的高速缓存阵列。