会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 72. 发明授权
    • Cache-based speculation of stores following synchronizing operations
    • 同步操作后,存储器中基于缓存的推测
    • US08683140B2
    • 2014-03-25
    • US13456420
    • 2012-04-26
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/0837G06F12/0895
    • A method of processing store requests in a data processing system includes enqueuing a store request in a store queue of a cache memory of the data processing system. The store request identifies a target memory block by a target address and specifies store data. While the store request and a barrier request older than the store request are enqueued in the store queue, a read-claim machine of the cache memory is dispatched to acquire coherence ownership of target memory block of the store request. After coherence ownership of the target memory block is acquired and the barrier request has been retired from the store queue, a cache array of the cache memory is updated with the store data.
    • 在数据处理系统中处理存储请求的方法包括在数据处理系统的高速缓存存储器的存储队列中引入存储请求。 存储请求通过目标地址识别目标存储器块并指定存储数据。 当存储请求和存储请求之前的屏障请求在存储队列中排队时,调度高速缓冲存储器的读取机器以获取存储请求的目标存储器块的一致性所有权。 在获取目标存储器块的一致性所有权并且屏障请求已经从存储队列中退出之后,用存储数据更新高速缓冲存储器的高速缓存阵列。
    • 74. 发明授权
    • Load request scheduling in a cache hierarchy
    • 在缓存层次结构中加载请求调度
    • US08521982B2
    • 2013-08-27
    • US12424207
    • 2009-04-15
    • Robert A. CargnoniGuy L. GuthrieThomas L. JeremiahStephen J. PowellWilliam J. StarkeJeffrey A. Steucheli
    • Robert A. CargnoniGuy L. GuthrieThomas L. JeremiahStephen J. PowellWilliam J. StarkeJeffrey A. Steucheli
    • G06F12/00
    • G06F12/123G06F12/084G06F12/0897
    • A system and method for tracking core load requests and providing arbitration and ordering of requests. When a core interface unit (CIU) receives a load operation from the processor core, a new entry in allocated in a queue of the CIU. In response to allocating the new entry in the queue, the CIU detects contention between the load request and another memory access request. In response to detecting contention, the load request may be suspended until the contention is resolved. Received load requests may be stored in the queue and tracked using a least recently used (LRU) mechanism. The load request may then be processed when the load request resides in a least recently used entry in the load request queue. CIU may also suspend issuing an instruction unless a read claim (RC) machine is available. In another embodiment, CIU may issue stored load requests in a specific priority order.
    • 用于跟踪核心负载请求并提供仲裁和请求排序的系统和方法。 当核心接口单元(CIU)从处理器核心接收到加载操作时,分配在CIU队列中的新条目。 响应于在队列中分配新条目,CIU检测加载请求和另一个存储器访问请求之间的争用。 响应于检测到争用,负载请求可以被暂停,直到争用被解决。 接收到的加载请求可以存储在队列中,并使用最近最少使用的(LRU)机制进行跟踪。 然后可以在加载请求驻留在加载请求队列中最近最少使用的条目中时处理加载请求。 除非读取权利要求(RC)机器可用,否则CIU也可以暂停发出指令。 在另一个实施例中,CIU可以以特定优先级顺序发布存储的加载请求。
    • 75. 发明授权
    • Performing a partial cache line storage-modifying operation based upon a hint
    • 基于提示执行部分缓存行存储修改操作
    • US08332588B2
    • 2012-12-11
    • US13349315
    • 2012-01-12
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/04
    • G06F12/0822
    • Analyzing pre-processed code includes identifying at least one storage-modifying construct specifying a storage-modifying memory access to a memory hierarchy of a data processing system and determining if more than one granule of a cache line of data containing multiple granules that is targeted by the storage-modifying construct is subsequently referenced by said pre-processed code. Post-processed code including a storage-modifying instruction corresponding to the at least one storage-modifying construct in the pre-processed code is generated and stored. Generating the post-processed code includes marking the storage-modifying instruction with a partial cache line hint indicating that said storage-modifying instruction targets less than a full cache line of data within a memory hierarchy if the analyzing indicates only one granule of the target cache line will be accessed while the cache line is held in the cache memory and otherwise refraining from marking the storage-modifying instruction with the partial cache line hint.
    • 分析预处理的代码包括识别指定对数据处理系统的存储器层次结构的存储修改存储器访问的至少一个存储修改结构,并且确定是否存在多个颗粒的高速缓存行数据,所述数据包含多个颗粒的高速缓存行是由 存储修改结构随后由所述预处理代码引用。 生成并存储包括与预处理代码中的至少一个存储修改结构对应的存储修改指令的后处理代码。 生成后处理代码包括用部分高速缓存线提示标记存储修改指令,指示所述存储修改指令的目标小于存储器层次结构内的完整高速缓存数据行,如果分析仅指示目标高速缓存的一个颗粒 将高速缓存线保持在高速缓存存储器中,并以其它方式避免使用部分高速缓存线提示来标记存储修改指令。
    • 76. 发明申请
    • PERFORMING A PARTIAL CACHE LINE STORAGE-MODIFYING OPERATION BASED UPON A HINT
    • 根据提示执行部分缓存线存储 - 修改操作
    • US20120265938A1
    • 2012-10-18
    • US13349315
    • 2012-01-12
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/08
    • G06F12/0822
    • Analyzing pre-processed code includes identifying at least one storage-modifying construct specifying a storage-modifying memory access to a memory hierarchy of a data processing system and determining if more than one granule of a cache line of data containing multiple granules that is targeted by the storage-modifying construct is subsequently referenced by said pre-processed code. Post-processed code including a storage-modifying instruction corresponding to the at least one storage-modifying construct in the pre-processed code is generated and stored. Generating the post-processed code includes marking the storage-modifying instruction with a partial cache line hint indicating that said storage-modifying instruction targets less than a full cache line of data within a memory hierarchy if the analyzing indicates only one granule of the target cache line will be accessed while the cache line is held in the cache memory and otherwise refraining from marking the storage-modifying instruction with the partial cache line hint.
    • 分析预处理的代码包括识别指定对数据处理系统的存储器层次结构的存储修改存储器访问的至少一个存储修改结构,并且确定是否存在多个颗粒的高速缓存行数据,所述数据包含多个颗粒的高速缓存行是由 存储修改结构随后由所述预处理代码引用。 生成并存储包括与预处理代码中的至少一个存储修改结构对应的存储修改指令的后处理代码。 生成后处理代码包括用部分高速缓存线提示标记存储修改指令,指示所述存储修改指令的目标小于存储器层次结构内的完整高速缓存数据行,如果分析仅指示目标高速缓存的一个颗粒 将高速缓存线保持在高速缓存存储器中,并以其它方式避免使用部分高速缓存线提示来标记存储修改指令。
    • 78. 发明授权
    • Lateral cache-to-cache cast-in
    • 横向缓存到缓存投入
    • US08225045B2
    • 2012-07-17
    • US12335975
    • 2008-12-16
    • Guy L. GuthrieAlvan W. NgMichael S. SiegelWilliam J. StarkeDerek E. WilliamsPhillip G. Williams
    • Guy L. GuthrieAlvan W. NgMichael S. SiegelWilliam J. StarkeDerek E. WilliamsPhillip G. Williams
    • G06F12/00
    • G06F12/0811G06F12/0804G06F12/0831G06F12/0862
    • A data processing system includes a first processing unit and a second processing unit coupled by an interconnect fabric. The first processing unit has a first processor core and associated first upper and first lower level caches, and the second processing unit has a second processor core and associated second upper and lower level caches. In response to a data request, a victim cache line is selected for castout from the first lower level cache. The first processing unit issues on the interconnect fabric a lateral castout (LCO) command that identifies the victim cache line to be castout from the first lower level cache and indicates that a lower level cache is an intended destination. In response to a coherence response indicating success of the LCO command, the victim cache line is removed from the first lower level cache and held in the second lower level cache.
    • 数据处理系统包括由互连结构耦合的第一处理单元和第二处理单元。 第一处理单元具有第一处理器核心和相关联的第一上部和第一下层高速缓存,并且第二处理单元具有第二处理器核心和相关联的第二上部和下部高速缓存。 响应于数据请求,选择受害者高速缓存行用于从第一较低级别缓存进行舍弃。 第一处理单元在互连结构上发出横向聚合(LCO)命令,该命令标识要从第一较低级缓存中抛出的受害缓存行,并且指示较低级缓存是预期目的地。 响应于指示LCO命令的成功的一致性响应,从第一低级缓存中删除受害者高速缓存行并保存在第二较低级高速缓存中。
    • 79. 发明授权
    • Updating partial cache lines in a data processing system
    • 更新数据处理系统中的部分缓存行
    • US08117390B2
    • 2012-02-14
    • US12424434
    • 2009-04-15
    • David W. CummingsGuy L. GuthrieHugh ShenWilliam J. StarkeDerek E. WilliamsPhillip G. Williams
    • David W. CummingsGuy L. GuthrieHugh ShenWilliam J. StarkeDerek E. WilliamsPhillip G. Williams
    • G06F13/00
    • G06F12/0822G06F12/0897G06F2212/507
    • A processing unit for a data processing system includes a processor core having one or more execution units for processing instructions and a register file for storing data accessed in processing of the instructions. The processing unit also includes a multi-level cache hierarchy coupled to and supporting the processor core. The multi-level cache hierarchy includes at least one upper level of cache memory having a lower access latency and at least one lower level of cache memory having a higher access latency. The lower level of cache memory, responsive to receipt of a memory access request that hits only a partial cache line in the lower level cache memory, sources the partial cache line to the at least one upper level cache memory to service the memory access request. The at least one upper level cache memory services the memory access request without caching the partial cache line.
    • 用于数据处理系统的处理单元包括具有一个或多个用于处理指令的执行单元的处理器核心和用于存储在指令处理中访问的数据的寄存器文件。 处理单元还包括耦合到并支持处理器核的多级高速缓存层级。 多级高速缓存层级包括具有较低访问延迟的至少一个高级缓存存储器和具有较高访问延迟的至少一个较低级别的高速缓存存储器。 响应于仅接收低级高速缓冲存储器中的部分高速缓存行的存储器访问请求的响应,较低级别的高速缓存存储器将部分高速缓存行源送至至少一个上级高速缓冲存储器来服务存储器访问请求。 至少一个上级缓存存储器服务于存储器访问请求,而不缓存部分高速缓存行。