会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 41. 发明授权
    • Protecting ownership transfer with non-uniform protection windows
    • 用不均匀的保护窗保护所有权转让
    • US08205024B2
    • 2012-06-19
    • US11560619
    • 2006-11-16
    • Leo J. ClarkJames S. Fields, Jr.Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Leo J. ClarkJames S. Fields, Jr.Guy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F3/00G06F13/00
    • G06F15/173
    • In a data processing system, a plurality of agents communicate operations therebetween. Each operation includes a request and a combined response representing a system-wide response to the request. Latencies of requests and combined responses between the plurality of agents are observed. Each of the plurality of agents is configured with a respective duration of a protection window extension by reference to the observed latencies. Each protection window extension is a period following receipt of a combined response during winch an associated one of the plurality of agents protects transfer of coherency ownership of a data granule between agents. The plurality of agents employing protection window extensions in accordance with the configuration, and at least two of the agents have protection window extensions of differing durations.
    • 在数据处理系统中,多个代理之间进行通信。 每个操作包括一个请求和组合的响应,代表对该请求的全系统响应。 观察到请求的延迟和多个代理之间的组合响应。 通过参考所观察到的延迟,多个代理中的每个被配置有保护窗口扩展的相应持续时间。 每个保护窗口扩展是在绞盘期间接收到组合响应之后的周期,多个代理之一相关联的一个代理保护代理之间的数据粒子的一致性所有权的传送。 多个代理根据配置​​使用保护窗口扩展,并且至少两个代理具有不同持续时间的保护窗口扩展。
    • 42. 发明授权
    • Virtual barrier synchronization cache castout election
    • 虚拟屏障同步缓存突发选举
    • US08095733B2
    • 2012-01-10
    • US12419343
    • 2009-04-07
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • G06F13/00G06F13/28
    • G06F12/0811G06F9/30101G06F9/3851G06F9/522
    • A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region. Each of the plurality of processing units includes a processor core and a cache memory including a cache controller and a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory. The cache controller of a first processing unit, responsive to a memory access request from its processor core that targets a first VBSR line, transfers responsibility for writing back to the virtual barrier synchronization region a second VBSR line contemporaneously held in the cache arrays of first, second and third processing units. The responsibility is transferred via an election held over the interconnect fabric.
    • 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括高速缓存控制器和从系统存储器的虚拟屏障同步区域缓存VBSR行的高速缓存阵列。 响应于来自其处理器核心的第一VBSR线路的存储器访问请求的第一处理单元的高速缓存控制器将负责向第一虚拟屏障同步区域写回同时保存在第一VBSR线路的高速缓存阵列中的第二VBSR线路, 第二和第三处理单元。 通过互连结构上的选举来转移责任。
    • 44. 发明申请
    • Aggregate Data Processing System Having Multiple Overlapping Synthetic Computers
    • 具有多重重合成计算机的综合数据处理系统
    • US20110153943A1
    • 2011-06-23
    • US12643800
    • 2009-12-21
    • Guy L. GuthrieCharles F. MarinoWilliam J. StarkeDerek E. Williams
    • Guy L. GuthrieCharles F. MarinoWilliam J. StarkeDerek E. Williams
    • G06F12/00G06F12/14G06F12/08
    • G06F12/0813G06F12/0284G06F15/167
    • A first SMP computer has first and second processing units and a first system memory pool, a second SMP computer has third and fourth processing units and a second system memory pool, and a third SMP computer has at least fifth and sixth processing units and third, fourth and fifth system memory pools. The fourth system memory pool is inaccessible to the third, fourth and sixth processing units and accessible to at least the second and fifth processing units, and the fifth system memory pool is inaccessible to the first, second and sixth processing units and accessible to at least the fourth and fifth processing units. A first interconnect couples the second processing unit for load-store coherent, ordered access to the fourth system memory pool, and a second interconnect couples the fourth processing unit for load-store coherent, ordered access to the fifth system memory pool.
    • 第一SMP计算机具有第一和第二处理单元和第一系统存储器池,第二SMP计算机具有第三和第四处理单元和第二系统存储器池,并且第三SMP计算机具有至少第五和第六处理单元,第三SMP计算机具有至少第五和第六处理单元, 第四和第五系统内存池。 第四系统存储器池对于第三,第四和第六处理单元是不可访问的,并且可访问至少第二和第五处理单元,并且第五系统存储器池对于第一,第二和第六处理单元是不可访问的,并且至少可访问 第四和第五处理单元。 第一互连耦合第二处理单元,用于对第四系统存储池进行加载存储相关的有序访问,并且第二互连耦合第四处理单元,用于加载存储相关的有序访问到第五系统存储池。
    • 45. 发明授权
    • Reducing number of rejected snoop requests by extending time to respond to snoop request
    • 通过延长响应窥探请求的时间来减少被拒绝的窥探请求数
    • US07818511B2
    • 2010-10-19
    • US11847941
    • 2007-08-30
    • Benjiman L. GoodmanGuy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • Benjiman L. GoodmanGuy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • G06F12/00
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. A “stall/reorder unit” in a cache receives a snoop request from an interconnect. Information, such as the address, of the snoop request is stored in a queue of the stall/reorder unit. The stall/reorder unit forwards the snoop request to a selector which also receives a request from a processor. An arbitration mechanism selects either the snoop request or the request from the processor. If the snoop request is denied by the arbitration mechanism, information, e.g., address, about the snoop request may be maintained in the stall/reorder unit. The request may be later resent to the selector. This process may be repeated up to “n” clock cycles. By providing the snoop request additional opportunities (n clock cycles) to be accepted by the arbitration mechanism, fewer snoop requests may ultimately be denied.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 缓存中的“停止/重新排序单元”从互连中接收窥探请求。 窥探请求的诸如地址的信息被存储在失速/重新排序单元的队列中。 停止/重新排序单元将窥探请求转发到也从处理器接收请求的选择器。 仲裁机制选择来自处理器的窥探请求或请求。 如果侦听请求被仲裁机制拒绝,关于窥探请求的信息(例如地址)可以被保留在停止/重新排序单元中。 请求可能会稍后重新发送到选择器。 该过程可以重复直到“n”个时钟周期。 通过提供窥探请求仲裁机制接受的额外机会(n个时钟周期),最终可能会拒绝更少的侦听请求。
    • 46. 发明申请
    • Virtual Barrier Synchronization Cache Castout Election
    • 虚拟障碍同步缓存铸造选举
    • US20100257316A1
    • 2010-10-07
    • US12419343
    • 2009-04-07
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • Ravi K. ArimilliGuy L. GuthrieMichael SiegelWilliam J. StarkeDerek E. Williams
    • G06F12/08G06F12/00
    • G06F12/0811G06F9/30101G06F9/3851G06F9/522
    • A data processing system includes an interconnect fabric, a system memory coupled to the interconnect fabric and including a virtual barrier synchronization region allocated to storage of virtual barrier synchronization registers (VBSRs), and a plurality of processing units coupled to the interconnect fabric and operable to access the virtual barrier synchronization region. Each of the plurality of processing units includes a processor core and a cache memory including a cache controller and a cache array that caches VBSR lines from the virtual barrier synchronization region of the system memory. The cache controller of a first processing unit, responsive to a memory access request from its processor core that targets a first VBSR line, transfers responsibility for writing back to the virtual barrier synchronization region a second VBSR line contemporaneously held in the cache arrays of first, second and third processing units. The responsibility is transferred via an election held over the interconnect fabric.
    • 数据处理系统包括互连结构,耦合到互连结构并包括分配给虚拟屏障同步寄存器(VBSR)的存储的虚拟屏障同步区域的系统存储器,以及耦合到互连结构的多个处理单元, 访问虚拟屏障同步区域。 多个处理单元中的每一个包括处理器核心和高速缓存存储器,其包括高速缓存控制器和从系统存储器的虚拟屏障同步区域缓存VBSR行的高速缓存阵列。 响应于来自其处理器核心的第一VBSR线路的存储器访问请求的第一处理单元的高速缓存控制器将负责向第一虚拟屏障同步区域写回同时保存在第一VBSR线路的高速缓存阵列中的第二VBSR线路, 第二和第三处理单元。 通过互连结构上的选举来转移责任。
    • 47. 发明授权
    • Reducing number of rejected snoop requests by extending time to respond to snoop request
    • 通过延长响应窥探请求的时间来减少被拒绝的窥探请求数
    • US07484046B2
    • 2009-01-27
    • US11950717
    • 2007-12-05
    • Benjiman L. GoodmanGuy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • Benjiman L. GoodmanGuy L. GuthrieWilliam J. StarkeJeffrey A. StuecheliDerek E. Williams
    • G06F12/00
    • G06F12/0831
    • A cache, system and method for reducing the number of rejected snoop requests. A “stall/reorder unit” in a cache receives a snoop request from an interconnect. Information, such as the address, of the snoop request is stored in a queue of the stall/reorder unit. The stall/reorder unit forwards the snoop request to a selector which also receives a request from a processor. An arbitration mechanism selects either the snoop request or the request from the processor. If the snoop request is denied by the arbitration mechanism, information, e.g., address, about the snoop request may be maintained in the stall/reorder unit. The request may be later resent to the selector. This process may be repeated up to “n” clock cycles. By providing the snoop request additional opportunities (n clock cycles) to be accepted by the arbitration mechanism, fewer snoop requests may ultimately be denied.
    • 用于减少拒绝的窥探请求数量的缓存,系统和方法。 缓存中的“停止/重新排序单元”从互连中接收窥探请求。 窥探请求的诸如地址的信息被存储在失速/重新排序单元的队列中。 停止/重新排序单元将窥探请求转发到也从处理器接收请求的选择器。 仲裁机制选择来自处理器的窥探请求或请求。 如果侦听请求被仲裁机制拒绝,关于窥探请求的信息(例如地址)可以被保留在停止/重新排序单元中。 请求可能会稍后重新发送到选择器。 该过程可以重复直到“n”个时钟周期。 通过提供窥探请求仲裁机制接受的额外机会(n个时钟周期),最终可能会拒绝更少的侦听请求。
    • 50. 发明申请
    • Data Processing System, Processor and Method of Data Processing in which Local Memory Access Requests are Serviced on a Fixed Schedule
    • 数据处理系统,处理器和数据处理方法,其中本地存储器访问请求在固定时间表上服务
    • US20080016278A1
    • 2008-01-17
    • US11457322
    • 2006-07-13
    • Leo J. ClarkGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • Leo J. ClarkGuy L. GuthrieWilliam J. StarkeDerek E. Williams
    • G06F12/00
    • G06F12/1425G06F12/0817G06F12/0897
    • A processing unit includes a local processor core and a cache memory coupled to the local processor core. The cache memory includes a data array, a directory of contents of the data array. The cache memory further includes one or more state machines that service a first set of memory access requests, an arbiter that directs servicing of a second set of memory access requests by reference to the data array and the directory on a fixed schedule, address collision logic that protects memory access requests in the second set by detecting and signaling address conflicts between active memory access requests in the second set and subsequent memory access requests, and dispatch logic coupled to the address collision logic. The dispatch logic dispatches memory access requests in the first set to the one or more state machines for servicing and signals the arbiter to direct servicing of memory access requests in the second set according to the fixed schedule.
    • 处理单元包括本地处理器核心和耦合到本地处理器核心的高速缓存存储器。 高速缓冲存储器包括数据阵列,数据阵列的内容目录。 缓存存储器还包括服务于第一组存储器访问请求的一个或多个状态机,通过参考数据阵列和固定时间表上的目录来指导第二组存储器访问请求的服务的仲裁器,地址冲突逻辑 其通过检测和发出第二组中的活动存储器访问请求与后续存储器访问请求之间的地址冲突以及耦合到地址冲突逻辑的调度逻辑来保护第二组中的存储器访问请求。 调度逻辑将第一组中的存储器访问请求分派到一个或多个状态机用于服务,并且向仲裁器发出信号,以根据固定的时间表对第二组中的存储器访问请求进行直接服务。