会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 2. 发明授权
    • PCI host bridge multi-priority fairness arbiter
    • PCI主机桥优先权公平仲裁器
    • US5905877A
    • 1999-05-18
    • US853776
    • 1997-05-09
    • Guy Lynn GuthrieKenneth Alan Riek
    • Guy Lynn GuthrieKenneth Alan Riek
    • G06F13/362G06F13/14
    • G06F13/362
    • A method and system for allowing one or more attached devices to access a computer bus. The objects of the method and system are achieved as is now described. At some particular instant in time, prioritized queues are loaded with one or more requests for access from one or more devices whose assigned priority levels correspond to the priority of the queue into which the requests for access are loaded. Requests for access, which are resident within a current queue, are preferentially granted in a sequential fashion until the current queue is emptied, after which at least one request for access from a lower in priority queue relative to the current queue is granted before responding to other requests for access, such that at least one request for access is periodically granted from a lower in priority queue relative to the current queue.
    • 一种用于允许一个或多个附接设备访问计算机总线的方法和系统。 方法和系统的对象如现在描述的那样实现。 在某个特定的时刻,优先级排队队列被加载有一个或多个请求,从一个或多个设备访问请求,其中一个或多个设备的分配的优先级与加载请求的队列的优先级相对应。 优先授予驻留在当前队列中的访问请求,直到当前队列被清空,然后在响应之前至少从一个较低优先级队列访问的请求相对于当前队列进行访问 其他访问请求,使得至少一个访问请求被定期地从相对于当前队列的较低优先级队列中被授予。
    • 5. 发明授权
    • Multi-node data processing system having a non-hierarchical interconnect architecture
    • 具有非分层互连架构的多节点数据处理系统
    • US06671712B1
    • 2003-12-30
    • US09436898
    • 1999-11-09
    • Ravi Kumar ArimilliJames Stephen Fields, Jr.Guy Lynn GuthrieJody Bern JoynerJerry Don Lewis
    • Ravi Kumar ArimilliJames Stephen Fields, Jr.Guy Lynn GuthrieJody Bern JoynerJerry Don Lewis
    • G06F1516
    • G06F13/4217
    • A data processing system includes a plurality of nodes, which each contain at least one agent, and data storage accessible to agents within the nodes. The plurality of nodes are coupled by a non-hierarchical interconnect including multiple non-blocking uni-directional address channels and at least one uni-directional data channel. The agents, which are each coupled to and snoop transactions on all of the plurality of address channels, can only issue transactions on an associated address channel. The uni-directional channels employed by the present non-hierarchical interconnect architecture permit high frequency pumped operation not possible with conventional bi-directional shared system buses. In addition, access latencies to remote (cache or main) memory incurred following local cache misses are greatly reduced as compared with conventional hierarchical systems because of the absence of inter-level (e.g., bus acquisition) communication latency. The non-hierarchical interconnect architecture also permits design flexibility in that the segment of the interconnect within each node can be independently implemented by a set of buses or as a switch, depending upon cost and performance considerations.
    • 数据处理系统包括多个节点,每个节点包含至少一个代理,以及节点内的代理可访问的数据存储。 多个节点通过包括多个非阻塞单向地址信道和至少一个单向数据信道的非分层互连来耦合。 在所有多个地址信道上分别耦合到并且窥探事务的代理只能在相关联的地址信道上发布事务。 当前的非分层互连架构采用的单向信道允许高频抽运操作对于传统的双向共享系统总线是不可能的。 另外,与传统分层系统相比,由于没有层间(例如,总线采集)通信延迟,与本地高速缓存未命中所产生的远程(高速缓存或主)存储器的访问延迟大大降低。 非分层互连架构还允许设计灵活性,因为根据成本和性能考虑,每个节点内的互连部分可以由一组总线或开关单独地实现。
    • 7. 发明授权
    • Multiprocessor system bus protocol with group addresses, responses, and priorities
    • 具有组地址,响应和优先级的多处理器系统总线协议
    • US06591321B1
    • 2003-07-08
    • US09437200
    • 1999-11-09
    • Ravi Kumar ArimilliJames Stephen Fields, Jr.Guy Lynn GuthrieJody Bern JoynerJerry Don Lewis
    • Ravi Kumar ArimilliJames Stephen Fields, Jr.Guy Lynn GuthrieJody Bern JoynerJerry Don Lewis
    • G06F1200
    • G06F12/0831
    • A multiprocessor system bus protocol system and method for processing and handling a processor request within a multiprocessor system having a number of bus accessible memory devices that are snooping on. at least one bus line. Snoop response groups which are groups of different types of snoop responses from the bus accessible memory devices are provided. Different transfer types are provided within each of the snoop response groups. A bus master device that provides a bus master signal is designated. The bus master device receives the processor request. One of the snoop response groups and one of the transfer types are appropriately designated based on the processor request. The bus master signal is formulated from a snoop response group, a transfer type, a valid request signal, and a cache line address. The bus master signal is sent to all of the bus accessible memory devices on the cache bus line and to a combined response logic system. All of the bus accessible memory devices on the cache bus line send snoop responses in response to the bus master signal based on the designated snoop response group. The snoop responses are sent to the combined response logic system. A combined response by the combined response logic system is determined based on the appropriate combined response encoding logic determined by the designated and latched snoop response group. The combined response is sent to all of the bus accessible memory devices on the cache bus line.
    • 一种用于处理和处理处理器请求的多处理器系统总线协议系统和方法,所述多处理器系统具有被窥探的多个总线可访问存储器件。 至少有一条总线。 提供了来自总线可访问存储器设备的不同类型的窥探响应的侦听响应组。 在每个窥探响应组中提供不同的传输类型。 指定提供总线主机信号的总线主设备。 总线主设备接收处理器请求。 根据处理器请求适当地指定其中一个侦听响应组和传输类型之一。 总线主机信号由侦听响应组,传输类型,有效请求信号和高速缓存线地址来制定。 总线主机信号被发送到高速缓存总线上的所有总线可访问存储器件和组合响应逻辑系统。 基于指定的窥探响应组,高速缓存总线上的所有总线可访问存储器件响应于总线主机信号发送窥探响应。 侦听响应被发送到组合的响应逻辑系统。 基于由指定和锁存的窥探响应组确定的适当的组合响应编码逻辑来确定组合响应逻辑系统的组合响应。 组合的响应被发送到高速缓存总线上的所有总线可访问存储器件。
    • 9. 发明授权
    • Multiprocessor system in which a cache serving as a highest point of coherency is indicated by a snoop response
    • 多处理器系统,其中作为最高点的一致性的缓存由窥探响应指示
    • US06405289B1
    • 2002-06-11
    • US09437196
    • 1999-11-09
    • Ravi Kumar ArimilliLeo James ClarkJames Stephen Fields, Jr.Guy Lynn Guthrie
    • Ravi Kumar ArimilliLeo James ClarkJames Stephen Fields, Jr.Guy Lynn Guthrie
    • G06F1200
    • G06F12/0831G06F12/0813G06F2212/2542
    • A method of maintaining cache coherency, by designating one cache that owns a line as a highest point of coherency (HPC) for a particular memory block, and sending a snoop response from the cache indicating that it is currently the HPC for the memory block and can service a request. The designation may be performed in response to a particular coherency state assigned to the cache line, or based on the setting of a coherency token bit for the cache line. The processing units may be grouped into clusters, while the memory is distributed using memory arrays associated with respective clusters. One memory array is designated as the lowest point of coherency (LPC) for the memory block (i.e., a fixed assignment) while the cache designated as the HPC is dynamic (i.e., changes as different caches gain ownership of the line). An acknowledgement snoop response is sent from the LPC memory array, and a combined response is returned to the requesting device which gives priority to the HPC snoop response over the LPC snoop response.
    • 通过将一个具有一行的高速缓存指定为特定存储器块的最高一致性(HPC),以及从高速缓存指示其当前是存储器块的HPC的高速缓存发送侦听响应的方法来维持高速缓存一致性的方法,以及 可以服务请求。 可以响应于分配给高速缓存行的特定一致性状态,或者基于高速缓存行的相关性令牌位的设置来执行指定。 处理单元可以被分组成群集,而存储器是使用与相应簇相关联的存储器阵列分布的。 一个存储器阵列被指定为存储器块的一致性(LPC)的最低点(即,固定分配),而指定为HPC的缓存是动态的(即,随着不同的高速缓存获得线的所有权而改变)。 从LPC存储器阵列发送确认窥探响应,并且将组合的响应返回给请求设备,该请求设备通过LPC窥探响应优先考虑HPC侦听响应。
    • 10. 发明授权
    • Optimized cache allocation algorithm for multiple speculative requests
    • 针对多个推测请求的优化缓存分配算法
    • US06393528B1
    • 2002-05-21
    • US09345714
    • 1999-06-30
    • Ravi Kumar ArimilliLakshminarayana Baba ArimilliLeo James ClarkJohn Steven DodsonGuy Lynn GuthrieJames Stephen Fields, Jr.
    • Ravi Kumar ArimilliLakshminarayana Baba ArimilliLeo James ClarkJohn Steven DodsonGuy Lynn GuthrieJames Stephen Fields, Jr.
    • G06F1200
    • G06F12/0862G06F12/127
    • A method of operating a computer system is disclosed in which an instruction having an explicit prefetch request is issued directly from an instruction sequence unit to a prefetch unit of a processing unit. In a preferred embodiment, two prefetch units are used, the first prefetch unit being hardware independent and dynamically monitoring one or more active streams associated with operations carried out by a core of the processing unit, and the second prefetch unit being aware of the lower level storage subsystem and sending with the prefetch request an indication that a prefetch value is to be loaded into a lower level cache of the processing unit. The invention may advantageously associate each prefetch request with a stream ID of an associated processor stream, or a processor ID of the requesting processing unit (the latter feature is particularly useful for caches which are shared by a processing unit cluster). If another prefetch value is requested from the memory hiearchy and it is determined that a prefetch limit of cache usage has been met by the cache, then a cache line in the cache containing one of the earlier prefetch values is allocated for receiving the other prefetch value.
    • 公开了一种操作计算机系统的方法,其中具有显式预取请求的指令直接从指令序列单元发送到处理单元的预取单元。 在优选实施例中,使用两个预取单元,第一预取单元是硬件独立的,并且动态地监视与由处理单元的核心执行的操作相关联的一个或多个活动流,并且第二预取单元知道较低级别 存储子系统,并用预取请求发送将预取值加载到处理单元的较低级缓存中的指示。 本发明可以有利地将每个预取请求与相关联的处理器流的流ID或请求处理单元的处理器ID相关联(后一特征对于由处理单元簇共享的高速缓存特别有用)。 如果从存储器hiearchy请求另一个预取值,并且确定高速缓存的高速缓存使用的预取限制已被满足,则包含先前预取值中的一个的高速缓存行中的高速缓存行被分配用于接收另一个预取值 。