会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 3. 发明授权
    • Optimization of ordered stores on a pipelined bus via self-initiated retry
    • 通过自发重试优化流水线总线上的有序存储
    • US06269360B1
    • 2001-07-31
    • US09066012
    • 1998-04-24
    • Thomas Albert PetersenJames Nolan Hardage, Jr.
    • Thomas Albert PetersenJames Nolan Hardage, Jr.
    • G06F1730
    • G06F13/161Y10S707/99932
    • Where a plurality of ordered transactions are received for data transfers on a pipelined bus, each transaction in the series is initiated before all prospective retry responses to the preceding ordered transactions may be asserted. The address responses to all preceding ordered transfers are then monitored in connection with performance of the newly initiated transfer. If a retry response to any preceding ordered transaction is asserted, a self-initiated retry response for all subsequent transactions, including the newly initiated transfer, is also asserted. The system-retried transactions and all succeeding, ordered transactions are immediately reattempted. The overlapping performance of the ordered transfers reduces the latency of non-retried transfers, achieving performance comparable to non-ordered transactions. Even where a retry response is asserted, the total latency required for completion of both transactions in the ordered pair is reduced by at least a portion of the address-to-response latency, so that the impact of ordering requirements on system performance is minimized. Strict ordering is thus enforced while taking full advantage of the pipelined nature of the bus to maximize utilization of the bus bandwidth.
    • 在对流水线总线上的数据传输接收到多个有序事务的情况下,系列中的每个事务在所有对先前有序事务的预期重试响应可能被断言之前启动。 然后,与新发起的传输的性能相关联地监视对所有先前有序传输的地址响应。 如果对任何先前有序事务的重试响应被断言,则还会断言所有后续事务的自发起的重试响应,包括新发起的传输。 系统重试的交易和所有后续的有序交易立即被重新尝试。 有序传输的重叠性能降低了非重试传输的延迟,实现与非有序事务相当的性能。 即使在重试响应被断言的情况下,完成有序对中的两个事务所需的总延迟也减少了地址到响应延迟的至少一部分,从而将排序要求对系统性能的影响降到最低。 因此,在充分利用总线的流水线特性的同时强制执行严格排序,以最大限度地利用总线带宽。
    • 4. 发明授权
    • Multi-stage pipelined data coalescing for improved frequency operation
    • 多级流水线数据合并,提高频率运行
    • US6023737A
    • 2000-02-08
    • US66014
    • 1998-04-24
    • Thomas Albert PetersenJames Nolan Hardage, Jr.
    • Thomas Albert PetersenJames Nolan Hardage, Jr.
    • G06F12/08G06F12/04G06F13/36G06F13/00
    • G06F13/36
    • To implement full gathering of data transfers from a processor to a system bus without adding many levels of logic to the write enable logic for transaction queue entries or reducing the processor operating frequency, gatherable combinations are divided and gathering is performed in multiple stages operating in parallel. During the first stage, a subset of the full gathering is performed between incoming transactions and the last transaction received, coalescing the two transfers into a single transaction entry if one of the possible combinations within the subset is satisfied. During the second stage, existing queue entries are tested for the remainder of the full gather combination set and merged if a combination within the remaining subset is satisfied. The gathering logic may thus be split between the write enable logic and the entry control logic for the transaction queue, reducing the depth of logic required for any one path and increasing the set size of gatherable combinations implemented and/or the processor operating frequency. Any additional processor cycles required to complete full gathering are typically hidden by bus latency.
    • 为了实现从处理器到系统总线的数据传输的全面收集,而不会为事务队列条目的写使能逻辑增加许多级别的逻辑,或降低处理器的工作频率,可收集的组合被分割并且以并行运行的多个阶段进行采集 。 在第一阶段期间,在传入事务和所接收的最后事务之间执行完整收集的子集,如果满足子集内的可能组合之一,则将两个传输合并成单个事务条目。 在第二阶段期间,如果剩余子集中的组合得到满足,则现有队列条目将针对完整集合组合集合的其余部分进行测试并合并。 因此,采集逻辑可以在写入使能逻辑和用于事务队列的条目控制逻辑之间分开,从而减少任何一个路径所需的逻辑深度并增加实现的可收集组合和/或处理器工作频率的集合大小。 总线延迟通常隐藏完成完全采集所需的任何其他处理器周期。
    • 5. 发明授权
    • Data processing system and method for maintaining translation lookaside
buffer TLB coherency without enforcing complete instruction
serialization
    • 用于保持翻译后备缓冲器TLB一致性的数据处理系统和方法,而不强制执行完整的指令串行化
    • US06119204A
    • 2000-09-12
    • US108157
    • 1998-06-30
    • Joseph Yih ChangJames Nolan Hardage, Jr.Jose Melanio NunezThomas Albert Petersen
    • Joseph Yih ChangJames Nolan Hardage, Jr.Jose Melanio NunezThomas Albert Petersen
    • G06F12/10G06F13/14G06F12/08
    • G06F12/1027G06F2212/682
    • A data processing system includes at least a first processor and a second processor that each have a respective translation lookaside buffer (TLB). In response to detection by the second processor of a TLB entry invalidation request, the second processor marks at least one memory referent instruction that is being processed by the second processor and invalidates a TLB entry in the TLB of the second processor. In response to receipt of a synchronization request at the second processor, the second processor indicates to the first processor that the second processor has invalidated the TLB entry if the second processor has completed processing the marked instruction. During the interval between receipt of the synchronization request and indicating to the first processor that the second processor has invalidated the TLB entry, the second processor continues to process instructions, including fetching instructions for processing. In this manner, the second processor is able to continue normal instruction processing during the process of TLB synchronization.
    • 数据处理系统至少包括第一处理器和第二处理器,每个具有相应的翻译后备缓冲器(TLB)。 响应于第二处理器检测到TLB条目无效请求,第二处理器标记正在由第二处理器处理的至少一个存储器指示指令,并使第二处理器的TLB中的TLB条目无效。 响应于在第二处理器处接收到同步请求,第二处理器向第一处理器指示如果第二处理器已经完成处理标记的指令,则第二处理器使TLB条目无效。 在接收到同步请求和向第一处理器指示第二处理器使TLB条目无效的间隔期间,第二处理器继续处理指令,包括取指令进行处理。 以这种方式,第二处理器能够在TLB同步的处理期间继续正常的指令处理。
    • 6. 发明授权
    • Data received before coherency window for a snoopy bus
    • 在Snoopy总线的一致性窗口之前收到的数据
    • US06898675B1
    • 2005-05-24
    • US09138380
    • 1998-08-24
    • Alexander Edward OkpiszThomas Albert Petersen
    • Alexander Edward OkpiszThomas Albert Petersen
    • G06F12/00G06F12/08
    • G06F12/0831
    • Where a null response can be expected from devices snooping a load operation, data may be used by a requesting processor prior to the coherency response window. A null snoop response may be determined, for example, from the availability of the data without a bus transaction. The capability of accelerating data in this fashion requires only a few simple changes in processor state transitions, required to permit entry of the data completion wait state prior to the response wait state. Processors may forward accelerated data to execution units with the expectation that a null snoop response will be received during the coherency response window. If a non-null snoop response is received, an error condition is asserted. Data acceleration of the type described allows critical data to get back to the processor without waiting for the coherency response window.
    • 在设备侦听加载操作时可以期望空响应的情况下,请求处理器可以在相干性响应窗口之前使用数据。 可以例如从没有总线事务的数据的可用性来确定空窥探响应。 以这种方式加速数据的能力仅需要处理器状态转换的几个简单的改变,这是允许在响应等待状态之前输入数据完成等待状态所需要的。 处理器可以将加速数据转发到执行单元,期望在一致性响应窗口期间将接收到空窥探响应。 如果接收到非空窥探响应,则会发出错误条件。 所描述类型的数据加速允许关键数据回到处理器,而不必等待一致性响应窗口。
    • 7. 发明授权
    • Static queue and index queue for storing values identifying static queue locations
    • 静态队列和索引队列,用于存储识别静态队列位置的值
    • US06317806B1
    • 2001-11-13
    • US09315612
    • 1999-05-20
    • Srinath AudityanThomas Albert PetersenRobert Charles Podnar
    • Srinath AudityanThomas Albert PetersenRobert Charles Podnar
    • G06F1200
    • G06F5/06G06F12/0855Y02D10/13
    • A queuing apparatus associated with a processor includes at least one static queue (11), an index generator (34), at least one index queue (37), and a static queue accessing arrangement. Each static queue (11) has a number of static queue locations (12), each for storing a static queue entry and an availability indicator (14) for indicating an availability status of the respective static queue location. The index generator (34) uses information from the static queue (11) to provide a unique index value for each static queue entry, the index value for a particular static queue entry identifying the static queue location (12) containing the particular static queue entry. Each index queue (37, 42) has a number of index queue locations (40), each for storing one of the index values provided by the index generator (34). The static queue accessing arrangement retrieves a selected index value from a particular index queue location (40), and uses the selected index value to retrieve the static queue entry with which the selected index value is associated. Multiple index queues (37, 42) facilitate prioritization of static queue entries, and reprioritization by transferring index queue values from one index queue to another. The index queues (37, 42) also facilitate compaction of unnecessary static queue entries.
    • 与处理器相关联的排队设备包括至少一个静态队列(11),索引生成器(34),至少一个索引队列(37)和静态队列访问布置。 每个静态队列(11)具有多个静态队列位置(12),每个静态队列位置(12)用于存储静态队列条目和用于指示相应静态队列位置的可用性状态的可用性指示符(14)。 索引生成器(34)使用来自静态队列(11)的信息为每个静态队列条目提供唯一的索引值,特定静态队列条目的索引值标识包含特定静态队列条目的静态队列位置(12) 。 每个索引队列(37,42)具有多个索引队列位置(40),每个索引队列位置(40)用于存储由索引生成器(34)提供的索引值之一。 静态队列访问安排从特定索引队列位置(40)检索所选择的索引值,并使用所选择的索引值来检索与所选择的索引值相关联的静态队列条目。 多个索引队列(37,42)有助于静态队列条目的优先化,并通过将索引队列值从一个索引队列传递到另一个索引队列来重新确定优先级。 索引队列(37,42)还有助于压缩不必要的静态队列条目。
    • 10. 发明授权
    • Queue resource tracking in a multiprocessor system
    • 多处理器系统中的队列资源跟踪
    • US06460133B1
    • 2002-10-01
    • US09315488
    • 1999-05-20
    • Jose Melanio NunezThomas Albert Petersen
    • Jose Melanio NunezThomas Albert Petersen
    • G06F900
    • G06F13/1642G06F9/3824G06F9/3836G06F12/0831
    • A multiprocessor computer system including a set of processors where each processor in the set includes an execution unit for issuing operations and a processor queue suitable for queuing previously issued and still pending operations. The multiprocessor further includes means for forwarding operations issued by the processor to the processor queue and to an operation block queue of a memory subsystem that is connected to the multiprocessor. The depth of (i.e., the number of entries in) the operation block queue matches the depth of the processor queue. The processor queue, when full, inhibits the processor from issuing additional operations. In this manner, an operation issued by the processor is guaranteed an available entry in the operation block queue of the memory subsystem thereby eliminating the need for operation retry circuitry and protocols such as handshaking. Preferably, each processor queue includes a processor load queue and a processor store queue and the operation block queue includes a load queue and a store queue. In this embodiment, the depth of each of the processor load and store queues matches the depth of the operation block load and store queues respectively. In the preferred embodiment, the operation block is comprised of a load miss block that includes the operation block load queue and a store miss block that includes the operation block store queue. Still further preferably, the operation block store queue includes a set of store queues corresponding to the set of processors and the operation block load queue includes a set of load queues corresponding to the set of processors. Each queue entry preferably includes state information indicative of the status of the corresponding entry.
    • 一种包括一组处理器的多处理器计算机系统,其中集合中的每个处理器包括用于发布操作的执行单元和适于排队先前发布和尚待处理的操作的处理器队列。 多处理器还包括用于将由处理器发出的操作转发到处理器队列和连接到多处理器的存储器子系统的操作块队列的装置。 操作块队列的深度(即条目数)与处理器队列的深度相匹配。 处理器队列在满时禁止处理器发出其他操作。 以这种方式,由处理器发出的操作保证在存储器子系统的操作块队列中的可用条目,从而消除对操作重试电路和协议(例如握手)的需要。 优选地,每个处理器队列包括处理器负载队列和处理器存储队列,并且操作块队列包括加载队列和存储队列。 在该实施例中,每个处理器加载和存储队列的深度分别与操作块加载和存储队列的深度匹配。 在优选实施例中,操作块包括包括操作块加载队列的负载未命中块和包括操作块存储队列的存储错误块。 更优选地,操作块存储队列包括与该组处理器对应的一组存储队列,并且操作块加载队列包括与该组处理器对应的一组加载队列。 每个队列条目优选地包括指示对应条目的状态的状态信息。