会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明申请
    • Mini-refresh processor recovery as bug workaround method using existing recovery hardware
    • 微型刷新处理器恢复作为使用现有恢复硬件的错误解决方法
    • US20060184771A1
    • 2006-08-17
    • US11055823
    • 2005-02-11
    • Michael FloydLarry LeitnerSheldon LevensteinScott SwaneyBrian Thompto
    • Michael FloydLarry LeitnerSheldon LevensteinScott SwaneyBrian Thompto
    • G06F9/30
    • G06F9/3863G06F9/3851
    • A method in a data processing system for avoiding a microprocessor's design defects and recovering a microprocessor from failing due to design defects, the method comprised of the following steps: The method detects and reports of events which warn of an error. Then the method locks a current checkpointed state and prevents instructions not checkpointed from checkpointing. After that, the method releases checkpointed state stores to a L2 cache, and drops stores not checkpointed. Next, the method blocks interrupts until recovery is completed. Then the method disables the power savings states throughout the processor. After that, the method disables an instruction fetch and an instruction dispatch. Next, the method sends a hardware reset signal. Then the method restores selected registers from the current checkpointed state. Next, the method fetches instructions from restored instruction addresses. Then the method resumes a normal execution after a programmable number of instructions.
    • 一种用于避免微处理器设计缺陷并由于设计缺陷而使微处理器故障恢复的数据处理系统中的方法,该方法包括以下步骤:该方法检测并报告发生错误的事件。 然后,该方法锁定当前的检查点状态,并防止从检查点进行检查点的指令。 之后,该方法将检查点状态存储发送到L2缓存,并且将不检查点丢弃存储。 接下来,该方法将阻止中断,直到恢复完成。 然后该方法将禁用整个处理器的省电状态。 之后,该方法禁用指令提取和指令分派。 接下来,该方法发送硬件复位信号。 然后,该方法将从当前检查点状态恢复所选寄存器。 接下来,该方法从恢复的指令地址获取指令。 然后,该方法在可编程指令数量之后恢复正常执行。
    • 2. 发明申请
    • Localized generation of global flush requests while guaranteeing forward progress of a processor
    • 本地化的全局刷新请求,同时保证处理器的进度
    • US20060184769A1
    • 2006-08-17
    • US11056692
    • 2005-02-11
    • Michael FloydHung LeLarry LeitnerBrian Thompto
    • Michael FloydHung LeLarry LeitnerBrian Thompto
    • G06F9/40
    • G06F9/3867G06F9/3814G06F9/3836G06F9/3859
    • Localized generation of global flush requests while providing a means for increasing the likelihood of forward progress in a controlled fashion. Local hazard (error) detection is accomplished with a trigger network situated between execution units and configurable state machines that track trigger events. Once a hazardous state is detected, a local detection mechanism requests a workaround flush from the flush control logic. The processor is flushed and a centralized workaround control is informed of the workaround flush. The centralized control blocks subsequent workaround flushes until forward progress has been made. The centralized control can also optionally send out a control to activate a set of localized workarounds or reduced performance modes to avoid the hazardous condition once instructions are re-executed after the flush until a configurable amount of forward progress has been made.
    • 本地化的全球冲刷请求,同时提供了一种增加以受控方式推进进展的可能性的手段。 本地危险(错误)检测是通过位于执行单元和跟踪触发事件的可配置状态机之间的触发网络完成的。 一旦检测到危险状态,本地检测机制就会从刷新控制逻辑请求解决冲突。 处理器被刷新,并且通知了一个集中的解决方法控制措施。 集中式控制块后续的解决方法会刷新,直到进行进一步的进展。 集中式控制还可以选择性地发送控制以激活一组本地化的解决方案或降低的性能模式,以避免在冲洗之后重新执行指令直到可配置的前进进度为止的危险状况。
    • 7. 发明申请
    • Branch lookahead prefetch for microprocessors
    • 用于微处理器的分支前瞻预取
    • US20080091928A1
    • 2008-04-17
    • US11953799
    • 2007-12-10
    • Richard EickemeyerHung LeDung NguyenBenjamin StoltBrian Thompto
    • Richard EickemeyerHung LeDung NguyenBenjamin StoltBrian Thompto
    • G06F9/38
    • G06F9/3842G06F9/3861
    • A method of handling program instructions in a microprocessor which reduces delays associated with mispredicted branch instructions, by detecting the occurrence of a stall condition during execution of the program instructions, speculatively executing one or more pending instructions which include at least one branch instruction during the stall condition, and determining the validity of data utilized by the speculative execution. Dispatch logic determines the validity of the data by marking one or more registers of an instruction dispatch unit to indicate which results of the pending instructions are invalid. The speculative execution of instructions can occur across multiple pipeline stages of the microprocessor, and the validity of the data is tracked during their execution in the multiple pipeline stages while monitoring a dependency of the speculatively executed instructions relative to one another during their execution in the multiple pipeline stages.
    • 一种处理微处理器中的程序指令的方法,其通过在执行程序指令期间检测到失速状态的发生来减少与错误预测的分支指令相关联的延迟,推测性地执行一个或多个未决指令,其中包括在失速期间包括至少一个分支指令 条件,并确定投机执行使用的数据的有效性。 调度逻辑通过标记指令调度单元的一个或多个寄存器来指示待处理指令的哪些结果无效来确定数据的有效性。 指令的推测执行可以在微处理器的多个流水线阶段发生,并且在多个流水线阶段的执行期间跟踪数据的有效性,同时在多个流水线阶段的执行期间监视推测性执行的指令相对于彼此的依赖性 流水线阶段
    • 8. 发明申请
    • Thread priority method, apparatus, and computer program product for ensuring processing fairness in simultaneous multi-threading microprocessors
    • 线程优先方法,装置和计算机程序产品,用于确保同时多线程微处理器的处理公平性
    • US20060184946A1
    • 2006-08-17
    • US11055850
    • 2005-02-11
    • James BishopHung LeDung NguyenBalaram SinharoyBrian ThomptoRaymond Yeung
    • James BishopHung LeDung NguyenBalaram SinharoyBrian ThomptoRaymond Yeung
    • G06F9/46
    • G06F9/30101G06F9/3834G06F9/3851G06F9/3861G06F9/3885G06F9/4818G06F9/485
    • A method, apparatus, and computer program product are disclosed in a data processing system for ensuring processing fairness in simultaneous multi-threading (SMT) microprocessors that concurrently execute multiple threads during each clock cycle. A clock cycle priority is assigned to a first thread and to a second thread during a standard selection state that lasts for an expected number of clock cycles. The clock cycle priority is assigned according to a standard selection definition during the standard selection state by selecting the first thread to be a primary thread and the second thread to be a secondary thread during the standard selection state. If a condition exists that requires overriding the standard selection definition, an override state is executed during which the standard selection definition is overridden by selecting the second thread to be the primary thread and the first thread to be the secondary thread. The override state is forced to be executed for an override period of time which equals the expected number of clock cycles plus a forced number of clock cycles. The forced number of clock cycles is granted to the first thread in response to the first thread again becoming the primary thread.
    • 在数据处理系统中公开了一种方法,装置和计算机程序产品,用于确保在每个时钟周期期间同时执行多个线程的同时多线程(SMT)微处理器中的处理公平性。 在持续预期数量的时钟周期的标准选择状态期间,将时钟周期优先级分配给第一线程和第二线程。 在标准选择状态期间,根据标准选择定义分配时钟周期优先级,通过在标准选择状态期间选择作为主线程的第一线程和第二线程作为次线程。 如果存在需要覆盖标准选择定义的条件,则执行超越状态,在该状态期间,通过选择第二个线程作为主线程,并将第一个线程作为次要线程来覆盖标准选择定义。 超时状态被强制执行超时时间等于预期的时钟周期数加上强制的时钟周期数。 响应于第一个线程再次成为主线程,强制的时钟周期数被授予第一个线程。
    • 9. 发明申请
    • Load lookahead prefetch for microprocessors
    • US20060149935A1
    • 2006-07-06
    • US11016236
    • 2004-12-17
    • Richard EickemeyerHung LeDung NguyenBenjamin StoltBrian Thompto
    • Richard EickemeyerHung LeDung NguyenBenjamin StoltBrian Thompto
    • G06F9/30
    • G06F9/3842G06F9/3804G06F9/383G06F9/3838G06F9/3851
    • The present invention allows a microprocessor to identify and speculatively execute future load instructions during a stall condition. This allows forward progress to be made through the instruction stream during the stall condition which would otherwise cause the microprocessor or thread of execution to be idle. The data for such future load instructions can be prefetched from a distant cache or main memory such that when the load instruction is re-executed (non speculative executed) after the stall condition expires, its data will reside either in the L1 cache, or will be enroute to the processor, resulting in a reduced execution latency. When an extended stall condition is detected, load lookahead prefetch is started allowing speculative execution of instructions that would normally have been stalled. In this speculative mode, instruction operands may be invalid due to source loads that miss the L1 cache, facilities not available in speculative execution mode, or due to speculative instruction results that are not available via forwarding and are not written to the architected registers. A set of status bits are used to dynamically keep track of the dependencies between instructions in the pipeline and a bit vector tracks invalid architected facilities with respect to the speculative instruction stream. Both sources of information are used to identify load instructions with valid operands for calculating the load address. If the operands are valid, then a load prefetch operation is started to retrieve data from the cache ahead of time such that it can be available for the load instruction when it is non-speculatively executed.
    • 10. 发明申请
    • Load Lookahead Prefetch for Microprocessors
    • 加载用于微处理器的前瞻预取
    • US20080077776A1
    • 2008-03-27
    • US11950495
    • 2007-12-05
    • Richard EickemeyerHung LeDung NguyenBenjamin StoltBrian Thompto
    • Richard EickemeyerHung LeDung NguyenBenjamin StoltBrian Thompto
    • G06F9/38
    • G06F9/3842G06F9/3804G06F9/383G06F9/3838G06F9/3851
    • The present invention allows a microprocessor to identify and speculatively execute future load instructions during a stall condition. This allows forward progress to be made through the instruction stream during the stall condition which would otherwise cause the microprocessor or thread of execution to be idle. The data for such future load instructions can be prefetched from a distant cache or main memory such that when the load instruction is re-executed (non speculative executed) after the stall condition expires, its data will reside either in the L1 cache, or will be enroute to the processor, resulting in a reduced execution latency. When an extended stall condition is detected, load lookahead prefetch is started allowing speculative execution of instructions that would normally have been stalled. In this speculative mode, instruction operands may be invalid due to source loads that miss the L1 cache, facilities not available in speculative execution mode, or due to speculative instruction results that are not available via forwarding and are not written to the architected registers. A set of status bits are used to dynamically keep track of the dependencies between instructions in the pipeline and a bit vector tracks invalid architected facilities with respect to the speculative instruction stream. Both sources of information are used to identify load instructions with valid operands for calculating the load address. If the operands are valid, then a load prefetch operation is started to retrieve data from the cache ahead of time such that it can be available for the load instruction when it is non-speculatively executed.
    • 本发明允许微处理器在失速状态期间识别并推测性地执行未来的加载指令。 这允许在停顿条件期间通过指令流进行正向进展,否则将导致微处理器或执行线程空闲。 可以从远程高速缓存或主存储器预取这样的未来加载指令的数据,使得当停止条件到期之后,当加载指令被重新执行(不推测执行)时,其数据将驻留在L1高速缓存中,或者将 进入处理器,导致执行时间缩短。 当检测到扩展失速条件时,启动加载前瞻预取,允许推测执行通常已经停止的指令。 在这种推测模式中,由于缺少L1高速缓存的源负载,设备在推测执行模式下不可用的设备,或由于不能通过转发而不能使用并且未写入到架构化寄存器的推测性指令结果,指令操作数可能无效。 一组状态位用于动态地跟踪流水线中的指令之间的依赖关系,并且位向量相对于推测性指令流跟踪无效的架构设施。 两个信息来源用于识别加载指令,其中包含用于计算加载地址的有效操作数。 如果操作数有效,则启动加载预取操作以提前从高速缓存中检索数据,使得当非推测性地执行加载指令时,可以对加载指令可用。