会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 4. 发明授权
    • Techniques for dynamically sharing a fabric to facilitate off-chip communication for multiple on-chip units
    • 用于动态共享结构以促进多个片上单元的片外通信的技术
    • US08346988B2
    • 2013-01-01
    • US12786716
    • 2010-05-25
    • Jian LiWilliam E. SpeightLixin Zhang
    • Jian LiWilliam E. SpeightLixin Zhang
    • G06F3/00G06F13/00
    • G06F1/3203G06F1/3253G06F9/5011Y02D10/151Y02D10/22
    • A technique for sharing a fabric to facilitate off-chip communication for on-chip units includes dynamically assigning a first unit that implements a first communication protocol to a first portion of the fabric when private fabrics are indicated for the on-chip units. The technique also includes dynamically assigning a second unit that implements a second communication protocol to a second portion of the fabric when the private fabrics are indicated for the on-chip units. In this case, the first and second units are integrated in a same chip and the first and second protocols are different. The technique further includes dynamically assigning, based on off-chip traffic requirements of the first and second units, the first unit or the second unit to the first and second portions of the fabric when the private fabrics are not indicated for the on-chip units.
    • 一种用于共享一个结构以促进片上单元的片外通信的技术包括:当针对片上单元指示专用结构时,动态分配实现第一通信协议的第一单元到该结构的第一部分。 该技术还包括当为片上单元指示专用结构时,动态地将实现第二通信协议的第二单元分配给该结构的第二部分。 在这种情况下,第一和第二单元集成在相同的芯片中,并且第一和第二协议是不同的。 该技术还包括:当私有结构未被指示用于片上单元时,基于第一单元或第二单元的片外流量要求将第一单元或第二单元动态地分配给该结构的第一和第二部分 。
    • 6. 发明授权
    • Hardware based dynamic load balancing of message passing interface tasks
    • 基于硬件的动态负载平衡消息传递接口任务
    • US08127300B2
    • 2012-02-28
    • US11846119
    • 2007-08-28
    • Lakshminarayana B. ArimilliRavi K. ArimilliRamakrishnan RajamonyWilliam E. Speight
    • Lakshminarayana B. ArimilliRavi K. ArimilliRamakrishnan RajamonyWilliam E. Speight
    • G06F9/46G06F15/173
    • G06F9/522G06F9/5083
    • Mechanisms for providing hardware based dynamic load balancing of message passing interface (MPI) tasks are provided. Mechanisms for adjusting the balance of processing workloads of the processors executing tasks of an MPI job are provided so as to minimize wait periods for waiting for all of the processors to call a synchronization operation. Each processor has an associated hardware implemented MPI load balancing controller. The MPI load balancing controller maintains a history that provides a profile of the tasks with regard to their calls to synchronization operations. From this information, it can be determined which processors should have their processing loads lightened and which processors are able to handle additional processing loads without significantly negatively affecting the overall operation of the parallel execution system. As a result, operations may be performed to shift workloads from the slowest processor to one or more of the faster processors.
    • 提供了提供消息传递接口(MPI)任务的基于硬件的动态负载平衡的机制。 提供了用于调整执行MPI作业任务的处理器的处理工作负载的平衡的机制,以便最小化等待所有处理器调用同步操作的等待时间。 每个处理器都有一个相关的硬件实现的MPI负载平衡控制器。 MPI负载平衡控制器维护一个历史记录,提供任务关于其对同步操作的调用的简档。 根据该信息,可以确定哪些处理器应该减轻其处理负载,哪些处理器能够处理额外的处理负载,而不会对并行执行系统的整体操作产生显着的负面影响。 因此,可以执行操作以将工作负载从最慢的处理器转移到一个或多个更快的处理器。
    • 9. 发明申请
    • Techniques for Indirect Data Prefetching
    • 间接数据预取技术
    • US20090198950A1
    • 2009-08-06
    • US12024239
    • 2008-02-01
    • Ravi K. ArimilliBalaram SinharoyWilliam E. SpeightLixin Zhang
    • Ravi K. ArimilliBalaram SinharoyWilliam E. SpeightLixin Zhang
    • G06F12/02G06F12/10
    • G06F12/0862G06F2212/6028
    • A processor includes a first address translation engine, a second address translation engine, and a prefetch engine. The first address translation engine is configured to determine a first memory address of a pointer associated with a data prefetch instruction. The prefetch engine is coupled to the first translation engine and is configured to fetch content, included in a first data block (e.g., a first cache line) of a memory, at the first memory address. The second address translation engine is coupled to the prefetch engine and is configured to determine a second memory address based on the content of the memory at the first memory address. The prefetch engine is also configured to fetch (e.g., from the memory or another memory) a second data block (e.g., a second cache line) that includes data at the second memory address.
    • 处理器包括第一地址转换引擎,第二地址转换引擎和预取引擎。 第一地址转换引擎被配置为确定与数据预取指令相关联的指针的第一存储器地址。 预取引擎被耦合到第一翻译引擎,并被配置为在第一存储器地址处提取包含在存储器的第一数据块(例如,第一高速缓存行)中的内容。 第二地址转换引擎耦合到预取引擎,并且被配置为基于第一存储器地址处的存储器的内容来确定第二存储器地址。 预取引擎还被配置为从第二存储器地址提取包括数据的第二数据块(例如,第二高速缓存行)(例如,从存储器或另一存储器)。