会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 5. 发明授权
    • Coalescing memory barrier operations across multiple parallel threads
    • 在多个并行线程之间合并记忆障碍操作
    • US09223578B2
    • 2015-12-29
    • US12887081
    • 2010-09-21
    • John R. NickollsSteven James HeinrichBrett W. CoonMichael C. Shebanow
    • John R. NickollsSteven James HeinrichBrett W. CoonMichael C. Shebanow
    • G06F9/46G06F9/38G06F9/30
    • G06F9/3834G06F9/3004G06F9/30087G06F9/3851
    • One embodiment of the present invention sets forth a technique for coalescing memory barrier operations across multiple parallel threads. Memory barrier requests from a given parallel thread processing unit are coalesced to reduce the impact to the rest of the system. Additionally, memory barrier requests may specify a level of a set of threads with respect to which the memory transactions are committed. For example, a first type of memory barrier instruction may commit the memory transactions to a level of a set of cooperating threads that share an L1 (level one) cache. A second type of memory barrier instruction may commit the memory transactions to a level of a set of threads sharing a global memory. Finally, a third type of memory barrier instruction may commit the memory transactions to a system level of all threads sharing all system memories. The latency required to execute the memory barrier instruction varies based on the type of memory barrier instruction.
    • 本发明的一个实施例提出了一种用于在多个并行线程之间聚合存储器屏障操作的技术。 来自给定并行线程处理单元的存储器屏障请求被合并以减少对系统其余部分的影响。 此外,存储器屏障请求可以指定针对其提交内存事务的一组线程的级别。 例如,第一类型的存储器障碍指令可以将存储器事务提交到共享L1(一级)高速缓存的一组协作线程的级别。 第二种类型的存储器障碍指令可以将存储器事务提交到共享全局存储器的一组线程的级别。 最后,第三种类型的存储器障碍指令可以将存储器事务提交到共享所有系统存储器的所有线程的系统级。 执行存储器屏障指令所需的延迟基于存储器屏障指令的类型而变化。
    • 6. 发明申请
    • COALESCING MEMORY BARRIER OPERATIONS ACROSS MULTIPLE PARALLEL THREADS
    • 通过多个并行线程来解决存储器障碍操作
    • US20110078692A1
    • 2011-03-31
    • US12887081
    • 2010-09-21
    • John R. NICKOLLSSteven James HeinrichBrett W. CoonMichael C. Shebanow
    • John R. NICKOLLSSteven James HeinrichBrett W. CoonMichael C. Shebanow
    • G06F9/46
    • G06F9/3834G06F9/3004G06F9/30087G06F9/3851
    • One embodiment of the present invention sets forth a technique for coalescing memory barrier operations across multiple parallel threads. Memory barrier requests from a given parallel thread processing unit are coalesced to reduce the impact to the rest of the system. Additionally, memory barrier requests may specify a level of a set of threads with respect to which the memory transactions are committed. For example, a first type of memory barrier instruction may commit the memory transactions to a level of a set of cooperating threads that share an L1 (level one) cache. A second type of memory barrier instruction may commit the memory transactions to a level of a set of threads sharing a global memory. Finally, a third type of memory barrier instruction may commit the memory transactions to a system level of all threads sharing all system memories. The latency required to execute the memory barrier instruction varies based on the type of memory barrier instruction.
    • 本发明的一个实施例提出了一种用于在多个并行线程之间聚合存储器屏障操作的技术。 来自给定并行线程处理单元的存储器屏障请求被合并以减少对系统其余部分的影响。 此外,存储器屏障请求可以指定针对其提交内存事务的一组线程的级别。 例如,第一类型的存储器障碍指令可以将存储器事务提交到共享L1(一级)高速缓存的一组协作线程的级别。 第二种类型的存储器障碍指令可以将存储器事务提交到共享全局存储器的一组线程的级别。 最后,第三种类型的存储器障碍指令可以将存储器事务提交到共享所有系统存储器的所有线程的系统级。 执行存储器屏障指令所需的延迟基于存储器屏障指令的类型而变化。
    • 9. 发明授权
    • Support for non-local returns in parallel thread SIMD engine
    • 支持并行线程SIMD引擎中的非本地返回
    • US08572355B2
    • 2013-10-29
    • US12881065
    • 2010-09-13
    • Guillermo Juan RozasBrett W. Coon
    • Guillermo Juan RozasBrett W. Coon
    • G06F9/30
    • G06F9/30058G06F9/3851
    • One embodiment of the present invention sets forth a method for executing a non-local return instruction in a parallel thread processor. The method comprises the steps of receiving, within the thread group, a first long jump instruction and, in response, popping a first token from the execution stack. The method also comprises determining whether the first token is a first long jump token that was pushed onto the execution stack when a first push instruction associated with the first long jump instruction was executed, and when the first token is the first long jump token, jumping to the second instruction based on the address specified by the first long jump token, or, when the first token is not the first long jump token, disabling the active thread until the first long jump token is popped from the execution stack.
    • 本发明的一个实施例提出了一种用于在并行线程处理器中执行非本地返回指令的方法。 该方法包括以下步骤:在线程组内接收第一长跳转指令,作为响应,从执行堆栈中弹出第一个令牌。 该方法还包括当与第一长跳转指令相关联的第一推送指令被执行时,确定第一令牌是否是被推送到执行堆栈上的第一长跳转令牌,以及当第一令牌是第一长跳转令牌时,跳转 基于由第一长跳转令牌指定的地址到第二指令,或者当第一令牌不是第一长跳转令牌时,禁用活动线程,直到从执行堆栈弹出第一个长跳转令牌。