会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 36. 发明申请
    • Snoop filtering system in a multiprocessor system
    • 多处理器系统中的Snoop过滤系统
    • US20060224835A1
    • 2006-10-05
    • US11093127
    • 2005-03-29
    • Matthias BlumrichDong ChenAlan GaraMark GiampapaPhilip HeidelbergerDirk HoenickeMartin OhmachtValentina SalapuraPavlos Vranas
    • Matthias BlumrichDong ChenAlan GaraMark GiampapaPhilip HeidelbergerDirk HoenickeMartin OhmachtValentina SalapuraPavlos Vranas
    • G06F13/28
    • G06F12/0831G06F12/0813Y02D10/13
    • A system and method for supporting cache coherency in a computing environment having multiple processing units, each unit having an associated cache memory system operatively coupled therewith. The system includes a plurality of interconnected snoop filter units, each snoop filter unit corresponding to and in communication with a respective processing unit, with each snoop filter unit comprising a plurality of devices for receiving asynchronous snoop requests from respective memory writing sources in the computing environment; and a point-to-point interconnect comprising communication links for directly connecting memory writing sources to corresponding receiving devices; and, a plurality of parallel operating filter devices coupled in one-to-one correspondence with each receiving device for processing snoop requests received thereat and one of forwarding requests or preventing forwarding of requests to its associated processing unit. Each of the plurality of parallel operating filter devices comprises parallel operating sub-filter elements, each simultaneously receiving an identical snoop request and implementing one or more different snoop filter algorithms for determining those snoop requests for data that are determined not cached locally at the associated processing unit and preventing forwarding of those requests to the processor unit. In this manner, a number of snoop requests forwarded to a processing unit is reduced thereby increasing performance of the computing environment.
    • 一种用于在具有多个处理单元的计算环境中支持高速缓存一致性的系统和方法,每个单元具有与其可操作耦合的相关联的高速缓存存储器系统 该系统包括多个互连的窥探过滤器单元,每个窥探过滤器单元对应于相应处理单元并与其通信,每个窥探过滤器单元包括用于在计算环境中从相应存储器写入源接收异步窥探请求的多个设备 ; 以及包括用于将存储器写入源直接连接到对应的接收设备的通信链路的点对点互连; 以及与每个接收设备一一对应地耦合的多个并行操作过滤器设备,用于处理在其上接收的窥探请求,并且转发请求之一或者阻止将请求转发到其相关联的处理单元。 多个并行操作过滤器装置中的每一个包括并行操作子滤波器元件,每个并行操作子滤波器元件同时接收相同的窥探请求,并且实现一个或多个不同的窥探滤波器算法,用于确定对于在相关处理中本地未被缓存的数据被确定的窥探请求 并且防止将这些请求转发到处理器单元。 以这种方式,减少了转发到处理单元的多个窥探请求,从而增加了计算环境的性能。
    • 40. 发明授权
    • DMA engine for repeating communication patterns
    • 用于重复通信模式的DMA引擎
    • US07802025B2
    • 2010-09-21
    • US11768795
    • 2007-06-26
    • Dong ChenAlan G. GaraMark E. GiampapaPhilip HeidelbergerBurkhard Steinmacher-BurowPavlos Vranas
    • Dong ChenAlan G. GaraMark E. GiampapaPhilip HeidelbergerBurkhard Steinmacher-BurowPavlos Vranas
    • G06F13/28
    • G06F15/163
    • A parallel computer system is constructed as a network of interconnected compute nodes to operate a global message-passing application for performing communications across the network. Each of the compute nodes includes one or more individual processors with memories which run local instances of the global message-passing application operating at each compute node to carry out local processing operations independent of processing operations carried out at other compute nodes. Each compute node also includes a DMA engine constructed to interact with the application via Injection FIFO Metadata describing multiple Injection FIFOs where each Injection FIFO may containing an arbitrary number of message descriptors in order to process messages with a fixed processing overhead irrespective of the number of message descriptors included in the Injection FIFO.
    • 并行计算机系统被构造为互连的计算节点的网络,以操作用于在整个网络上执行通信的全局消息传递应用。 每个计算节点包括具有存储器的一个或多个单独处理器,该存储器运行在每个计算节点处操作的全局消息传递应用的本地实例,以独立于在其他计算节点执行的处理操作来执行本地处理操作。 每个计算节点还包括构造成通过描述多个注入FIFO的注入FIFO元数据与应用交互的DMA引擎,其中每个注入FIFO可以包含任意数量的消息描述符,以便处理具有固定处理开销的消息,而不管消息的数量 描述符包含在注入FIFO中。