会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 31. 发明申请
    • NETWORK DIRECT MEMORY ACCESS
    • 网络直接存储器访问
    • WO2008021530A2
    • 2008-02-21
    • PCT/US2007018329
    • 2007-08-17
    • PA SEMI INCDESAI SHAILENDRA SHAYTER MARK DGO DOMINIC
    • DESAI SHAILENDRA SHAYTER MARK DGO DOMINIC
    • H04L12/56G06F13/28H04L29/08
    • H04L49/35H04L49/109H04L49/352H04L69/324
    • In one embodiment, a system comprises at least a first node and a second node coupled to a network. The second node comprises a local memory and a direct memory access (DMA) controller coupled to the local memory. The first node is configured to transmit at least a first packet to the second node to access data in the local memory and at least one other packet that is not coded to access the local memory. The second node is configured to capture the packet from a data link layer of a protocol stack, and wherein the DMA controller is configured to perform one or more transfers with the local memory to access the data specified by the first packet responsive to the first packet received from the data link layer. The second node is configured to process the other packet to a top of the protocol stack.
    • 在一个实施例中,系统至少包括耦合到网络的第一节点和第二节点。 第二节点包括本地存储器和耦合到本地存储器的直接存储器存取(DMA)控制器。 第一节点被配置为向第二节点传输至少第一分组以访问本地存储器中的数据以及未被编码以访问本地存储器的至少一个其他分组。 所述第二节点被配置为从协议栈的数据链路层捕获所述分组,并且其中所述DMA控制器被配置为响应于所述第一分组与所述本地存储器进行一次或多次传送以访问由所述第一分组指定的数据 从数据链路层接收。 第二个节点配置为将另一个数据包处理到协议栈的顶部。
    • 34. 发明申请
    • NETWORK-ON-CHIP ENVIRONMENT AND METHOD FOR REDUCTION OF LATENCY
    • 网络在线环境和减少延迟的方法
    • WO2006106476A1
    • 2006-10-12
    • PCT/IB2006/051013
    • 2006-04-04
    • KONINKLIJKE PHILIPS ELECTRONICS N. V.GANGWAL, Om Prakash
    • GANGWAL, Om Prakash
    • H04L12/56H04Q11/04
    • H04L49/254G06F15/7825H04L45/40H04L49/109H04L49/205
    • The invention relates to an integrated circuit comprising a plurality of processing modules (21, 23, M, S; IP) and a network (NoC) arranged for coupling processing modules (21, 23, M, S; IP), comprising: the processing module (21, 23, M, S; IP) includes an associated network interface (NI) which is provided for transmitting data to the network (NoC) supplied by the associated processing module and for receiving data from the network (NoC) destined for the associated processing module; wherein the data transmission between processing modules (21, 23, M, S; IP) operates based on time division multiple access (TDMA) using time slots; each network interface (NI) includes a slot table for storing an allocation of a time slot to a connection (C1-C4), wherein multiple connections (C1-C4) are provided between a first processing module (21, M, IP) and a second processing module ( 23, S, IP) and a sharing of time slots allocated to these multiple connections between the first and a second processing modules is provided. The invention use the idea to utilize all or a part of time slots in common, which are allocated for multiple connections between a first and a second processing module, in order to reduce the latency of such connections. By sharing of slots assigned to multiple connections between two processing module a large pool of slots during one revolution of a slot table is formed. Thus the latency to access a burst of data could be reduced.
    • 本发明涉及一种集成电路,它包括多个处理模块(21,23,M,S; IP)和一个网络(NoC),用于耦合处理模块(21,23,M,S; IP),包括: 处理模块(21,23,M,S; IP)包括相关联的网络接口(NI),其被提供用于将数据发送到由相关处理模块提供的网络(NoC),并用于从目标地址(NoC)接收数据 用于相关处理模块; 其中处理模块(21,23,M,S; IP)之间的数据传输基于使用时隙的时分多址(TDMA)操作; 每个网络接口(NI)包括用于存储对连接(C1-C4)的时隙分配的时隙表,其中在第一处理模块(21,M,IP)和 提供第二处理模块(23,S,IP)以及分配给第一和第二处理模块之间的这些多个连接的时隙的共享。 本发明使用该思想来利用共同的全部或部分时隙,这些时隙被分配用于第一和第二处理模块之间的多个连接,以便减少这种连接的等待时间。 通过共享分配给两个处理模块之间的多个连接的时隙,形成在一个时隙表的一圈内的大量时隙池。 因此,可以减少访问数据突发的延迟。
    • 37. 发明申请
    • MASSIVELY PARALLEL COMPUTER, ACCELERATED COMPUTING CLUSTERS, AND TWO DIMENSIONAL ROUTER AND INTERCONNECTION NETWORK FOR FIELD PROGRAMMABLE GATE ARRAYS, AND APPLICATIONS
    • 大型并行计算机,加速计算集群,以及用于现场可编程门阵列的二维路由器和互连网络,以及应用
    • WO2017120270A1
    • 2017-07-13
    • PCT/US2017/012230
    • 2017-01-04
    • GRAY RESEARCH LLC
    • GRAY, Jan Stephen
    • H04L12/933H04L12/931
    • G06F13/36G06F13/4068H04L49/109
    • An embodiment of a massively parallel computing system comprising a plurality of processors, which may be subarranged into clusters of processors, and interconnected by means of a configurable directional 2D router for Networks on Chips (NOCs) is disclosed. The system further comprises diverse high bandwidth external I/O devices and interfaces, which may include without limitation Ethernet interfaces, and dynamic RAM (DRAM) memories. The system is designed for implementation in programmable logic in FPGAs, but may also be implemented in other integrated circuit technologies, such as non programmable circuitry, and in integrated circuits such as application specific integrated circuits (ASICs). The system enables the practical implementation of diverse FPGA computing accelerators to speed up computation for example in data centers or telecom networking infrastructure. The system uses the NOC to interconnect processors, clusters, accelerators, and/or external interfaces. A great diversity of NOC client cores, for communication amongst various external interfaces and devices, and on-chip interfaces and resources, may be coupled to a router in order to efficiently communicate with other NOC client cores. The system, router, and NOC enable feasible FPGA implementation of large integrated systems on chips, interconnecting hundreds of client cores over high bandwidth links, including compute and accelerator cores, industry standard IP cores, DRAM/HBM/HMC channels, PCI Express channels, and 10G/25G/40G/100G/400G networks.
    • 公开了包括多个处理器的大规模并行计算系统的实施例,其可以被子排列成处理器群并且通过用于芯片上网络(NOC)的可配置定向2D路由器互连 。 该系统还包括不同的高带宽外部I / O设备和接口,其可以包括但不限于以太网接口和动态RAM(DRAM)存储器。 该系统设计用于在FPGA中的可编程逻辑中实现,但也可以在其他集成电路技术(例如非可编程电路)以及诸如专用集成电路(ASIC)的集成电路中实现。 该系统能够实现多种FPGA计算加速器的实际实施,以加速计算,例如在数据中心或电信网络基础设施中。 系统使用NOC互连处理器,集群,加速器和/或外部接口。 为了与其他NOC客户端核心进行有效通信,可以将各种NOC客户端核心(用于各种外部接口和设备之间的通信以及片上接口和资源)耦合到路由器。 系统,路由器和NOC可以在芯片上实现大型集成系统的可行FPGA实现,通过高带宽链路(包括计算和加速器内核,工业标准IP内核,DRAM / HBM / HMC通道,PCI Express通道等)互连数百个客户端内核。 和10G / 25G / 40G / 100G / 400G网络。
    • 39. 发明申请
    • MITIGATING TRAFFIC STEERING INEFFICIENCIES IN DISTRIBUTED UNCORE FABRIC
    • 在分销的UNCORE FABRIC中减少交通转向失效
    • WO2016105967A1
    • 2016-06-30
    • PCT/US2015/065138
    • 2015-12-11
    • INTEL CORPORATION
    • NAGARAJAN, RamadassFREDERICK, Michael Todd
    • G06F13/16G06F12/08
    • H04L49/25G06F15/7807H04L45/7453H04L49/109H04L49/356
    • In an example, selected portions of an uncore fabric of a system-on-a-chip (SoC) or other embedded system is divided into two independent pipelines. Each pipeline operates independently of the other pipeline, and each accesses only one-half of the system memory, such as even or odd addresses in an interleaved memory. The two pipelines do not reconverge until after memory values have been returned. However, the uncore fabric may still present a single, monolithic interface to requesting devices. This allows system designers to treat the uncore fabric as a "black box" without modifying existing designs. Each incoming address may be processed by a deterministic hash, assigned to one of the pipelines, processed through memory, and then passed to a credit return.
    • 在一个示例中,片上系统(SoC)或其他嵌入式系统的非空心结构的选定部分被分成两个独立的管道。 每个流水线独立于其他流水线运行,并且每个流水线仅访问系统存储器的一半,例如交错存储器中的偶数或奇数地址。 在返回内存值之后,两条管道不会重新映射。 然而,不织布可能仍然呈现出一个单一的接口来请求设备。 这使得系统设计人员可以将不干净的织物视为“黑匣子”,而无需修改现有设计。 每个传入地址可以通过确定性散列来处理,分配给一条管道,通过存储器处理,然后传递给信用回报。