会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 2. 发明授权
    • Assembler tool for processor-coprocessor computer systems
    • 处理器 - 协处理器计算机系统的汇编工具
    • US06588008B1
    • 2003-07-01
    • US09546755
    • 2000-04-11
    • Marco C. HeddesRoss Boyd LeavensMark Anthony Rinaldi
    • Marco C. HeddesRoss Boyd LeavensMark Anthony Rinaldi
    • G06F945
    • G06F9/30043G06F8/47G06F9/30101G06F9/30145G06F9/30167G06F9/3885
    • A central processor-coprocessor assembly comprising an assembler software tool for extending the base central processor tasks into at least one coprocessor. What is important is that the assembler software tool does not need to be rebuilt when changes are made to the coprocessor elements. The invention allows assembly time extension of a base core language processing (CLP) programming model, without the need to rebuild the assembler tool itself. The assembler tool comprises a set of commands which enable the central processor to manipulate the coprocessor registers, and a coprocessor execute instruction, which initiates command processing on the coprocessor. The present invention simplifies the maintenance of the assembler tool through multiple hardware revisions by enabling hardware designers to update their coprocessor definition files to reflect new or modified coprocessors.
    • 一种中央处理器 - 协处理器组件,其包括用于将基本中央处理器任务扩展到至少一个协处理器的汇编器软件工具。 重要的是,当对协处理器元素进行更改时,汇编程序软件工具不需要重新构建。 本发明允许基本核心语言处理(CLP)编程模型的组装时间扩展,而不需要重建汇编器工具本身。 汇编器工具包括使得中央处理器能够操纵协处理器寄存器的一组命令,以及协处理器执行指令,其在协处理器上启动命令处理。 本发明通过使得硬件设计者能够更新其协处理器定义文件以反映新的或修改的协处理器,通过多个硬件修订来简化了汇编工具的维护。
    • 4. 发明授权
    • Instruction memory system for multi-processor environment and disjoint tasks
    • 指令存储系统,用于多处理器环境和不相交任务
    • US06760743B1
    • 2004-07-06
    • US09477757
    • 2000-01-04
    • Marco C. HeddesMark Anthony RinaldiBrian Alan Youngman
    • Marco C. HeddesMark Anthony RinaldiBrian Alan Youngman
    • G06F900
    • G06F9/3802G06F9/3851
    • An instruction memory system is shared by a plurality of processors and the system utilizes an increased bandwidth to support the combined number of processors. The total instruction address space is divided into code segments according to the disjoint tasks to be performed. The instruction codes of each processor are consolidated into one copy for control instructions and duplicate copies for other disjoint tasks such as inbound requests and outbound requests that have greater processor contention. Interleaving of the memory arrays for certain disjoint tasks serves to provide a larger number of instructions for these tasks. The system utilizes arbiters to receive all disjoint tasks and to control multiplexors that send addresses to memory arrays.
    • 指令存储器系统由多个处理器共享,并且系统利用增加的带宽来支持组合的处理器数量。 总指令地址空间根据要执行的不相交任务分为代码段。 每个处理器的指令代码被合并到一个副本中,用于控制指令和其他不相交任务的重复副本,例如具有更大处理器争用的入站请求和出站请求。 用于某些不相交任务的存储器阵列的交织用于为这些任务提供更多数量的指令。 该系统利用仲裁器接收所有不相交的任务,并控制向存储器阵列发送地址的多路复用器。
    • 6. 发明授权
    • Network processor/software control architecture
    • 网络处理器/软件控制架构
    • US06898179B1
    • 2005-05-24
    • US09544896
    • 2000-04-07
    • Brian Mitchell BassJean Louis CalvignacAnthony Matteo GalloMarco C. HeddesMark Anthony RinaldiMichael Steven SiegelColin Beaton VerrilliFabrice Jean Verplanken
    • Brian Mitchell BassJean Louis CalvignacAnthony Matteo GalloMarco C. HeddesMark Anthony RinaldiMichael Steven SiegelColin Beaton VerrilliFabrice Jean Verplanken
    • G06F11/00H04B7/216H04L1/16H04L12/26
    • G06F15/17
    • The transport protocol for communicating between general purpose processors acting as contact points and network processors in a packet processing environment such as Ethernet is provided. In such an environment, there is at least one single control point processor (CP) and a plurality of network processors (NP), sometimes referred to as blades. A typical system could contain two to sixteen network processors, and each network processor connects to a plurality of devices which communicate with each other over a network transport, such as Ethernet. The CP typically controls the functionality and the functioning of the network processors to function in a way that connects one end user with another, whether or not the end user is on the same network processor or a different network processor. There are three types of communication provided; first, there is communication generally referred to as control services and normally there will be only one pico processor which operates as a GCH (guided cell handler) and only one that operates as a guided tree handler (GTH). A path is provided for the controls to the GCH and the GTH commands, and a separate path is provided for the data frames between the GDH's (general data handler) and the CP.
    • 提供了用于在诸如以太网的分组处理环境中用作接触点的通用处理器和网络处理器之间进行通信的传输协议。 在这样的环境中,存在至少一个单个控制点处理器(CP)和多个网络处理器(NP),有时称为刀片。 典型的系统可以包含两到十六个网络处理器,并且每个网络处理器连接到通过诸如以太网的网络传输彼此通信的多个设备。 CP通常控制网络处理器的功能和功能,以使终端用户与另一终端用户相连的方式起作用,无论终端用户是否在同一个网络处理器或不同的网络处理器上。 提供三种通讯方式; 首先,通常被称为控制服务的通信,并且通常将只有一个微微处理器作为GCH(引导的单元处理器)操作,并且只有一个作为引导树处理器(GTH)操作。 为GCH和GTH命令的控制提供路径,并为GDH(通用数据处理程序)和CP之间的数据帧提供单独的路径。
    • 8. 发明授权
    • Mixed queue scheduler
    • 混合队列调度程序
    • US06728253B1
    • 2004-04-27
    • US09405691
    • 1999-09-24
    • Clark Debs JeffriesMarco C. HeddesMark Anthony RinaldiMichael Steven Siegel
    • Clark Debs JeffriesMarco C. HeddesMark Anthony RinaldiMichael Steven Siegel
    • H04L1256
    • H04L47/50H04L2012/5682
    • A method and system are disclosed for allocating data input bandwidth from a source link to a plurality of N data queues each having a variable occupancy value, Qi(t), and a constant decrement rate, Di, where i designated the ith queue among the N queues. First, a threshold occupancy value, T, is designated for the N queues. During each time step of a repeating time interval, &Dgr;t, the occupancy value, Qi, is compared with T. In response to each and every of said N data queues having occupancy values exceeding T, pausing data transmission from the source link to the N data queues, such that overflow within the data queues is minimized. In response to at least one of the N data queues having an occupancy value less than or equal to T, selecting one among the N data queues to be incremented, and incrementing the selected data queue, such that underflow of the selected queue is minimized. In the context of scheduling one cell per time step, the value of T is one. Furthermore, the method of the present invention guarantees that output port occupancy shall never, in that context, exceed two cells.
    • 公开了一种用于将数据输入带宽从源链路分配给多个N个数据队列的方法和系统,每个N个数据队列具有可变占用值Qi(t)和常数递减率Di,其中i指定i < 排队N队列。 首先,为N个队列指定阈值占有率T。 在重复时间间隔的每个时间步长,Deltat,占用值Qi与T进行比较。响应于具有超过T的占用值的所述N个数据队列中的每一个,暂停从源链路到N的数据传输 数据队列,使数据队列中的溢出最小化。 响应于具有小于或等于T的占用值的N个数据队列中的至少一个,选择要增加的N个数据队列中的一个,并增加所选择的数据队列,使得所选队列的下溢最小化。 在每个时间步长调度一个单元格的上下文中,T的值为1。 此外,本发明的方法保证输出端口占用在这方面永远不超过两个小区。
    • 10. 发明授权
    • Apparatus and method to access computer memory by processing object data as sub-object and shape parameter
    • US06463500B1
    • 2002-10-08
    • US09477775
    • 2000-01-04
    • Marco C. HeddesPiyush Chunilal PatelMark Anthony Rinaldi
    • Marco C. HeddesPiyush Chunilal PatelMark Anthony Rinaldi
    • G06F1200
    • G06F17/30607G06F12/0207G06F17/3061
    • A method is provided for utilizing a memory system which allows for the fast and efficient writing and reading of objects to and from diverse memory chips. A computer system and memory system complex according to method is also provided. The invention defines objects in terms of “shapes.” The shape of an object is defined by two parameters: “Width” and “Height.” Memory system memory chips may comprise sets of different kinds of memory modules which vary in terms of access speed, latency and memory width, such as for example DRAM or SRAM memory modules. The Height of an object denotes the number of consecutive address locations at which the object is stored on a memory module. The Width of an object denotes the number of memory modules at which the object is stored. An advantage of the invention is that objects are defined in terms of “sub-objects” optimized for the memory system memory modules. The sub-objects match the line-width of the memory, thereby allowing the objects to be efficiently written to different memories or memory banks. A further advantage of the invention is that the sub-object shape is transparent to the requester (i.e. transparent to application assembly language). A further advantage of the invention is that sub-objects are handled independently by the memory arbiter, i.e. they can be written to the memory or read from the memory in any order. The complex may further comprise a Tree Search Memory (TSM) system that utilizes a “tree” object hierarchy to perform high-speed memory lookups. Shaping is used to specify how an object is stored in the TSM. The trees consist of different kinds of objects with different shapes. An important advantage of the invention is that the concept of shapes can be used for memory bandwidth distribution and performance increase, allowing objects that are frequently read from memory to be distributed in specific sub-object ordering.