会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 32. 发明授权
    • High performance recoverable communication method and apparatus for
write-only networks
    • 用于只写网络的高性能可恢复通信方法和装置
    • US6049889A
    • 2000-04-11
    • US6115
    • 1998-01-13
    • Simon C. Steely, Jr.Glenn P. GarveyRichard B. Gillett, Jr.
    • Simon C. Steely, Jr.Glenn P. GarveyRichard B. Gillett, Jr.
    • H04L29/06H04L29/14G06F3/00
    • H04L29/06H04L69/40
    • A multi-node computer network includes a plurality of nodes coupled together via a data link. Each of the nodes includes a local memory, which further comprises a shared memory. Certain items of data that are to be shared by the nodes are stored in the shared portion of memory. Associated with each of the shared data items is a data structure. When a node sharing data with other nodes in the system seeks to modify the data, it transmits the modifications over the data link to the other nodes in the network. Each update is received in order by each node in the cluster. As part of the last transmission by the modifying node, an acknowledgement request is sent to the receiving nodes in the cluster. Each node that receives the acknowledgment request returns an acknowledgement to the sending node. The returned acknowledgement is written to the data structure associated with the shared data item. If there is an error during the transmission of the message, the receiving node does not transmit an acknowledgement, and the sending node is thereby notified that an error has occurred.
    • 多节点计算机网络包括通过数据链路耦合在一起的多个节点。 每个节点包括本地存储器,其还包括共享存储器。 要由节点共享的某些数据项存储在存储器的共享部分中。 与每个共享数据项相关联的是数据结构。 当与系统中的其他节点共享数据的节点寻求修改数据时,它将数据链路上的修改发送到网络中的其他节点。 群集中的每个节点按顺序接收每个更新。 作为修改节点的最后一次传输的一部分,向群集中的接收节点发送确认请求。 接收确认请求的每个节点向发送节点返回确认。 返回的确认被写入与共享数据项相关联的数据结构。 如果在消息的发送期间存在错误,则接收节点不发送确认,并且由此通知发送节点发生了错误。
    • 33. 发明授权
    • Multi-index multi-way set-associative cache
    • 多索引多路组合关联缓存
    • US5509135A
    • 1996-04-16
    • US951623
    • 1992-09-25
    • Simon C. Steely, Jr.
    • Simon C. Steely, Jr.
    • G06F12/08G06F13/00
    • G06F12/0864
    • A plurality of indexes are provided for a multi-way set-associate cache of a computer system. The cache is organized as a plurality of blocks for storing data which are a copies of main memory data. Each block has an associated tag for uniquely identifying the block. The blocks and the tags are addressed by indexes. The indexes are generated by a Boolean hashing function which converts a memory address to cache indexes by combining the bits of the memory address using an exclusive OR function. Different combination of bits are used to generate a plurality of different indexes to address the tags and the associated blocks to transfer data between the cache and the central processing unit of the computer system.
    • 为计算机系统的多路集合相关缓存提供多个索引。 高速缓存被组织为用于存储作为主存储器数据的副本的数据的多个块。 每个块都具有用于唯一标识块的关联标签。 块和标签由索引寻址。 索引由布尔散列函数生成,该函数通过使用异或函数组合存储器地址的位来将存储器地址转换为缓存索引。 使用不同的比特组合来生成多个不同的索引以寻址标签和相关联的块以在计算机系统的高速缓存和中央处理单元之间传送数据。
    • 34. 发明授权
    • Set prediction cache memory system using bits of the main memory address
    • 使用主存储器地址的位设置预测高速缓存存储器系统
    • US5235697A
    • 1993-08-10
    • US956827
    • 1992-10-05
    • Simon C. Steely, Jr.John H. Zurawski
    • Simon C. Steely, Jr.John H. Zurawski
    • G06F12/08
    • G06F12/0864G06F2212/6082
    • The set-prediction cache memory system comprises an extension of a set-associative cache memory system which operates in parallel to the set-associative structure to increase the overall speed of the cache memory while maintaining its performance. The set prediction cache memory system includes a plurality of data RAMs and a plurality of tag RAMs to store data and data tags, respectively. Also included in the system are tag store comparators to compare the tag data contained in a specific tag RAM location with a second index comprising a predetermined second portion of a main memory address. The elements of the set prediction cache memory system which operate in parallel to the set-associative cache memory include: a set-prediction RAM which receives at least one third index comprising a predetermined third portion of the main memory address, and stores such third index to essentially predict the data cache RAM holding the data indexed by the third index; a data-select multiplexer which receives the prediction index and selects a data output from the data cache RAM indexed by the prediction index; and a mispredict logic device to determine if the set prediction RAM predicted the correct data cache RAM and if not, issue a mispredict signal which may comprise a write data signal, the write data signal containing information intended to correct the prediction index contained in the set prediction RAM.
    • 设置预测高速缓冲存储器系统包括与集合关联结构并行操作的集合关联高速缓冲存储器系统的扩展,以在保持其性能的同时增加高速缓冲存储器的总体速度。 集合预测高速缓冲存储器系统包括分别存储数据和数据标签的多个数据RAM和多个标签RAM。 还包括在系统中的标签存储比较器,用于将包含在特定标签RAM位置中的标签数据与包含主存储器地址的预定第二部分的第二索引进行比较。 与设置关联高速缓存存储器并行操作的集合预测高速缓冲存储器系统的元件包括:设置预测RAM,其接收包含主存储器地址的预定第三部分的至少一个第三索引,并存储这样的第三索引 以基本预测由第三指标索引的数据的数据缓存RAM; 数据选择多路复用器,其接收预测索引并选择从由预测索引索引的数据高速缓存RAM输出的数据; 以及用于确定所设置的预测RAM是否预测正确的数据高速缓存RAM的错误预测逻辑设备,如果不是,则发出可能包括写入数据信号的错误预测信号,所述写入数据信号包含旨在校正包含在该组中的预测索引的信息 预测RAM。
    • 36. 发明授权
    • Source request arbitration
    • 源请求仲裁
    • US07340565B2
    • 2008-03-04
    • US10755919
    • 2004-01-13
    • Simon C. Steely, Jr.Gregory Edward Tierney
    • Simon C. Steely, Jr.Gregory Edward Tierney
    • G06F9/00G06F9/38G06F13/00
    • G06F12/0811G06F12/0815G06F12/0828G06F12/084G06F2212/507
    • Multiprocessor systems and methods are disclosed. One embodiment may comprise a plurality of processor cores. A given processor core may be operative to generate a request for desired data in response to a cache miss at a local cache. A shared cache structure may provide at least one speculative data fill and a coherent data fill of the desired data to at least one of the plurality of processor cores in response to a request from the at least one processor core. A processor scoreboard arbitrates the requests for the desired data. A speculative data fill of the desired data is provided to the at least one processor core. The coherent data fill of the desired data may be provided to the at least one processor core in a determined order.
    • 公开了多处理器系统和方法。 一个实施例可以包括多个处理器核。 给定的处理器核心可以用于响应于本地高速缓存处的高速缓存未命中而产生对期望数据的请求。 响应于来自至少一个处理器核心的请求,共享高速缓存结构可以向所述多个处理器核心中的至少一个提供期望数据的至少一个推测数据填充和相干数据填充。 处理器记分板对所需数据的请求进行仲裁。 将所需数据的推测数据填充提供给至少一个处理器核。 期望数据的相干数据填充可以以确定的顺序提供给至少一个处理器核心。
    • 37. 发明授权
    • Method and apparatus for adaptively bypassing one or more levels of a cache hierarchy
    • 用于自适应地绕过高速缓存层级的一个或多个级别的方法和装置
    • US06647466B2
    • 2003-11-11
    • US09769552
    • 2001-01-25
    • Simon C. Steely, Jr.
    • Simon C. Steely, Jr.
    • G06F1200
    • G06F12/0897G06F12/0811G06F12/0888G06F12/0891
    • A system for adaptively bypassing one or more higher cache levels following a miss in a lower level of a cache hierarchy is described. Each cache level preferably includes a tag store containing address and state information for each cache line resident in the respective cache. When an invalidate request is received at a given cache hierarchy, each cache level is searched for the address specified by the invalidate request. When an address match is detected, the state of the respective cache line is changed to the invalid state, although the address of the cache line is left in the tag store. Thereafter, if the processor or entity associated with this cache hierarchy issues its own request for this same cache line, the cache hierarchy begins searching the tag store of each level starting with the lowest cache level. Since the address of the invalidated cache line was left in the respective tag store, a match will be detected at one of the cache levels, although the corresponding state of this cache line is invalid. This condition is specifically detected and is considered to be an “inval_miss” occurrence. In response, to an inval_miss, the cache hierarchy calls off searching any higher levels, and instead, issues a memory reference request for the desired cache line. In a further embodiment, the entity that sourced an invalidate request is stored, and a subsequent memory reference request for the same cache line is sent directly to the source entity.
    • 描述了用于在高速缓存层级的较低级别中错过之后自适应地绕过一个或多个更高的高速缓存级别的系统。 每个高速缓存级别优选地包括标签存储,其包含驻留在相应高速缓存中的每个高速缓存行的地址和状态信息。 当在给定的缓存层次结构中接收到无效请求时,将搜索每个高速缓存级别以查找由无效请求指定的地址。 当检测到地址匹配时,尽管高速缓存行的地址被留在标签存储器中,但各个高速缓存行的状态被改变为无效状态。 此后,如果与该高速缓存层级相关联的处理器或实体发出其对该相同高速缓存行的自身请求,则高速缓存层级开始以最低高速缓存级别开始搜索每个级别的标签存储。 由于无效高速缓存行的地址被留在相应的标签存储中,所以在高速缓存级别之一处将检测到匹配,尽管该高速缓存行的相应状态是无效的。 该条件被特别检测并被认为是“inval_miss”事件。 作为响应,对于inval_miss,缓存层次结构调用搜索任何更高级别,而是发出所需高速缓存行的内存引用请求。 在另一个实施例中,存储了源自无效请求的实体,并且将相同高速缓存行的后续存储器引用请求直接发送到源实体。
    • 38. 发明授权
    • System for passing an index value with each prediction in forward
direction to enable truth predictor to associate truth value with
particular branch instruction
    • 用于向前传递每个预测的索引值的系统,以使真实预测器能够将真值与特定分支指令相关联
    • US6081887A
    • 2000-06-27
    • US191869
    • 1998-11-12
    • Simon C. Steely, Jr.Edward J. McLellanJoel S. Emer
    • Simon C. Steely, Jr.Edward J. McLellanJoel S. Emer
    • G06F9/38G06F9/32
    • G06F9/3844
    • A technique for predicting the result of a conditional branch instruction for use with a processor having instruction pipeline. A stored predictor is connected to the front end of the pipeline and is trained from a truth based predictor connected to the back end of the pipeline. The stored predictor is accessible in one instruction cycle, and therefore provides minimum predictor latency. Update latency is minimized by storing multiple predictions in the front end stored predictor which are indexed by an index counter. The multiple predictions, as provided by the back end, are indexed by the index counter to select a particular one as current prediction on a given instruction pipeline cycle. The front end stored predictor also passes along to the back end predictor, such as through the instruction pipeline, a position value used to generate the predictions. This further structure accommodates ghost branch instructions that turn out to be flushed out of the pipeline when it must be backed up. As a result, the front end always provides an accurate prediction with minimum update latency.
    • 一种用于预测与具有指令流水线的处理器一起使用的条件转移指令的结果的技术。 存储的预测器连接到管道的前端,并且从连接到管道后端的基于真实的预测器训练。 存储的预测器可以在一个指令周期中访问,因此提供最小预测器延迟。 通过将多个预测存储在由索引计数器索引的前端存储的预测器中来最小化更新延迟。 由后端提供的多个预测由索引计数器索引,以选择特定的预测作为给定指令流水线周期上的当前预测。 前端存储的预测器还将传递到后端预测器,例如通过指令流水线,用于产生预测的位置值。 这种进一步的结构可以容纳重影分支指令,当它必须被备份时,这些指令将被清除流出管道。 因此,前端总是以最小的更新延迟提供准确的预测。