会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 1. 发明授权
    • Method for reducing directory writes and latency in a high performance, directory-based, coherency protocol
    • 在高性能,基于目录的一致性协议中减少目录写入和延迟的方法
    • US06654858B1
    • 2003-11-25
    • US09652324
    • 2000-08-31
    • David H. AsherBrian LillyRichard E. KesslerMichael Bertone
    • David H. AsherBrian LillyRichard E. KesslerMichael Bertone
    • G06F1200
    • G06F12/0817G06F12/0824
    • A computer system has a plurality of processors wherein each processor preferably has its own cache memory. Each processor or group of processors may have a memory controller that interfaces to a main memory. Each main memory includes a “directory” that maintains the directory coherence state of each block of that memory. One or more of the processors are members of a “local” group of processors. Processors outside a local group are referred to as “remote” processors with respect to that local group. Whenever a remote processor performs a memory reference for a particular block of memory, the processor that maintains the directory for that block normally updates the directory to reflect that the remote processor now has exclusive ownership of the block. However, memory references between processors within a local group do not result in directory writes. Instead, the cache memory of the local processor that initiated the memory requests places or updates a copy of the requested data in its cache memory and also sets associated tag control bits to reflect the same or similar information as would have been written to the directory. If a subsequent request is received for that same block, the local processor that previously accessed the block examines its cache for the associated tag control bits. Using those bits, that processor will determine that it currently has the block exclusive and provides the requested data to the new processor that is requesting the data.
    • 计算机系统具有多个处理器,其中每个处理器优选地具有其自己的高速缓冲存储器。 每个处理器或处理器组可以具有与主存储器接口的存储器控​​制器。 每个主存储器包括维护该存储器的每个块的目录一致状态的“目录”。 一个或多个处理器是“本地”处理器组的成员。 本地组外的处理器称为“本地组”的“远程”处理器。 每当远程处理器执行特定内存块的内存引用时,维护该块的目录的处理器通常会更新目录,以反映远程处理器现在拥有该块的独占所有权。 但是,本地组内处理器之间的内存引用不会导致目录写入。 相反,启动存储器请求的本地处理器的高速缓存存储器将所请求的数据的副本放置或更新在其高速缓冲存储器中,并且还设置相关联的标签控制位以反映与已经写入目录的相同或相似的信息。 如果接收到该相同块的后续请求,则先前访问该块的本地处理器检查其相关标签控制位的高速缓存。 使用这些位,该处理器将确定其当前具有块排他性,并且向请求数据的新处理器提供所请求的数据。
    • 2. 发明授权
    • Fast lane prefetching
    • 快速车道预取
    • US06681295B1
    • 2004-01-20
    • US09652451
    • 2000-08-31
    • Stephen C. RootRichard E. KesslerDavid H. AsherBrian Lilly
    • Stephen C. RootRichard E. KesslerDavid H. AsherBrian Lilly
    • G06F1200
    • G06F9/30047G06F9/383G06F9/3832G06F9/3836G06F9/384G06F9/3855G06F9/3857G06F9/3859G06F12/0862G06F12/0864G06F2212/6028
    • A computer system has a set-associative, multi-way cache system, in which at least one way is designated as a fast lane, and remaining way(s) are designated slow lanes. Any data that needs to be loaded into cache, but is not likely to be needed again in the future, preferably is loaded into the fast lane. Data loaded into the fast lane is earmarked for immediate replacement. Data loaded into the slow lanes preferably is data that may not needed again in the near future. Slow data is kept in cache to permit it to be reused if necessary. The high-performance mechanism of data access in a modem microprocessor is with a prefetch; data is moved, with a special prefetch instruction, into cache prior to its intended use. The prefetch instruction requires less machine resources, than carrying out the same intent with an ordinary load instruction. So, the slow-lane, fast-lane decision is accomplished by having a multiplicity of prefetch instructions. By loading “not likely to be needed again” data into the fast lane, and designating such data for immediate replacement, data in other cache blocks, in the other ways, may not be evicted, and overall system performance is increased.
    • 计算机系统具有集合关联的多路缓存系统,其中至少一种方式被指定为快速通道,并且剩余方式被指定为慢车道。 任何需要加载到缓存中但不太可能再次需要的数据最好被加载到快速通道中。 加载到快速通道的数据被指定用于立即更换。 加载到慢车道中的数据优选地是在不久的将来可能不再需要的数据。 慢数据保存在缓存中,以便在必要时重新使用它。 调制解调器微处理器中数据访问的高性能机制具有预取功能; 在预期使用之前,将数据用特殊的预取指令移动到缓存中。 预取指令比普通加载指令执行相同的意图要求较少的机器资源。 因此,通过具有多个预取指令来实现慢通道,快速通道决定。 通过将“不太可能需要再次”的数据加载到快速通道中,并且指定这样的数据以立即替换,以其他方式在其他高速缓存块中的数据可能不被驱逐,并且整体系统性能增加。
    • 3. 发明授权
    • Method and system for absorbing defects in high performance microprocessor with a large n-way set associative cache
    • 用大型n路组合关联缓存来吸收高性能微处理器缺陷的方法和系统
    • US07370151B2
    • 2008-05-06
    • US10690137
    • 2003-10-21
    • David H. AsherBrian LillyJoel GrodsteinPatrick M. Fitzgerald
    • David H. AsherBrian LillyJoel GrodsteinPatrick M. Fitzgerald
    • G06F13/00
    • G11C29/76G06F12/0864G11C15/00
    • A method and architecture for improving the usability and manufacturing yield of a microprocessor having a large on-chip n-way set associative cache. The architecture provides a method for working around defects in the portion of the die allocated to the data array of the cache. In particular, by adding a plurality of muxes to a way or ways in the data array of an associative cache having the shorter paths to the access control logic, each way in a bank can be selectively replaced or remapped to the ways with the shorter paths without adding any latency to the system. This selective remapping of separate ways in individual banks of the set associative cache provides a more efficient way to absorb defects and allows more defects to be absorbed in the data array of a set associative cache.
    • 一种用于提高具有大片上n路组相关高速缓存的微处理器的可用性和制造产量的方法和架构。 该架构提供了一种用于处理分配给高速缓存的数据阵列的芯片部分中的缺陷的方法。 特别地,通过向具有到访问控制逻辑的较短路径的关联高速缓存的数据阵列中的方式或方式添加多个多路复用器,可以选择性地将存储体中的每个路径替换或重新映射到具有较短路径的路径 而不会对系统增加任何延迟。 在集合关联高速缓存的各个组中的单独方式的这种选择性重新映射提供了一种更有效的方式来吸收缺陷,并允许在集合关联高速缓存的数据阵列中吸收更多的缺陷。
    • 4. 发明授权
    • Method and system for absorbing defects in high performance microprocessor with a large n-way set associative cache
    • 用大型n路组合关联缓存来吸收高性能微处理器缺陷的方法和系统
    • US06671822B1
    • 2003-12-30
    • US09651948
    • 2000-08-31
    • David H. AsherBrian LillyJoel GrodsteinPatrick M. Fitzgerald
    • David H. AsherBrian LillyJoel GrodsteinPatrick M. Fitzgerald
    • G06F1100
    • G11C29/76G06F12/0864G11C15/00
    • A method and architecture for improving the usability and manufacturing yield of a microprocessor having a large on-chip n-way set associative cache. The architecture provides a method for working around defects in the portion of the die allocated to the data array of the cache. In particular, by adding a plurality of muxes to a way or ways in the data array of an associative cache having the shorter paths to the access control logic, each way in a bank can be selectively replaced or remapped to the ways with the shorter paths without adding any latency to the system. This selective remapping of separate ways in individual banks of the set associative cache provides a more efficient way to absorb defects and allows more defects to be absorbed in the data array of a set associative cache.
    • 一种用于提高具有大片上n路组相关高速缓存的微处理器的可用性和制造产量的方法和架构。 该架构提供了一种用于处理分配给高速缓存的数据阵列的芯片部分中的缺陷的方法。 特别地,通过向具有到访问控制逻辑的较短路径的关联高速缓存的数据阵列中的方式或方式添加多个多路复用器,可以选择性地将存储体中的每个路径替换或重新映射到具有较短路径的路径 而不会对系统增加任何延迟。 在集合关联高速缓存的各个组中的单独方式的这种选择性重新映射提供了一种更有效的方式来吸收缺陷,并允许在集合关联高速缓存的数据阵列中吸收更多的缺陷。