会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 43. 发明授权
    • Supporting late DRAM bank hits
    • 支持晚期DRAM银行点击
    • US08375163B1
    • 2013-02-12
    • US12326060
    • 2008-12-01
    • John H. EdmondsonShane Keil
    • John H. EdmondsonShane Keil
    • G06F12/00G06F13/00G06F13/28
    • G06F13/28
    • One embodiment of the invention sets forth a mechanism to transmit commands received from an L2 cache to a bank page within the DRAM. An arbiter unit determines which commands from a command sorter to transmit to a command queue. An activate command associated with the bank page related to the commands is also transmitted to an activate queue. The last command in the command queue is marked as “last.” An interlock counter stores a count of “last” commands in the read/write command queue. A DRAM controller transmits activate and commands from the activate queue and the command queue to the DRAM. Each time a command marked as “last” is encountered, the DRAM controller decrements the interlock counter. If the count in the interlock counter is zero, then the command marked as “last” is marked as “auto-precharge.” The “auto-precharge” command, when processed, causes the bank page to be closed.
    • 本发明的一个实施例提出了一种将从L2高速缓存接收的命令发送到DRAM内的存储体页面的机制。 仲裁器单元确定哪些命令从命令分拣机发送到命令队列。 与该命令相关联的存储体页面的激活命令也被发送到激活队列。 命令队列中的最后一个命令被标记为last。 互锁计数器存储读/写命令队列中最后命令的计数。 DRAM控制器将激活和命令从激活队列和命令队列传送到DRAM。 每次遇到标记为最后的命令时,DRAM控制器递减联锁计数器。 如果联锁计数器中的计数为零,则标记为最后的命令被标记为自动预充电。 自动预充电命令在处理时会导致银行页面被关闭。
    • 44. 发明授权
    • Managing conflicts on shared L2 bus
    • 管理共享L2总线上的冲突
    • US08321618B1
    • 2012-11-27
    • US12510987
    • 2009-07-28
    • Shane KeilJohn H. Edmondson
    • Shane KeilJohn H. Edmondson
    • G06F13/00
    • G06F13/1605G06F12/0859
    • One embodiment of the present invention sets forth a mechanism to schedule read data transmissions and write data transmissions to/from a cache to frame buffer logic on the L2 bus. When processing a read or a write command, a scheduling arbiter examines a bus schedule to determine that a read-read conflict, a read-write conflict or a write-read exists, and allocates an available memory space in a read buffer to store the read data causing the conflict until the read return data transmission can be scheduled. In the case of a write command, the scheduling arbiter then transmits a write request to a request buffer. When processing a write request, the request arbiter examines the request buffers to determine whether a write-write conflict. If so, then the request arbiter allocates a memory space in a request buffer to store the write request until the write data transmission can be scheduled.
    • 本发明的一个实施例提出了一种机制,用于在L2总线上调度读取数据传输和向高速缓存写入数据传输到帧缓冲器逻辑。 在处理读取或写入命令时,调度仲裁器检查总线调度以确定存在读取冲突,读取冲突或写入读取,并且在读取缓冲器中分配可用的存储器空间以存储 读取导致冲突的数据,直到可以调度读取返回数据传输。 在写命令的情况下,调度仲裁器然后向请求缓冲器发送写请求。 在处理写入请求时,请求仲裁器检查请求缓冲区以确定是否写入写入冲突。 如果是,则请求仲裁器在请求缓冲器中分配存储空间以存储写入请求,直到可以调度写入数据传输。
    • 46. 发明授权
    • Systems for efficient retrieval from tiled memory surface to linear memory display
    • 用于从平铺记忆表面到线性记忆体显示的高效检索系统
    • US07986327B1
    • 2011-07-26
    • US11552082
    • 2006-10-23
    • John H. Edmondson
    • John H. Edmondson
    • G06F12/10G06F13/00G06F13/28G06F9/26G06F9/34
    • G09G5/395G09G5/363G09G2350/00G09G2360/122
    • Embodiments of the present invention set forth a technique for optimizing the on-chip data path between a memory controller and a display controller within a graphics processing unit (GPU). A row selection field and a sector mask are included within a memory access command transmitted from the display controller to the memory controller indicating which row of data is being requested from memory. The memory controller responds to the memory access command by returning only the row of data corresponding to the requested row to the display controller over the on-chip data path. Any extraneous data received by the memory controller in the process of accessing the specifically requested row of data is stripped out and not transmitted back to the display controller. One advantage of the present invention is that the width of the on-chip data path can be reduced by a factor of two or more as a result of the greater operational efficiency gained by stripping out extraneous data before transmitting the data to the display controller.
    • 本发明的实施例提出了一种用于优化图形处理单元(GPU)内的存储器控​​制器和显示控制器之间的片上数据路径的技术。 从显示控制器向存储器控制器发送的指示从存储器请求哪一行数据的存储器访问命令中包括行选择字段和扇区掩码。 存储器控制器通过仅通过片上数据路径仅将与所请求的行相对应的数据行返回到显示控制器来响应存储器访问命令。 在访问特定请求的数据行的过程中由存储器控制器接收的任何无关数据被剥离并且不被传送回显示控制器。 本发明的一个优点在于,由于在将数据发送到显示控制器之前剥离外来数据而获得更大的操作效率,片上数据路径的宽度可以减少2倍或更多。
    • 49. 发明申请
    • Page stream sorter for poor locality access patterns
    • 页面流排序器用于不良的局部访问模式
    • US20080109613A1
    • 2008-05-08
    • US11592540
    • 2006-11-03
    • David A. JaroshSonny S. YeohColyn S. CaseJohn H. Edmondson
    • David A. JaroshSonny S. YeohColyn S. CaseJohn H. Edmondson
    • G06F12/00G06F17/00
    • G06F13/1626
    • In some applications, such as video motion compression processing for example, a request pattern or “stream” of requests for accesses to memory (e.g., DRAM) may have, over a large number of requests, a relatively small number of requests to the same page. Due to the small number of requests to the same page, conventionally sorting to aggregate page hits may not be very effective. Reordering the stream can be used to “bury” or “hide” much of the necessary precharge/activate time, which can have a highly positive impact on overall throughput. For example, separating accesses to different rows of the same bank by at least a predetermined number of clocks can effectively hide the overhead involved in precharging/activating the rows.
    • 在一些应用中,例如视频运动压缩处理,例如,对存储器(例如,DRAM)访问的请求的请求模式或“流”可以在大量请求中具有相对较少数量的请求 页。 由于对同一页面的请求数量不多,常规排序以汇总页面命中可能不是很有效。 重新排序流可以用于“埋葬”或“隐藏”大量必要的预充/激活时间,这对整体吞吐量可能产生很大的积极影响。 例如,将对相同存储体的不同行的访问分离至少预定数量的时钟可以有效地隐藏预充电/激活行所涉及的开销。
    • 50. 发明授权
    • Fast area-efficient multi-bit binary adder with low fan-out signals
    • 具有低扇出信号的快速区域效率的多位二进制加法器
    • US5278783A
    • 1994-01-11
    • US969124
    • 1992-10-30
    • John H. Edmondson
    • John H. Edmondson
    • G06F7/50G06F7/508
    • G06F7/508G06F2207/5063
    • A carry look-ahead adder obtains high speed with minimum gate fan-in and a regular array of area-efficient logic cells in a datapath by including a first row of propagate-generate bit cells, a second row of block-propagate bit cells generating a hierarchy of block-propagate and block-generate bits, a third row of carry bit cells: and a bottom level of sum bit cells. The second row of block-propagate bit cells supply the block-propagate and block-generate bits to the first carry bit cells in chained segments of carry bit cells. In a preferred embodiment for a 32-bit complementary metal-oxide semiconductor (CMOS) adder, the logic gates are limited to a fan-in of three, and the block-propagate bit cells in the second row are interconnected to form two binary trees, each including fifteen cells, and the carry cells are chained in segments including up to four cells. In general, the interconnections between the block-propagate bit cells are derived from a graph which is optimized to meet the constraints of fast static complementary metal-oxide-semiconductor (CMOS) circuit design: low fan-out and small capacitance load on most signals. Sufficient gain stages are present in the binary trees to build-up to a large drive capability where the large drive capability is needed.
    • 携带前视加法器通过包括第一行传播生成位单元,在数据通路中获得具有最小栅极扇入和区域有效逻辑单元的规则阵列的高速度,第二行块传播位单元生成 块传播和块生成位的层次,进位位单元的第三行和和位位单元的底层。 第二行块传播位单元将块传播和块生成位提供给进位位单元的链接段中的第一进位位单元。 在用于32位互补金属氧化物半导体(CMOS)加法器的优选实施例中,逻辑门被限制为三个扇形,并且第二行中的块传播位单元被互连以形成两个二叉树 ,每个包括十五个小区,并且携带单元被链接到包括多达四个小区的段中。 通常,块传播比特单元之间的互连是从图形中导出的,该曲线被优化以满足快速静态互补金属氧化物半导体(CMOS)电路设计的约束:大多数信号的低扇出和小电容负载 。 在二叉树中存在足够的增益级,以构建需要大驱动能力的大型驱动能力。