会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 3. 发明专利
    • CACHE AFFINITY SCHEDULER
    • GB2258933A
    • 1993-02-24
    • GB9217740
    • 1992-08-19
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08
    • A computing system 50 includes N computing engines having N cache memories connected by a system bus 12. The computing system includes a global run queue (fig 5 not shown), an FPA global run queue, and N affinity run queues, each associated with one computing engine. When a process first becomes runnable, it is typically attached to one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. Each time a process becomes runnable, the scheduler estimates the remaining cache context for the process in the cache of the engine. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure, therefore, increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity e.g. engines on which the process was previously running, or will run on the best available engine when there is insufficient cache context.
    • 4. 发明专利
    • DE4227345C2
    • 1999-01-14
    • DE4227345
    • 1992-08-18
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08G06F15/76G06F9/38G06F9/46
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.
    • 5. 发明专利
    • Cache affinity scheduler
    • GB2258933B
    • 1995-10-04
    • GB9217740
    • 1992-08-19
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08G06F9/46
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.
    • 6. 发明专利
    • DE4227345A1
    • 1993-02-25
    • DE4227345
    • 1992-08-18
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.