会员体验
专利管家(专利管理)
工作空间(专利管理)
风险监控(情报监控)
数据分析(专利分析)
侵权分析(诉讼无效)
联系我们
交流群
官方交流:
QQ群: 891211   
微信请扫码    >>>
现在联系顾问~
热词
    • 2. 发明专利
    • CACHE AFFINITY SCHEDULER
    • GB9217740D0
    • 1992-09-30
    • GB9217740
    • 1992-08-19
    • SEQUENT COMPUTER SYSTEMS INC
    • G06F9/50G06F12/08
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.
    • 4. 发明专利
    • DE4227346A1
    • 1993-02-25
    • DE4227346
    • 1992-08-18
    • SEQUENT COMPUTER SYSTEMS INC
    • COULSON RICHARD LO'MALLEY VINCENT GSAFRANEK ROBERT J
    • G06F1/30G06F1/32G06F13/28G06F13/42
    • An apparatus and a method facilitate the sharing of a SCSI address ID between a SCSI initiator and a target device. The SCSI standard does not prohibit ID sharing under certain operating conditions. The shared ID is preferably used by a monitor device 70 that controls the power of disk drives 30A, 30B, 30C on the SCSI bus, detects their status and configuration, and monitors environmental conditions within a mass storage system cabinet. Power to individual disk drives is switched under control of the computer operating system so that start-up surge current to the disk drives is limited. This allows defective drives to be powered down and replaced without powering down the remaining good drives. SCSI termination power distribution is improved so that connector reversals can be detected and resulting current surges limited. Termination power control and status is under program control, 28, 100. The reliability, utility, and cost-effectiveness of large SCSI bus-based mass storage systems is improved through the use of this invention.
    • 7. 发明专利
    • DE4227345C2
    • 1999-01-14
    • DE4227345
    • 1992-08-18
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08G06F15/76G06F9/38G06F9/46
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.
    • 8. 发明专利
    • Cache affinity scheduler
    • GB2258933B
    • 1995-10-04
    • GB9217740
    • 1992-08-19
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08G06F9/46
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.
    • 9. 发明专利
    • DE4227345A1
    • 1993-02-25
    • DE4227345
    • 1992-08-18
    • SEQUENT COMPUTER SYSTEMS INC
    • VALENCIA ANDREW J
    • G06F9/50G06F12/08
    • A computing system (50) includes N number of symmetrical computing engines having N number of cache memories joined by a system bus (12). The computing system includes a global run queue (54), an FPA global run queue, and N number of affinity run queues (58). Each engine is associated with one affinity run queue, which includes multiple slots. When a process first becomes runnable, it is typically attached one of the global run queues. A scheduler allocates engines to processes and schedules the processes to run on the basis of priority and engine availability. An engine typically stops running a process before it is complete. When the process becomes runnable again the scheduler estimates the remaining cache context for the process in the cache of the engine. The scheduler uses the estimated amount of cache context in deciding in which run queue a process is to be enqueued. The process is enqueued to the affinity run queue of the engine when the estimated cache context of the process is sufficiently high, and is enqueued onto the global run queue when the cache context is sufficiently low. The procedure increases computing system performance and reduces bus traffic because processes will run on engines having sufficient cache affinity, but will also run on the best available engine when there is insufficient cache context.