Results 1 - 10
of
1,221
Paired Gang Scheduling
- IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS
, 2003
"... Conventional gang scheduling has the disadvantage that when processes perform I/O or blocking communication, their processors remain idle, because alternative processes cannot be run independently of their own gangs. To alleviate this problem we suggest a slight relaxation of this rule: match gangs ..."
Abstract
-
Cited by 52 (14 self)
- Add to MetaCart
that make heavy use of the CPU with gangs that make light use of the CPU (presumably due to I/O or communication activity), and schedule such pairs together, allowing the local scheduler on each node to select either of the two processes at any instant. As I/O-intensive gangs make light use of the CPU
Packing Schemes for Gang Scheduling
- In Job Scheduling Strategies for Parallel Processing
, 1996
"... . Jobs that do not require all processors in the system can be packed together for gang scheduling. We examine accounting traces from several parallel computers to show that indeed many jobs have small sizes and can be packed together. We then formulate a number of such packing algorithms, and e ..."
Abstract
-
Cited by 116 (22 self)
- Add to MetaCart
. Jobs that do not require all processors in the system can be packed together for gang scheduling. We examine accounting traces from several parallel computers to show that indeed many jobs have small sizes and can be packed together. We then formulate a number of such packing algorithms
Gang Scheduling Performance Benefits for Fine-Grain Synchronization
- Journal of Parallel and Distributed Computing
, 1992
"... Abstract Multiprogrammed multiprocessors executing fine-grained parallel programs appear to require new scheduling policies. A promising new idea is gang scheduling, where a set of threads are scheduled to execute simultaneously on a set of processors. This has the intuitive appeal of supplying the ..."
Abstract
-
Cited by 157 (12 self)
- Add to MetaCart
Abstract Multiprogrammed multiprocessors executing fine-grained parallel programs appear to require new scheduling policies. A promising new idea is gang scheduling, where a set of threads are scheduled to execute simultaneously on a set of processors. This has the intuitive appeal of supplying
Improved Utilization and Responsiveness with Gang Scheduling.
- Scheduling Strategies for Parallel Processing,
, 1997
"... Abstract Most commercial multicomputers use space-slicing schemes in which each scheduling decision has an unknown impact on the future: should a job be scheduled, risking that it will block other larger jobs later, or should the processors be left idle for now in anticipation of future arrivals? T ..."
Abstract
-
Cited by 119 (20 self)
- Add to MetaCart
? This dillema is solved by using gang scheduling, because then the impact of each decision is limited to its time slice, and future arrivals can be accommodated in other time slices. This added flexibility y is shown to improve overall system utilization and responsiveness. Empirical evidence from using gang
Hedera: Dynamic flow scheduling for data center networks
- In Proc. of Networked Systems Design and Implementation (NSDI) Symposium
, 2010
"... Today’s data centers offer tremendous aggregate bandwidth to clusters of tens of thousands of machines. However, because of limited port densities in even the highest-end switches, data center topologies typically consist of multi-rooted trees with many equal-cost paths between any given pair of hos ..."
Abstract
-
Cited by 223 (7 self)
- Add to MetaCart
Today’s data centers offer tremendous aggregate bandwidth to clusters of tens of thousands of machines. However, because of limited port densities in even the highest-end switches, data center topologies typically consist of multi-rooted trees with many equal-cost paths between any given pair
Gang Scheduling with Memory Considerations
- in Proc. of the 14th Intl. Parallel and Distributed Processing Symp., 2000
"... A major problem with time slicing on parallel machines is memory pressure, as the resulting paging activity damages the synchronism among a job’s processes. An alternative is to impose admission controls, and only admit jobs that fit into the available memory. Despite suffering from delayed executio ..."
Abstract
-
Cited by 59 (2 self)
- Add to MetaCart
A major problem with time slicing on parallel machines is memory pressure, as the resulting paging activity damages the synchronism among a job’s processes. An alternative is to impose admission controls, and only admit jobs that fit into the available memory. Despite suffering from delayed execution, this leads to better overall performance by preventing the harmful effects of paging and thrashing. 1.
Dynamic Critical-Path Scheduling: An Effective Technique for Allocating Task Graphs to Multiprocessors
- IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS
, 1996
"... In this paper, we propose a static scheduling algorithm for allocating task graphs to fullyconnected multiprocessors. We discuss six recently reported scheduling algorithms and show that they possess one drawback or the other which can lead to poor performance. The proposed algorithm, which is calle ..."
Abstract
-
Cited by 163 (16 self)
- Add to MetaCart
In this paper, we propose a static scheduling algorithm for allocating task graphs to fullyconnected multiprocessors. We discuss six recently reported scheduling algorithms and show that they possess one drawback or the other which can lead to poor performance. The proposed algorithm, which
An approximation algorithm for the generalized assignment problem
, 1993
"... The generalized assignment problem can be viewed as the following problem of scheduling parallel machines with costs. Each job is to be processed by exactly one machine; processing job j on machine i requires time pif and incurs a cost of c,f, each machine / is available for 7", time units, ..."
Abstract
-
Cited by 200 (5 self)
- Add to MetaCart
The generalized assignment problem can be viewed as the following problem of scheduling parallel machines with costs. Each job is to be processed by exactly one machine; processing job j on machine i requires time pif and incurs a cost of c,f, each machine / is available for 7", time units
Reconfigurable Gang Scheduling Algorithm
"... Using a single traditional gang scheduling algorithm cannot provide the best performance for all workloads and parallel architectures. A solution for this problem is the use of an algorithm that is capable of dynamically changing its form (configuration) into a more appropriate one, according to env ..."
Abstract
-
Cited by 7 (4 self)
- Add to MetaCart
Using a single traditional gang scheduling algorithm cannot provide the best performance for all workloads and parallel architectures. A solution for this problem is the use of an algorithm that is capable of dynamically changing its form (configuration) into a more appropriate one, according
Bounds on Gang Service Scheduling
, 1999
"... We define the bounds of two gang service algorithms, namely time balanced and instruction balanced, using interactivity and throughput based metrics. The workload is composed of V-RAM jobs, and the architectural model chosen is LogP. We also analyze the effect of workload and space sharing strategie ..."
Abstract
- Add to MetaCart
We define the bounds of two gang service algorithms, namely time balanced and instruction balanced, using interactivity and throughput based metrics. The workload is composed of V-RAM jobs, and the architectural model chosen is LogP. We also analyze the effect of workload and space sharing
Results 1 - 10
of
1,221