Results 1  10
of
40
Speed Scaling Functions for Flow Time Scheduling based on Active Job Count
"... Abstract. We study online scheduling to minimize flow time plus energy usage in the dynamic speed scaling model. We devise new speed scaling functions that depend on the number of active jobs, replacing the existing speed scaling functions in the literature that depend on the remaining work of activ ..."
Abstract

Cited by 46 (12 self)
 Add to MetaCart
(Show Context)
Abstract. We study online scheduling to minimize flow time plus energy usage in the dynamic speed scaling model. We devise new speed scaling functions that depend on the number of active jobs, replacing the existing speed scaling functions in the literature that depend on the remaining work of active jobs. The new speed functions are more stable and also more efficient. They can support better job selection strategies to improve the competitive ratios of existing algorithms [5,8], and, more importantly, to remove the requirement of extra speed. These functions further distinguish themselves from others as they can readily be used in the nonclairvoyant model (where the size of a job is only known when the job finishes). As a first step, we study the scheduling of batched jobs (i.e., jobs with the same release time) in the nonclairvoyant model and present the first competitive algorithm for minimizing flow time plus energy (as well as for weighted flow time plus energy); the performance is close to optimal. 1
Poweraware scheduling for makespan and flow
 In Proc. 18th Annual ACM Symp. Parallelism in Algorithms and Architectures
, 2006
"... We consider offline scheduling algorithms that incorporate speed scaling to address the bicriteria problem of minimizing energy consumption and a scheduling metric. For makespan, we give a lineartime algorithm to compute all nondominated solutions for the general uniprocessor problem and a fast a ..."
Abstract

Cited by 44 (1 self)
 Add to MetaCart
(Show Context)
We consider offline scheduling algorithms that incorporate speed scaling to address the bicriteria problem of minimizing energy consumption and a scheduling metric. For makespan, we give a lineartime algorithm to compute all nondominated solutions for the general uniprocessor problem and a fast arbitrarilygood approximation for multiprocessor problems when every job requires the same amount of work. We also show that the multiprocessor problem becomes NPhard when jobs can require different amounts of work. For total flow, we show that the optimal flow corresponding to a particular energy budget cannot be exactly computed on a machine supporting exact real arithmetic, including the extraction of roots. This hardness result holds even when scheduling equalwork jobs on a uniprocessor. We do, however, extend previous work by Pruhs et al. to give an arbitrarilygood approximation for scheduling equalwork jobs on a multiprocessor. 1
Improved bounds for speed scaling in devices obeying the cuberoot rule
 Proc. 36th International Colloqium on Automata, Languages and Programming
, 2009
"... ..."
(Show Context)
A.: The bell is ringing in speedscaled multiprocessor scheduling
 In: Proceedings of ACM Symposium on Parallelism in Algorithms and Architectures (SPAA
, 2009
"... This paper investigates the problem of scheduling jobs on multiple speedscaled processors without migration, i.e., we have constant α> 1 such that running a processor at speed s results in energy consumption s α per time unit. We consider the general case where each job has a monotonously increas ..."
Abstract

Cited by 20 (0 self)
 Add to MetaCart
(Show Context)
This paper investigates the problem of scheduling jobs on multiple speedscaled processors without migration, i.e., we have constant α> 1 such that running a processor at speed s results in energy consumption s α per time unit. We consider the general case where each job has a monotonously increasing cost function that penalizes delay. This includes the so far considered cases of deadlines and flow time. For any type of delay cost functions, we obtain the following results: Any βapproximation algorithm for a single processor yields a randomized βBαapproximation algorithm for multiple processors without migration, where Bα is the αth Bell number, that is, the number of partitions of a set of size α. Analogously, we show that any βcompetitive online algorithm for a single processor yields a βBαcompetitive online algorithm for multiple processors without migration. Finally, we show that any βapproximation algorithm for multiple processors with migration yields a deterministic βBαapproximation algorithm for multiple processors without migration. These facts improve several approximation ratios and lead to new results. For instance, we obtain the first constant factor online and offline approximation algorithm for multiple processors without migration for arbitrary release times, deadlines, and job sizes.
Competitive Nonmigratory Scheduling for Flow Time and Energy
 SPAA'08
, 2008
"... Energy usage has been an important concern in recent research on online scheduling. In this paper we extend the study of the tradeoff between flow time and energy from the singleprocessor setting [8, 6] to the multiprocessor setting. Our main result is an analysis of a simple nonmigratory online ..."
Abstract

Cited by 19 (7 self)
 Add to MetaCart
(Show Context)
Energy usage has been an important concern in recent research on online scheduling. In this paper we extend the study of the tradeoff between flow time and energy from the singleprocessor setting [8, 6] to the multiprocessor setting. Our main result is an analysis of a simple nonmigratory online algorithm called CRR (classified round robin) on m ≥ 2 processors, showing that its flow time plus energy is within O(1) times of the optimal nonmigratory offline algorithm, when the maximum allowable speed is slightly relaxed. This result still holds even if the comparison is made against the optimal migratory offline algorithm (the competitive ratio increases by a factor of 2.5). As a special case, our work also contributes to the traditional online flowtime scheduling. Specifically, for minimizing flow time only, CRR can yield a competitive ratio one or even arbitrarily smaller than one, when using sufficiently faster processors. Prior to our work, similar result is only known for online algorithms that needs migration [21, 23], while the best nonmigratory result can achieve an O(1) competitive ratio [14]. The above result stems from an interesting observation that there always exists some optimal migratory schedule S that can be converted (in an offline sense) to a nonmigratory schedule S ′ with a moderate increase in flow time plus energy. More importantly, this nonmigratory schedule always dispatches jobs in the same way as CRR.
Average rate speed scaling
 In Latin American Theoretical Informatics Symposium, 2008. Nikhil Bansal, HoLeung Chan, Kirk Pruhs, and Dmitriy RogozhnikovKatz. Improved
, 2007
"... Speed scaling is a power management technique that involves dynamically changing the speed of a processor. This gives rise to dualobjective scheduling problems, where the operating system both wants to conserve energy and optimize some Quality of Service (QoS) measure of the resulting schedule. Yao ..."
Abstract

Cited by 15 (7 self)
 Add to MetaCart
(Show Context)
Speed scaling is a power management technique that involves dynamically changing the speed of a processor. This gives rise to dualobjective scheduling problems, where the operating system both wants to conserve energy and optimize some Quality of Service (QoS) measure of the resulting schedule. Yao, Demers, and Shenker [4] considered the problem where the QoS constraint is deadline feasibility and the objective is to minimize the energy used. They proposed an online speed scaling algorithm Average Rate (AVR) that runs each job at a constant speed between its release and its deadline. They showed that the competitive ratio of AVR is at most (2α) α /2 if a processor running at speed s uses power s α. We show the competitive ratio of AVR is at least ((2 − δ)α) α /2, where δ is a function of α that approaches zero as α approaches infinity. This shows that the competitive analysis of AVR by Yao, Demers, and Shenker is essentially tight, at least for large α. We also give an alternative proof that the competitive ratio of AVR is at most (2α) α /2 using a potential function argument. We believe that this analysis is significantly simpler and more elementary than the original analysis of AVR in [4]. 1
Deadline Scheduling and Power Management for Speed Bounded Processors
"... Energy consumption has become an important issue in the study of processor scheduling. Energy reduction can be achieved by allowing a processor to vary the speed dynamically (dynamic speed scaling) [2–4, 7, 10] or to enter a sleep state [1, 5, 8]. In the past, these two mechanisms are often studied ..."
Abstract

Cited by 13 (1 self)
 Add to MetaCart
(Show Context)
Energy consumption has become an important issue in the study of processor scheduling. Energy reduction can be achieved by allowing a processor to vary the speed dynamically (dynamic speed scaling) [2–4, 7, 10] or to enter a sleep state [1, 5, 8]. In the past, these two mechanisms are often studied separately. It is indeed natural to consider an integrated model in which a
Energy Efficient Deadline Scheduling in Two Processor Systems
"... Abstract. The past few years have witnessed different scheduling algorithms for a processor that can manage its energy usage by scaling dynamically its speed. In this paper we attempt to extend such work to the twoprocessor setting. Specifically, we focus on deadline scheduling and study online alg ..."
Abstract

Cited by 12 (0 self)
 Add to MetaCart
(Show Context)
Abstract. The past few years have witnessed different scheduling algorithms for a processor that can manage its energy usage by scaling dynamically its speed. In this paper we attempt to extend such work to the twoprocessor setting. Specifically, we focus on deadline scheduling and study online algorithms for two processors with an objective of maximizing the throughput, while using the smallest possible energy. The motivation comes from the fact that dualcore processors are getting common nowadays. Our first result is a new analysis of the energy usage of the speed function OA [15,4,8] with respect to the optimal twoprocessor schedule. This immediately implies a trivial twoprocessor algorithm that is 16competitive for throughput and O(1)competitive for energy. A more interesting result is a new online strategy for selecting jobs for the two processors. Together with OA, it improves the competitive ratio for throughput from 16 to 3, while increasing that for energy by a factor of 2. Note that even if the energy usage is not a concern, no algorithm can be better than 2competitive with respect to throughput. 1
Algorithms for dynamic speed scaling
 In STACS 2011, volume 9 of LIPIcs. Schloss Dagstuhl  LeibnizZentrum fuer Informatik
, 2011
"... Many modern microprocessors allow the speed/frequency to be set dynamically. The general goal is to execute a sequence of jobs on a variablespeed processor so as to minimize energy consumption. This paper surveys algorithmic results on dynamic speed scaling. We address settings where (1) jobs have ..."
Abstract

Cited by 11 (0 self)
 Add to MetaCart
Many modern microprocessors allow the speed/frequency to be set dynamically. The general goal is to execute a sequence of jobs on a variablespeed processor so as to minimize energy consumption. This paper surveys algorithmic results on dynamic speed scaling. We address settings where (1) jobs have strict deadlines and (2) job flow times are to be minimized.
Nonclairvoyant speed scaling for batched parallel jobs on multiprocessors
 In CF
, 2009
"... Energy consumption and heat dissipation have become key considerations for modern high performance computer systems. In this paper, we focus on nonclairvoyant speed scaling to minimize flow time plus energy for batched parallel jobs on multiprocessors. We consider a common scenario where the to ..."
Abstract

Cited by 6 (5 self)
 Add to MetaCart
(Show Context)
Energy consumption and heat dissipation have become key considerations for modern high performance computer systems. In this paper, we focus on nonclairvoyant speed scaling to minimize flow time plus energy for batched parallel jobs on multiprocessors. We consider a common scenario where the total power consumption cannot exceed a given budget and the power consumed on each processor is sα when running at speed s. Extending the Equi processor allocation policy, we propose two algorithms: UEqui and NEqui, which use respectively a uniformspeed and a nonuniform speed scaling function for the allocated processors. Using competitive analysis, we show that UEqui is O(P (α−1)/α 2)competitive for flow time plus energy, and NEqui is O ( α lnP)competitive for the same metric when given sufficient power, where P is the total number of processors. Our simulation results confirm that UEqui and NEqui achieve better performance than a straightforward fixedspeed Equi strategy. Moreover, moderate power constraint does not significantly affect the performance of our algorithms.