2 Citations (Scopus)

Abstract

In many computing and networking applications, arriving tasks have to be routed to one of many servers, with the goal of minimizing queueing delays. When the number of processors is very large, a popular routing algorithm works as follows: Select two servers at random and route an arriving task to the least loaded of the two. It is well known that this algorithm dramatically reduces queueing delays compared to an algorithm, which routes to a single randomly selected server. In recent cloud computing applications, it has been observed that even sampling two queues per arriving task can be expensive and can even increase delays due to messaging overhead. So there is an interest in reducing the number of sampled queues per arriving task. In this paper, we show that the number of sampled queues can be dramatically reduced by using the fact that tasks arrive in batches (called jobs). In particular, we sample a subset of the queues such that the size of the subset is slightly larger than the batch size (thus, on average, we only sample slightly more than one queue per task). Once a random subset of the queues is sampled, we propose a new load-balancing method called batch-filling to attempt to equalize the load among the sampled servers. We show that, asymptotically, our algorithm dramatically reduces the sample complexity compared to previously proposed algorithms.

Original languageEnglish (US)
Pages (from-to)692-722
Number of pages31
JournalMathematics of Operations Research
Volume42
Issue number3
DOIs
StatePublished - Aug 1 2017

Fingerprint

Load Balancing
Resource allocation
Queue
Servers
Server
Batch
Queueing
Subset
Routing algorithms
Cloud computing
Routing Algorithm
Sampling
Cloud Computing
Networking
Load balancing
Computing

Keywords

  • Batch arrival
  • Interchange of the limits
  • Kurtz's theorem
  • Mean-field analysis
  • Randomized load balancing
  • Sample complexity

ASJC Scopus subject areas

  • Mathematics(all)
  • Computer Science Applications
  • Management Science and Operations Research

Cite this

The power of slightly more than one sample in randomized load balancing. / Ying, Lei; Srikant, R.; Kang, Xiaohan.

In: Mathematics of Operations Research, Vol. 42, No. 3, 01.08.2017, p. 692-722.

Research output: Contribution to journalArticle

Ying, Lei ; Srikant, R. ; Kang, Xiaohan. / The power of slightly more than one sample in randomized load balancing. In: Mathematics of Operations Research. 2017 ; Vol. 42, No. 3. pp. 692-722.
@article{36859e7f188b413db72d9858b3d4011d,
title = "The power of slightly more than one sample in randomized load balancing",
abstract = "In many computing and networking applications, arriving tasks have to be routed to one of many servers, with the goal of minimizing queueing delays. When the number of processors is very large, a popular routing algorithm works as follows: Select two servers at random and route an arriving task to the least loaded of the two. It is well known that this algorithm dramatically reduces queueing delays compared to an algorithm, which routes to a single randomly selected server. In recent cloud computing applications, it has been observed that even sampling two queues per arriving task can be expensive and can even increase delays due to messaging overhead. So there is an interest in reducing the number of sampled queues per arriving task. In this paper, we show that the number of sampled queues can be dramatically reduced by using the fact that tasks arrive in batches (called jobs). In particular, we sample a subset of the queues such that the size of the subset is slightly larger than the batch size (thus, on average, we only sample slightly more than one queue per task). Once a random subset of the queues is sampled, we propose a new load-balancing method called batch-filling to attempt to equalize the load among the sampled servers. We show that, asymptotically, our algorithm dramatically reduces the sample complexity compared to previously proposed algorithms.",
keywords = "Batch arrival, Interchange of the limits, Kurtz's theorem, Mean-field analysis, Randomized load balancing, Sample complexity",
author = "Lei Ying and R. Srikant and Xiaohan Kang",
year = "2017",
month = "8",
day = "1",
doi = "10.1287/moor.2016.0823",
language = "English (US)",
volume = "42",
pages = "692--722",
journal = "Mathematics of Operations Research",
issn = "0364-765X",
publisher = "INFORMS Inst.for Operations Res.and the Management Sciences",
number = "3",

}

TY - JOUR

T1 - The power of slightly more than one sample in randomized load balancing

AU - Ying, Lei

AU - Srikant, R.

AU - Kang, Xiaohan

PY - 2017/8/1

Y1 - 2017/8/1

N2 - In many computing and networking applications, arriving tasks have to be routed to one of many servers, with the goal of minimizing queueing delays. When the number of processors is very large, a popular routing algorithm works as follows: Select two servers at random and route an arriving task to the least loaded of the two. It is well known that this algorithm dramatically reduces queueing delays compared to an algorithm, which routes to a single randomly selected server. In recent cloud computing applications, it has been observed that even sampling two queues per arriving task can be expensive and can even increase delays due to messaging overhead. So there is an interest in reducing the number of sampled queues per arriving task. In this paper, we show that the number of sampled queues can be dramatically reduced by using the fact that tasks arrive in batches (called jobs). In particular, we sample a subset of the queues such that the size of the subset is slightly larger than the batch size (thus, on average, we only sample slightly more than one queue per task). Once a random subset of the queues is sampled, we propose a new load-balancing method called batch-filling to attempt to equalize the load among the sampled servers. We show that, asymptotically, our algorithm dramatically reduces the sample complexity compared to previously proposed algorithms.

AB - In many computing and networking applications, arriving tasks have to be routed to one of many servers, with the goal of minimizing queueing delays. When the number of processors is very large, a popular routing algorithm works as follows: Select two servers at random and route an arriving task to the least loaded of the two. It is well known that this algorithm dramatically reduces queueing delays compared to an algorithm, which routes to a single randomly selected server. In recent cloud computing applications, it has been observed that even sampling two queues per arriving task can be expensive and can even increase delays due to messaging overhead. So there is an interest in reducing the number of sampled queues per arriving task. In this paper, we show that the number of sampled queues can be dramatically reduced by using the fact that tasks arrive in batches (called jobs). In particular, we sample a subset of the queues such that the size of the subset is slightly larger than the batch size (thus, on average, we only sample slightly more than one queue per task). Once a random subset of the queues is sampled, we propose a new load-balancing method called batch-filling to attempt to equalize the load among the sampled servers. We show that, asymptotically, our algorithm dramatically reduces the sample complexity compared to previously proposed algorithms.

KW - Batch arrival

KW - Interchange of the limits

KW - Kurtz's theorem

KW - Mean-field analysis

KW - Randomized load balancing

KW - Sample complexity

UR - http://www.scopus.com/inward/record.url?scp=85026870697&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85026870697&partnerID=8YFLogxK

U2 - 10.1287/moor.2016.0823

DO - 10.1287/moor.2016.0823

M3 - Article

AN - SCOPUS:85026870697

VL - 42

SP - 692

EP - 722

JO - Mathematics of Operations Research

JF - Mathematics of Operations Research

SN - 0364-765X

IS - 3

ER -