Научная статья на тему 'Resource allocation in cloud computing via optimal control to queuing systems'

Resource allocation in cloud computing via optimal control to queuing systems Текст научной статьи по специальности «Математика»

CC BY
116
28
i Надоели баннеры? Вы всегда можете отключить рекламу.
Ключевые слова
CLOUD COMPUTING / MULTIPLE QUEUEING SYSTEM / MULTIPLE JOB CLASSES / STOCHASTIC CONTROL POLICY / ОБЛАЧНЫЕ ВЫЧИСЛЕНИЯ / СИСТЕМА МНОЖЕСТВЕННЫХ ОЧЕРЕДЕЙ / НЕСКОЛЬКО КЛАССОВ РАБОТЫ / СТОХАСТИЧЕСКАЯ ПОЛИТИКА УПРАВЛЕНИЯ

Аннотация научной статьи по математике, автор научной работы — Madankan A., Delavarkhalafi A., Karbassi S.M., Adibnia F.

We consider resource allocation problem in the cloud computing. We use queuing model to model the process of entering into the cloud and to schedule and to serve incoming jobs. In this paper, the main problem is to allocate resources in the queuing systems as a general optimization problem for controlled Markov process with finite state space. For this purpose, we study a model of cloud computing where the arrival jobs follow a stochastic process. We reduce this problem to a routing problem. In the case of minimizing, cost is given as a mixture of an average queue length and number of lost jobs. We use dynamic programming approach. Finally, we obtain the explicit form of the optimal control by the Bellman equation.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

РАСПРЕДЕЛЕНИЕ РЕСУРСОВ В ОБЛАЧНЫХ ВЫЧИСЛЕНИЯХ С ПОМОЩЬЮ ОПТИМАЛЬНОГО УПРАВЛЕНИЯ СИСТЕМАМИ ПЕРЕДАЧИ

Рассматривается задача выделения ресурсов в облачных вычислениях. Мы используем модель очередей для моделирования процесса входа в облако, а также для планирования и обслуживания входящих заданий. Основной задачей, с которой мы сталкиваемся в данной статье, является задача распределения ресурсов в системах массового обслуживания как общая задача оптимизации для управляемого марковского процесса с конечным пространством состояний. Для этой цели мы изучаем модель облачных вычислений, в которой задания по прибытию следуют случайному процессу. Мы сводим эту задачу к задаче маршрутизации. В случае минимизации стоимость выражается через среднюю длину очереди и количество потерянных заданий. Мы используем подход динамического программирования и получаем явную форму оптимального управления по уравнению Беллмана.

Текст научной работы на тему «Resource allocation in cloud computing via optimal control to queuing systems»

MSC 49J20, 93E20, 68M20, 90B22, 68Q85

DOI: 10.14529/ mmp190405

RESOURCE ALLOCATION IN CLOUD COMPUTING VIA OPTIMAL CONTROL TO QUEUING SYSTEMS

A. Madankan1, A. Delavarkhalafi1, S.M. Karbassi1, F. Adibnia1

1Yazd University, Yazd, Iran

E-mails: [email protected], [email protected], [email protected], [email protected]

We consider resource allocation problem in the cloud computing. We use queuing model to model the process of entering into the cloud and to schedule and to serve incoming jobs. In this paper, the main problem is to allocate resources in the queuing systems as a general optimization problem for controlled Markov process with finite state space. For this purpose, we study a model of cloud computing where the arrival jobs follow a stochastic process. We reduce this problem to a routing problem. In the case of minimizing, cost is given as a mixture of an average queue length and number of lost jobs. We use dynamic programming approach. Finally, we obtain the explicit form of the optimal control by the Bellman equation.

Keywords: cloud computing; multiple queueing system; multiple job classes; stochastic control policy.

Introduction

Resource Allocating (RA) is the process of assigning available resources to the applicants. There are several techniques in parallel computing and sharing the computing resources where cloud computing is the latest technique that shares resources. Cloud computing provides distributed computing over a network that allocates resources to the end-users. Since users are distributed geographically over a wide area, the resources are also distributed. Therefore the cloud management needs to manage resources to allocate resources efficiently. A comprehensive survey on the cloud computing can be found in [1-3].

In [4], resource allocation strategy is a considered as an integrating cloud provider activity for utilizing and allocating resources of cloud environment and a classification of resource allocation strategy is presented. The paper [5] analyses the architecture of the cloud with considering the Quality of Service (QoS). Also, a sequence of M/M/1 and M/M/m queues is considered to model the cloud architecture. The paper [6] proposes different workload types with different characteristics that should be supported by cloud computing, but there is no single solution where it can allocate resources to all imaginable demands optimally.

Stolyar A. [7] uses resource pooling to a generalized switch model, where it was used to study the heavy traffic optimality of the Max-Weight algorithm. Stolyar A. considered a scaled version of queue lengths and time to obtain these results, where it leaded to a regulated Brownian motion. In [8], the authors propose another method in unscaled time for heavy traffic optimality. In addition, this method directly obtains heavy-traffic optimality in steady state. The problem reduces to a routing problem, which is well studied in [9-13].

The optimization and the parameter for evaluating the service in cloud computing are studied in [14]. Also, [14] uses a queuing model to study the performance of services and

develops a method to optimize it. The paper [15] also considers queuing system for cloud computing and studies a routing problem regarding to reduced workload, response time and the average queue length.

In [16], authors use a stochastic model for load balancing and scheduling in cloud computing clusters where jobs arrive according to a stochastic process and request resources such as CPU, memory, and storage space. They show that the performance of JSQ1 routing and two-choice routing algorithms with Max-Weight scheduling policy have optimal throughput. In [17], the authors introduce a cloud resource allocating algorithm called CRAA/FA2, which creates a market for cloud resources and makes the resource agents and service agents bargain in that market.

Winston [18] proves that under some conditions, the JSQ strategy is optimal. These conditions are as follows: (1) we do not have any prior knowledge about the job, (2) the jobs are not allowed to preempt, (3) the router sends job immediately when it comes, (4) First-Come-First-Served policy is the only policy for each servers, (5) the job size has exponential distribution. Although, the JSQ is optimal under these conditions, the model that we considered does not meet one or more conditions, therefore we need to study a methodology over these conditions.

There are several types of networks, such as computer multiprocessor networks and communications data networks in the queuing model. Queues are the main part of various network components, such as the input and output buffers of packets. We often want to find the optimal performance of queuing systems, for example, queue length, waiting time, workload, and probabilities of certain states. Since the performance parameters are nonlinear as functions of the arrival and service rates , finding the optimal performance of queues is a difficult problem, whereas efficient control to work loud is one of the highlighted problems of the computer networks.

In this paper, we consider the problem of providing resources in cloud computing. Costumers request resources and set the size of their request, such as a memory, processor power, ..., upon the arrival. The entry point (EP) of the cloud computing puts these requests into a queue and then distribute them to a processing server (PS). These PSs have several kinds of resources where there is a limit of each kind in each PS. Due to this limitation, we have finite number of jobs of each kind on a PS. The simplest architecture of cloud is parallel queues with a router at the front to schedule incoming jobs to servers (see Fig.1).

Fig. 1. Queuing model of cloud computing

1 Join-the shortest-queue

2Cloud Resource Allocating Algorithm via Fitness-enabled Auction

For a wide class of criteria, we demonstrate that the optimal control problem can be reduced to the solution of the system of ordinary differential equations. Moreover, the optimal control exists within the class of Markov strategies and therefore can be calculated for each possible state of controlled Markov chain. The existence of the optimal solution was proved in [19]. The papers [20] and [21] consider the general optimization problem for jump Markov process and propose the reduction to a problem with complete information for a wide class of optimal control problems.

1. The Mathematical Modelling and State Equation

In this paper, the considered system is a multi-server queuing network where there is a single entry point (Entry Server) at the front. This entry server works as a load balancer and sends the user job requests to one of the processing servers PSs, where s = 1,..., S. The purpose of this server is to chose a PSs to deliver the job requests.

A processing server PSs can be a node, core or processor representing the physical computational resources of cloud architecture where the services are computed. The selected processing server executes all services that are demanded by an assigned job. The processing servers are identical and are modelled as a M/M/m queuing system. The system has S processing servers (PS) indexed by s = 1,..., S.

Let L be the number of different kinds of resources of each PSs. These resources can be processing power, disk space, memory, etc. Let Rl>s be the amount of the resource l of PSs for l = 1,...,L and s = 1,...,S. We suppose that each job needs rl, l = 1,...,L, units of the resource l to complete its service.

Let r(t) be the set of jobs that arrive at the time slot t and 7(t) = ier(i) Di be the size of r(t) which means the total time needed for the jobs in r. It is clear that 7(t) is a stochastic process which is i.i.d. over time slots, with E[7(t)] = A.

In each time slot, the Entering Server disputes the arrivals to one of the processing servers. In the time slot t, we use qs to show the queue length at the server s. For a given server m, denote by Ys the state of the queue. If we define Yi as the size of the i-th job at the server s then the total backlogged job size is given by qs(t) = Yl i Y/ which is a function of the state Ys.

In a Markov chain, one can control arrivals to each queues. This control is called load restriction. For load restriction, control parameter us is a probability of accepting the new arrival to the queue of the server s. Now, let all processes be defined on a probability space (Q, F, P) with right-continuous set of complete a-algebras F = {Ft}t>0 generated by Xs(t). Let the process {Xs(t)} be the random variable of queue length of queue of the server s. Similar to [22] and [23], we assume that trajectories of the random processes Xs(t) are piecewise right continuous nonnegative jump-like Markov processes and have finite left limits. Suppose that the state is described by a vector Xs. We write the standard problem statement equation as follows:

(1)

where

Y(t) if s = s*(t), 0 otherwise,

and s* is the server that router chooses it.

2. Optimal Control Problems 2.1. Optimality Criterion

We assume that the classical performance criterion takes the form

tf(a) = Ea{Co(T, Xt) + i c(t, Xt, a)dt

(2)

where Ea is the expectation of its argument and the cost functions c0(.) and c(.) are two continuous nonnegative functions. In the cost functional (2) the time T is the time horizon. The optimal control problem is to find controls at £ A, where achieves the extremum of the functional (2).

Boel R. and Varaia P. [22] propose a method for using Bellman equation to control a jump-like process. We follow the model presented in [22,24] to drive Bellman equations.

The function V(t,x) is called the value function of the control problem. The function V(t,x) satisfies the Hamilton-Jacobi-Bellman (HJB) equation (also called the optimality equation)

rT

V (t, x) = min EN co(T,Xt )+/ c(s, Xs, aa)ds|Xt- = x

{as&A}t<s<r { Jt

(3)

One can find the derivation of equation (3) for a Markov queuing network in [24]. If we define the cost function's difference derivatives as at point (t, x) Y+(t, x) = V(t,x + es) — V(t,x) and Y- (t, x) = V(t,x — es) — V(t,x), then the value function V(t,x) can also be presented as the following equation:

= rmn < c(t,x,a) + ^ afcAfcT+(r, x) + ^ afcTj(r, x)l(Qfc>0)

fees

fees

where the initial condition in (4) is

V (T,x) = Co(T,x).

(4)

3. Controlling Servicing Intensity

In order to find the control to classical queuing system, let us consider the state Xt be queue length which is a jump-like Markov process. Obviously, the state Xt is a birth and death process. Our goal is to minimize the following functional:

Ea [ Xt dt,

o

(5)

Therefore, we need to find the optimal strategy akt = a(Xt)|Xt = k to optimize this functional considered as an accumulated delay functional. Regarding to this functional, the value function (4) takes the following form:

ÖV. ,

——{t,x)= mm {x + a dt a€[0,(U]

V(t, x - 1) - V(t,x) + A

V(t, x + 1) - V(t, x)

(x>0)

(6)

under the initial condition V(T, x) = 0.

o

1

We consider in the region [0,T] x W, and (t) = V(t, k). With this consideration, (6) reduces to an infinite system of ODEs with respect to the functions (t):

d7r0(t) dt dni(i) dt

A[no(t) - ni(t)],

min |ai[no(t) - ni(t)]}A[ni(t) - n(t)] + 1, «1 е[0,м]

dni(t)

= mill {йг[7Tj_i(i) - 7Tî(Î)]}A[7Tî(Î) - 7Ti+i(i)] + i, dt «i e[o,^]

(7)

where the initial conditions are ni(T) = 0 for each i.

Also, if we consider Vi(t) = ni(t) — ni-1(t) then we can represent the optimal control by the following equation:

a (t) = ^1(i>o)a-1Vi (t). And one can find the following expression for P0(t, i) in queueing theory in [25]:

(8)

Po (t, i)

e

-(A+M)r

+ (1 - p) £ P-k/21k(2r, fc=i+2

(9)

where p = A/^.

Definition 1. We use the notation ® to define the convolution of the functions A(x) and A(x) as

A(x) ® A(x) = A(x - y)A(y)dy. Jo

Remark 1. By solving the Chapman-Kolmogorov equations

1РЛг,)

AP0-i(t, i) - (^ + A)P0(r, i) + ^P0+i(r, i), o > 0, ^Pi(r,i) - APo(t, i),

under the initial condition Po(0,i) = 1(o=i), one can find that

(10)

p+(r, i) = P{Xo = i} = 1 — Po(t, i), Po(t, i) = P{Xr = 0|Xo = i}.

Theorem 1. Let XT be a birth and death process with constant birth and death rates A and ^ with respect to t = T — t. For the problem with functional (5), the optimal control is a Markov control and can be reprresented as

dit = ^1(i),

(11)

эо

t

where the value function is

<x

V (t,x) = J] ni (r)1(x=ei), (12)

i=0

and

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

J2

7Ti(i) = ir - ßT®p+(r,i) + Ay. (13)

Proof. In order to show the Laplace transform of the function A(x)

/><x

A*(s) = [A]* = e-sx A(x)dx,

0

we use the symbol *. We use backward method of dynamic programming to find the optimal control. Let us define a new time variable t as t = T — t which takes values from the time T to 0. We consider the function n(t) = n(t), therefore we have

dr]i(r) diii{t)

~ ■ Vi+l(r) - Vi{T) = 7Ti+l(i) -

dr dt

and n»(0) = n(T). Let t = 0. We present optimal control (8) as follows

a (0) = j, Vi > 0. Hence, system (7) can be rewritten as follows: dno(T)

dr drjijr) dr

drjiij) dr

AMt) - ni(T)),

Mni(T) - no(T)) - AMt) - ni(T)),

^(ni(r) - ni-i(r)) - A(ni+i(r) - ni(T)), (14)

where n»(0) = 0. In order to solve (14) we follow the method used in [25]. Consider the generator function to be:

<x

n(z,t) = J] ni(T )zi.

i=0

multiply both sides of (14) by z for all i > 0 and take the sum of the equations:

= ^ ^ + ^r]i-i(r) — (A + n)rii(r) + \r]i+i\z\

i.e.

д

— [ф,т) -гю(т)] = J2îzl +

i=o

fiz + ^ - (A + fi)

n(z,T) -

no(T) - Azni (t)-

With respect to the generator function, °=0 iz*

(1-^)2

and the first equation of system

(14), we have (fo?°(-r-> = A(r/i(r) — xt]q{t)), therefore we have following linear equation:

d

z—rj(z, r) = [A - z(A + /i) + /iz2]rq(z, r) + kiz - AWt) + dr

By using the Laplace transform of (15) for t, we have

(1 - z)2'

(15)

z[sn*(z, s) - n(z, 0+)] = [A - z(A + j) + jz2]n*(z, s) + [jz - A]n0(s) + With regards to the initial condition, we have

n(z, 0+) = J] n^(0+)zi = 0

where (16) means that

s(1 - z)2'

i=o

n*(z,s)

z2-(A-p,zMl-z)4*(e) (1 - z)2(zs - A(1 - z) + fiz( 1 - z)) '

(16)

(17)

If we set the denominator of fraction (17) to be zero then we have the following roots:

A) ft

02

1, which ia a double root,

^-(s + A + ц + v/(A-p. + s)2 + 4p,s),

^-(s + A + - v/(A-^ + S)2 + 4/xs).

2M

According to the Rouche's theorem [25], only 02(s) is inside the unit circle on the complex plane. Under the condition that poles of function (17) correspond to a root of top of the fraction i.e., numerator turns to 0, the function (17) has bounds. Since for x = 02(s) the top of fraction (17) turns to 0, we have the following equality:

e2(s)

s A(1 - в2(s))2 - 002(s)(1 - 02(s)):

no(s)

After some algebra and considering q = ^ — A, we obtain the final representation for the Laplace transform:

n*(z,s)

1

1

tf* (s)

(1 - z)2s2 (1 - z)s2 (ei(s) - z)s2 (1 - z)s3

(18)

о

z

2

z

2

z

where

tf*(s)

+ ç - y/(s - ç)2 + 4lit

It is easy to inverse equation (18) with regarding to table of generator functions [25], to get the following representation for the Laplace transforms (t):

i+1 1 ç 1

Vt(s) = ^ - 3 " ^ " (19)

s2 s2 s3 s2 It can be easily seen that

P

tf*(s) =

P2(s) - P' P(s)P2(s) = p follows that

,,(s)|/Ji(8)|-(.+I) = el^fll _ Mf! (2o)

P2 (s) - p Pi (s) - 1

Next, we show that

— i

ig(M)= ... PI)

^(Pi(s) - 1)

and Pi(t, i)s are solutions to system of equations (10), which these are widely known [25] in queueing theory. We consider k = 2Ty/Xjj, and use equation (9) for Pq(t, i), which is

Po(T,i) = e

— (A+M)r

p—i/21i(K)+ p—(A+^)/21i+i(K) + (1 - P) p—fc/2/fc(k)

fc=i+2

where (x) is a modified Bessel function for which we have the following simple relation:

2k

h+i{x) = h-i{x)--h{x).

x

After some simple algebra, we have

Po(t, i) = p-fc/2[/fc (k) - Ifc+2(k)] ,

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

fc=i

and, therefore,

1 ^

P0(r,z) = e-(A+it)r- J] fcr"V"fc/24(/i)

a

^ fc=i+1

The following Laplace transform is known:

[kT—1 p—k/2e—/fc (k)]* = [Pi (s)]"

s

where

Pi(s) = (^s + A + fi + V(A - + s)2 + .

Then

^ fc=i+1

Comparing (19), (20), and (21), we have

= 4 + (22)

s2 s2 s3

After taking inverse Laplace transform, we obtain equation (12) for the value function V (t) = ^¿(t ).

In order to complete the proof, we need to show that the optimal control does not switch for all t > 0. To show this, with regards to (8) it is enough to show ni+1(t) —(t) > 0 for t < T .To this end, we prove that Ti(r) = ^i+1(r) — ^¿(t ) is an increasing function. Now by applying both (21) and (22), we find a representation for the Laplace transform of the first differentiation of Yi(T):

s2

By inverting the expression [^1(s)]-(i+1), we find

Pi(T) = (i + 1)t-1 p(i+1)/2 e-(A+M)r /¿+1(k), while the expression for (t) for i = 0,

Po(r) = -^ze-^W,

is known [25]. For i > 0 one can easy find that (t) is a i-fold convolution of the density function for the busy period

Pi(t) = pO(t) * ... *pO(t),

N-v-'

i

i.e., from [25], it is the distribution density function for the generalized busy period, defined as the busy period starting from a moment when the queuing system has i claims, one of which is starting to be serviced. Therefore, we have shown that

d

-Tl(r) = l-Fl(r), (23)

where (t) = fjpi(x)dx is the distribution function for the generalized busy period.

(t) < 1, hence Yi(t) are strictly increasing functions. Since Yi(0) = 0, we get that Yi(T) > 0 for t > 0, and this completes the proof. n

Fig. 2. Router diagram

Fig. 3. Transition intensity diagram on the i-th queue

4. Optimal Load Balancing (OLB)

Let us consider the EP which sends arrival jobs with the probability as G [0,1], X)f=i as = 1 to the PSs. Denote the S-tuple of queue lengths by X1 = (Xj, ...,XS) with the set of states x = (x1,xS), x G [W]S. Under these assumptions, each of the processing servers individually is a queuing system with load restriction. In [26], these processes are studied and it proved that the processes XS are controllable Markov processes. The scheme of these processes is given in Fig. 3.

Our problem is to find optimal strategy a(t, x) = a(t, Xt)|Xt = x to minimize

pT S

J (a) = Ea ^2x1 dt.

Jo 8 = 1

In order to find the optimal Markovian control, the value function is

x)\ = mmaeA I £f=i(z* + Aa8[V(t, x + e8) - V(t, x)] + +^8 [V(t, x - es) - V(t, x)]1(xs>0) ^, where the initial condition is V(T, x) = 0. The other conditions are as follows:

A = jas G [0,1], ^ as = 1

(24)

(25)

(26)

Since value function (25) and also conditions (26) are linear with respect to as, we can present the optimal strategy by a cost function. We define the set of indices s as

Jo{t,x) = {j : j = argmin1<s<sx)} (27)

in which the indices represent minimum of derivatives at the given point (t, x) by

Y+ (t,x) = V(t,x + es) - V(t,x). (28) Now we can rewrite the optimal control as follows:

a = 1(m = jo (t,x)). (29)

The idea given in [24] to find the explicit expression of the cost function:

V (t,x)

r-T ç s S r-T S S r-T К

Jt Jt s=l J* s=1 s=1 J* J*

Regarding to (26), one can easy find that the vectors of optimal control are vectors with unity as the j0(x)-th element and zeros elsewhere. As a result, the cost function with some replacement such as t = T — t, z = 9 — t, y = Z — t can be formed as follows:

S S

J=s J=s

F(i,x) = Ay + -

(30)

where p+(n,x) = P{xsn > 0|xo = x,

â(n,xn)} are probabilities of non-zero queue

lengths at the given time n, whereas at the initial time, x is the vector of queue lengths. These probabilities depend on the optimal control.

Now if we can write the equations for (t, x) then we can find the closed representation for the cost function:

p+(r,x) = 1 - e-(A+^)T

pXs/21xs Ы + P-(Xs+1)/2/xs+i(b) + (1 - Ps)

те

E

s=Xi+2

P- S/2/s(Çs)

As a result, the optimal control strategy is to route arrival jobs from users at the given time t to the processing server such that the difference derivative Y+(t, x) is minimal. The Markovian optimal strategy is not uniform. However, the properties of Y+(t, x) for large t imply the optimal controls depend on time only near the optimal control horizon, i.e., for values of t that are close to T.

5. Simulation

The system that we consider to simulate is a sequence of M/M/1 and M/M/K queuing sytem. The first server, which is called EP, routes incoming jobs to the servers (load balancer). Each server has its own queue to accept the coming new jobs when the server is currently busy. To illustrate the effectiveness of this control, we consider the following assumptions for the system.

• The arrival process is a Poisson process with the arrival intensity A.

• Servers are considered to have the same service rate

• The router sends jobs immediately as they arrive.

To evaluate the presented methodology, we compare it with Join Shortest Queue (JSQ); which is a common strategy of load balancing. We use Matlab Smulink to simulate and to compare these strategies. We compare them with regard to the waiting time for routing policies for a range of different job-size distributions such as Deterministic and Exponential distributions. The results are shown in Figs 4-7.

We consider several strategies to compare OLB method versus JSQ method. We consider two different queuing system, M/M/N and M/D/N, and then for each of them

JSQ ! MM4 DJS S MM4 DJS -

— OLE

0 2000 4000 6000 8000 10000

Fig. 4. The waiting time of sampling 10000 entities to the system where the load balancers are JSQ and OLB and the service rate depends on the service time needed for entities for all servers and the job size follows deterministic distribution

1

JSQ MD4 EJS

OLB MD4 EJS

)

0 2000 4000 6000 8000 10000

Fig. 6. The waiting time of sampling 10000 entities to the system where the load

balancers are JSQ and OLB and the service rate is constant and equals 1 for all servers and the job size follows exponential distribution

1

JSC ! MD4 DJS I MD4 DJS _

-OLE

0 2000 4000 6000 8000 10000

Fig. 5. The waiting time of sampling 10000 entities to the system where the load

balancers are JSQ and OLB and the service rate is constant and equals 1 for all servers and the job size follows deterministic distribution

1

JSQ MM4 EJS -OLB MM4 EJS

0 2000 4000 6000 8000 10000

Fig. 7. The waiting time of sampling 10000 entities to the system where the load balancers are JSQ and OLB and the service rate depends on the service time needed for entities for all servers and the job size follows exponential distribution with mean 8

we consider two different job size distributions, i.e. deterministic job size distribution and exponential job size distribution. The result of comparison for M/D/N queuing system when job size follows deterministic distribution is shown in Fig. 5 and when job size follows exponential distribution is shown in Fig. 6. Also, the result of comparison for M/M/N queuing system when job size follows deterministic distribution is shown in Fig. 4 and when job size follows exponential distribution is shown in Fig. 7.

Figs. 4 and 5 show that the performance of the OLB strategy is similar to the JSQ strategy when the job size distribution is deterministic and the queue length is the number of jobs in the queue. But the OLB strategy shows better performance rather than the JSQ strategy when job size distribution is exponential as shown in Figs. 7 and 6.

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

Regarding to the results provided by simulation, the OLB strategy has less waiting time and as a consequence less queue length than the JSQ strategy. The reason is because of that the OLB picks the queue to send the new job based on the value function of queue lengths if either queue length is the number of jobs in the queue or the queue length equals to the sum of the job sizes. Despite of OLB, JSQ picks the queue that has the minimum number of jobs between queues. As a result, the OLB strategy is insensitive to the job size distribution, while the JSQ routing policy is sensitive.

Conclusion

We study the resource allocation in a private cloud computing via queuing theory and optimal control. In this paper, in order to analyse the resources allocation in cloud computing, we propose a queueing model for cloud computing and develope a synthetical optimal control method to optimize the performance of services. The considered system is a stochastic system for load balancing and scheduling in cloud computing clusters. We consider the state vector of system to be available to observe at any time. We pose and solve the problem of constructing joint control strategies for the queuing system state. For shortness, we consider a setting with the same type jobs and find the optimal control strategy. In the future we will study our findings in a region containing private and public cloud.

References

1. Foster I., Zhao Y., Raicu I., Lu S. Cloud Computing and Grid Computing 360-Degree Compared. Grid Computing Environments Workshop, 2008, vol. 32, pp. 1-10.

2. Armbrust M., Fox A., Griffith R., Joseph A., Katz R., Konwinski A., Lee G., Patterson D., Rabkin A., Stoica I. Above the Clouds: a Berkeley View of Cloud Computing. Technical Reports, 2009, p. UCB/EECS-2009-28.

3. Menasce D.A., Ngo P. Understanding Cloud Computing: Experimentation and Capacity Planning. Computer Measurement Group Conference, 2009, Dallas, 7-11 December, 11 p.

4. Vinothina V., Sridaran R., Padmavathi Ganapathi, Resource Allocation Strategies in Cloud Computing, International Journal of Advanced Computer Science and Applications, 2012, vol. 3, no. 6, pp. 18-22.

5. Vilaplana J., Solsona F., Teixidy I., Mateo J., Abella F., Rius J. A Queuing Theory Model for Cloud Computing. The Journal of Supercomputing, 2014, vol. 69, pp. 492-507.

6. Salehpour M., Shahbahrami A., Alleviating Dynamic Resource Allocation for Bag of Tasks Applications in Cloud Computing. International Journal of Grid and Distributed Computing, 2012, vol. 5, no. 3, pp. 95-110.

7. Stolyar A. Maxweight Scheduling in a Generalized Switch: State Space Collapse and Workload Minimization in Heavy Traffic. Applied Probability Journals, 2004, vol. 14, no. 1, pp. 1-53.

8. Eryilmaz A., Srikant R. Asymptotically Tight Steady-State Queue Length Bounds Implied by Drift Conditions. Queueing Systems, 2012, vol. 1, pp. 1-49.

9. Mitzenmacher M. The Power of Two Choices in Randomized Load Balancing. PhD thesis, University of California at Berkeley, Harvard University, 1996.

10. Bramson M., Lu Y., Prabhakar B. Randomized Load Balancingwith General Service Time Distributions. ACM SIGMETRICS Performance Evaluation Review, 2010, vol. 8, no. 1, pp. 275-286.

11. Hong Chen, Heng-Qing Ye. Asymptotic Optimality of Balanced Routing. Operation Research, 2010, vol. 60, no. 1, pp. 163-179. DOI: 10.1287/opre.1110.1011

12. Yu-Tong He, Down D.G. Limited Choice and Locality Considerations for Load Balancing. Performance Evaluation, 2008, vol. 65, pp. 670-687. DOI: 10.1016/j.peva.2008.03.001

13. Vvedenskaya N.D., Karpelevich F.I., Queueing System with Selection of the Shortest of Two Queues: an Asymptotic Approach. Problems of Information Transmission, 1996, vol. 32, pp. 15-27.

14. Guo L., Yan T., Zhao S., Jiang C. Dynamic Performance Optimization for Cloud Computing Using M/M/m Queueing System. Journal of Applied Mathematics, 2014, vol. 2014, article ID: 756592, 8 p. DOI: 10.1155/2014/756592

15. Eisa M., Esedimy E.I., Rashad M.Z. Enhancing Cloud Computing Scheduling Based on Queuing Models. International Journal of Computer Applications, 2014, vol. 85, no. 2, pp. 17-23.

16. Siva Theja Maguluri, Srikant R., Lei Ying. Heavy Traffic Optimal Resource Allocation Algorithms for Cloud Computing Clusters. Perfomance Evolution, 2014, vol. 81, pp. 20-39. DOI: 10.1016/j.peva.2014.08.002

17. Zuling Kang, Hongbing Wang. A Novel Approach to Allocate Cloud Resource with Different Performance Traits. Services Computing, 2013, article ID: 13878874, 6 p. DOI: 10.1109/ SCC.2013.109

18. Winston W. Optimality of the Shortest Line Discipline. Journal of Applied Probability, 1977, vol. 14, pp. 181-189.

19. Wan C., Davis M. Existence of Optimal Control for Stochastic Jump Processes. SIAM Journal on Control and Optimization, 1979, vol. 17, pp. 511-524.

20. Elliott R. A Partially Observed Control Problem for Markov Chains. Applied Mathematics and Optimization, 1992, vol. 25, pp. 151-169.

21. Elliott R., Aggoun L., Moore J. Hidden Markov Models. Estimation and Control. N.Y., Springer, 1995.

22. Boel R., Varaia P. Optimal Control of Jump Processes. SIAM Journal on Control and Optimization, 1977, vol. 15, pp. 92-119.

23. Miller B.M. Optimization of Queuing System via Stochastic Control. Automatica, 2009, vol. 45, pp. 1423-1430.

24. Solodyannikov Yu.V. Control and Observation for Dynamical Queueing Networks. Automation and Remote Control, 2014, vol. 75, pp. 422-446.

25. Kleinrock L. Queueing Systems. N.Y., Springer, 1976.

26. Miller A.B. Using Methods of Stochastic Control to Prevent Overloads in Data Transmission Networks. Automation and Remote Control, 2010, vol. 71, pp. 1804-1815.

Received May 5, 2019

УДК 519.872 БС! 10.14529/mmp190405

РАСПРЕДЕЛЕНИЕ РЕСУРСОВ В ОБЛАЧНЫХ ВЫЧИСЛЕНИЯХ С ПОМОЩЬЮ ОПТИМАЛЬНОГО УПРАВЛЕНИЯ СИСТЕМАМИ ПЕРЕДАЧИ

А. Маданкан1, А. Делавархалафи1, С.М. Карбасси1, Ф. Адибния1

йездский университет, г. Йезд, Иран

Рассматривается задача выделения ресурсов в облачных вычислениях. Мы используем модель очередей для моделирования процесса входа в облако, а также для планирования и обслуживания входящих заданий. Основной задачей, с которой мы сталкиваемся в данной статье, является задача распределения ресурсов в системах массового обслуживания как общая задача оптимизации для управляемого марковского процесса с конечным пространством состояний. Для этой цели мы изучаем модель облачных вычислений, в которой задания по прибытию следуют случайному процессу. Мы сводим эту задачу к задаче маршрутизации. В случае минимизации стоимость выражается через среднюю длину очереди и количество потерянных заданий. Мы используем подход динамического программирования и получаем явную форму оптимального управления по уравнению Беллмана.

Ключевые слова: облачные вычисления; система множественных очередей; несколько классов работы; стохастическая политика управления.

Али Маданкан, аспирант, кафедра «Математические науки>, Йездский университет (г. Йезд, Иран), [email protected].

Али Делавархалафи, PhD, доцент, кафедра «Математические науки>, Йездский университет (г. Йезд, Иран), [email protected].

Сейед Медиа Карбасси, PhD, профессор, кафедра «Математические науки>, Йездский университет (г. Йезд, Иран), [email protected].

Фазлолла Адибния, PhD, доцент, кафедра «Математические науки>, Йездский университет (г. Йезд, Иран), [email protected].

Поступила в редакцию 5 мая 2019 г.

i Надоели баннеры? Вы всегда можете отключить рекламу.