Lecture Notes: (R15A0529) B.Tech Iv Year - I Sem (R15) (2019 - 20)
Lecture Notes: (R15A0529) B.Tech Iv Year - I Sem (R15) (2019 - 20)
[R15A0529]
DEPARTMENT OF
COMPUTER SCIENCE AND ENGINEERING
UNIT – I
Unit- II
UNIT—III
Infra s t r u c t u r e as a servi c e
Virtual mac hi n e s provisionin g and Migr a tio n s service s, On the Man a g e m e n t of
Virtual machin e s for Cloud Infra s t r u c t u r e s , Enha n ci n g cloud comp u ti n g
environ m e n t s using a clust e r as a service, secu r e distrib u t e d dat a stor a g e in
cloud comp u ti n g
Platf or m as a servi c e
Aneka, Comet cloud, T-syste m s Work flow engin e for cloud, Unde r s t a n d i n g
scientific Application s for cloud Environ m e n t s
Unit IV
UNIT- V
Gover n a n c e and cas e stu d i e s : Orga niz a tio n al Readin e s s and chan g e
Mana g e m e n t in the cloud age, Data secu rity in the cloud, Legal Issue s in cloud
comp u ti n g , Achieving produ c tio n rea di n e s s for cloud service s.
Text Boo k s
1. Cloud Compu ti n g: Principle s and Par a di g m s by Rajku m a r Buyya, Wiley, 2011.
2. Distribu t e d and cloud comp u ti n g , Kai Hwa n g, Geoffrey C. Fox,Jack
J.Donn a g a r r a , El s e vi e r , 2 0 1 2
Reference Books:
1. Cloud Comp u ti n g : A Prac tic al Appro ac h, Anthony T.Velte, Toby J.Velte,
Rober t Elsen p e t e r , Tata McGra w Hill, rp20 1 1 .
2. Ente r p ri s e Cloud Comp u ti n g , Gaut a m Shroff, Camb ri d g e Universi ty Pres s,
2010.
OUTCOMES:
Ability to unde r s t a n d the virtualiza tio n and cloud comp u ti n g conce p t s .
INDEX
UNIT
TOPIC PAGE NO
NO
virtu a l i z a t i o n 24 – 33
Intro d u c t i o n to clo u d co m p u t i n g 34 - 51
Mi gr a t i n g into a Clou d 52 - 55
Enric h i n g th e ‘Inte g r a t i o n as a
II Servi c e ’ Parad i g m for th e clo u d 56 – 60
era
Infra s t r u c t u r e as a servi c e 64 - 98
III
Platf or m as a servi c e 99 - 12 2
IV Mo n i t o r i n g and Man a g e m e n t 12 3 - 16 4
UNIT 1
INTROD U CTIO N
Parallel comp u ti n g: In par allel comp u ti n g, all proce s s o r s are eithe r tightly
couple d with cent r alize d sha r e d me mo ry or loosely couple d with distrib u t e d
mem o ry. Some autho r s refe r to this discipline as parallel proce s si n g . Inte r
proc e s s o r comm u ni c a tio n is accom plis h e d thro u g h shar e d me mo ry or via
mess a g e pas sin g. A comp u t e r syste m capa bl e of parallel comp u ti n g is
com m o nly know n as a parallel comp u t e r . Progr a m s run ni n g in a par allel
comp u t e r are called parallel progr a m s . The proc e s s of writing parallel
prog r a m s is often refer r e d to as parallel progr a m m i n g .
De g r e e s of Parall e l i s m :
Bit- leve l parall e l i s m (BLP) conve r t s bit- serial proce s si n g to word- level
proc e s si n g grad u a lly. Over the year s, use r s gra d u a t e d from 4- bit
micro p r o c e s s o r s to 8- ,16- , 32- , and 64- bit CPUs.
Data- lev el parall e l i s m (DLP) was mad e popula r throu g h SIMD (single
instr u c tio n, multiple data) and vector machin e s using vector or arr ay type s
of instr u c tio n s . DLP req ui r e s even more har d w a r e suppo r t and compile r
assist a n c e to work prop e rly.
arr ays have exce e d e d 3 TB in capa city. The rapid grow t h of flash me mo ry
and solid- stat e drives (SSDs) also impac t s the futur e of HPC and HTC
syste m s .
The host machin e is equipp e d with the physical hard w a r e . The VM is built
with virtu al resou r c e s man a g e d by a gues t OS to run a specific applic ation.
Betw e e n the VMs and the host platfor m , one nee d s to deploy a middle w a r e
layer called a virtual machin e monitor (VMM).
Distrib u t e d and cloud comp u ti n g syste m s are built over a large num b e r of
auton o m o u s comp u t e r node s. Thes e node mac hin e s are interc o n n e c t e d by
SANs, LANs, or WANs in a hier a r c hi c al man n e r . Massive syste m s are
consid e r e d highly scala bl e, and can reac h web- scale conn e c tivity, eithe r
physically or logically. Massive syste m s are classified into four grou p s:
clust e r s , P2Pn e t w o r k s , comp u ti n g grids, and Inte r n e t clouds over huge dat a
cent e r s . In ter m s of node num b e r , thes e four syste m class e s may involve
hund r e d s , thous a n d s , or even millions of comp u t e r s as particip a ti n g node s.
Thes e mac hi n e s work collectively, coope r a tiv ely, or collabo r a t iv ely at
variou s levels.
Clu s t e r Archi t e c t u r e
In a P2P syste m , every node acts as both a client and a serve r, providin g
part of the syste m reso u r c e s . Peer machin e s are simply client comp u t e r s
conn e c t e d to the Inte r n e t . All client mac hin e s act auto no m o u s ly to join or
leave the syste m freely. This implies that no mas t e r- slave relations hi p exists
amon g the peer s . No cent r al coordi n a ti o n or cent r al data b a s e is nee d e d . In
othe r words, no peer machi n e has a global view of the entire P2P syste m .
The syste m is self- orga nizin g with distrib u t e d contr ol. Unlike the clust e r or
grid, a P2P net wo r k does not use a dedic a t e d inte rc o n n e c t i o n net wo r k. The
physical netw o r k is simply an ad hoc netw o r k forme d at various Inte r n e t
domai n s ran do m ly using the TCP/IP and NAI protocols
|7
P2P perfor m a n c e is affecte d by routin g efficiency and self- orga niz a tio n by
particip a ti n g pee r s. Fault tolera n c e , failur e man a g e m e n t , and load
balanci n g are othe r import a n t issue s in using overlay netw o r k s. Lack of
trus t amon g pee r s poses anoth e r proble m . Peer s are stra n g e r s to one
anot h e r . Secu rity, privacy, and copyrigh t violations are major worrie s
provide r supplie s the API and softw a r e tools (e.g., Java, Pytho n, Web
2.0, .NET). The user is freed from man a gi n g the cloud infras t r u c t u r e .
Soft w a r e as a Servi c e (Sa a S ) This refer s to brow s e r- initiat e d
applica tion softwa r e over thous a n d s of paid cloud custo m e r s . The
SaaS model applies to busine s s proce s s e s , indus t ry applica tion s ,
cons u m e r relation s hi p man a g e m e n t (CRM), ente r p r i s e resou r c e s
plan ni n g (ERP),hu m a n reso u r c e s (HR), and collabo r a t ive applica tion s .
On the custo m e r side, the r e is no upfron t invest m e n t in serve r s or
softw a r e licensin g. On the provide r side, costs are rat h e r low,
comp a r e d with conve n tio n al hosting of user applic ation s.
Runs with a virtu aliza tio n layer in the Linux environ m e n t This layer provide s
a parti al single- syste m image to user applica tio n s This is mainly due to the
fact that all node mac hin e s run with an indep e n d e n t oper a ti n g syste m
Suppo r t s both sequ e n t i al and par allel applic ation s, and discove r s reso u r c e s
and migr a t e s softw a r e proce s s e s amon g Linux node s Can man a g e a Linux
clust e r or a grid of multiple clust e r s
Overview
PARALLEL AND DISTRI B U T E D PROGRA M M I N G MODELS
|9
Now the sam e prog r a m is partition e d for par allel execu tio n on a cluste r of
many node s
α T + (1 - α )T/n
whe r e , the first ter m is the sequ e n ti al execu tio n time on a single proc e s s o r
and the secon d ter m is the par allel execu tion time on n proc e s si n g node s
| 10
Overview
Amd a h l' s Law: Sp e e d u p fact o r
The spe e d u p factor of using the n- proce s s o r syste m over the use of asingle
proc e s s o r is expr e s s e d by:
= 1/[ α + (1 - α)/n]
The maxim u m spee d u p of n is achieve d only if the code is fully par alleliza bl e
with α = 0
As the clust e r beco m e s sufficien tly larg e, that is, n -> ∞, S appr o a c h e s 1/ α,
an uppe r boun d on the spe e d u p S
Scaling the proble m size to matc h the clust e r capa bility (scale d- workloa d
spee d u p )
W’ = α W + (1 - α) n W
S’ = W’/W = [ α W + (1 - α) n W]/W
= α + (1 - α)n
Thus efficiency is
E’ = S’ / n = α /n + (1 - α)
For α = 0:25 and n = 256, E = 75%
Availa b i l i t y
A syste m is highly available if it has a long mea n time to failure (MTTF) and
a short mea n time to rep ai r (MTTR)
Appli c a t i o n Layer: Until now, most user applica tio n s in scienc e, busine s s ,
engin e e r i n g , and financi al are a s tend to incre a s e a syste m’s spe e d or
quality. By introd u ci n g ener gy- awa r e applica tion s , the challen g e is to desig n
sophistic a t e d multilevel and multi- domai n ene r gy man a g e m e n t applic ation s
witho u t hurtin g perfor m a n c e .
DVFS, ene r gy savings are achieve d bas e d on the fact that the powe r
cons u m p t i o n in CMOS circuits has a direct relation s hi p with frequ e n c y and
the squa r e of the voltag e supply.
Ne t w o r k Layer: Routin g and tran sfe r ri n g packe t s and ena bling netw o r k
servic es to the reso u r c e layer are the main respo n s i bility of the netw o r k
layer in distrib u t e d comp u ti n g syste m s . The major challen g e to build
ene r gy- efficient netw o r k s is, again, dete r m i ni n g how to mea s u r e , pre dict,
and crea t e a balanc e betw e e n ener gy cons u m p t io n and perfor m a n c e .
• Being expos e d , intra clust e r comm u ni c a tio n is not secu r e , unles s the
com m u n ic a t io n
subsys t e m perfor m s addition al work to ens u r e privacy and secu rity.
• Outsid e com m u ni c a tio n s may disru p t intra clust e r com m u n ic a t io n s
in an unpr e di c t a bl e
fashion.
• Stan d a r d com m u n ic a tio n protocols tend to have high overh e a d . A
disadva n t a g e is that
ther e is curr e n tly no stan d a r d for efficien t, enclos e d intra cluste r
com m u n ic a t io n.
atta c h e d . The node s are typically geog r a p h i c ally distrib u t e d , and are
not nece s s a rily in the sam e room or even in the sam e building. The
nodes are individu ally owne d by multiple owne r s .
3. Availa b i l i t y
Su p p o r t: Cluste r s can provide cost- effective HA
capa bility with lots of red u n d a n c y in proce s s o r s , mem o ry, disks, I/O
devices, netw o r k s , and oper a ti n g syste m imag e s
A Ba s i c Clu s t e r Archi t e c t u r e
| 16
Figur e show s simple clust e r of comp u t e r s built with com m o dity compo n e n t s
and fully suppo r t e d with desir e d SSI featu r e s and HA capa bility. The
proc e s si n g node s are com m o di ty workst a ti o n s , PCs, or serve r s . The node
oper a ti n g syste m s should be desig n e d for multius e r, multit a s ki n g, and
multith r e a d e d applica tio n s . The node s are inte rc o n n e c t e d by one or more
fast com m o dity netwo r k s . Thes e netwo r k s use sta n d a r d com m u ni c a t io n
protocols and oper a t e at a spee d that should be two orde r s of mag nit u d e
faste r tha n that of the curr e n t TCP/IP spee d over Ethe r n e t .
The netwo r k interfa c e card is conn e c t e d to the node’s stan d a r d I/O bus
(e.g., PCI). When the
proc e s s o r or the oper a ti n g syste m is cha n g e d , only the driver softw a r e
nee d s to chan g e
clust e r middlew a r e combin e s toge t h e r all node platfor m s at the use r spac e.
An availa bility middle w a r e offers HA service s. An SSI layer provide s a single
entry point, a single file hiera r c h y, a single point of cont rol, and a single job
man a g e m e n t syste m . In addition to
run ni n g sequ e n ti al user prog r a m s , the cluste r suppo r t s parallel
prog r a m m i n g base d on stan d a r d
lang u a g e s and com m u ni c a tio n libra ri e s using PVM, MPI, or OpenM P . The
prog r a m m i n g environ m e n t also includ e s tools for debu g gi n g , profiling,
monito ri n g, and so forth. A user inte rfa c e subsys t e m is nee d e d to combin e
the adva n t a g e s of the web interfa c e and the Window s GUI. It should also
provide user- friendly links to various prog r a m m i n g environ m e n t s , job
man a g e m e n t tools, hyper t e x t , and sear c h suppo r t so that use r s can easily
get help in prog r a m m i n g the comp u t e r clust e r .
The sha r e d- mem o ry clust e r in Part (c) is much more difficult to realize. The
nodes could be conn e c t e d by a scala ble cohe r e n c e interfa c e (SCI) ring,
which is conn e c t e d to the memo ry bus of eac h node throu g h an NIC module.
In the othe r two arc hit e c t u r e s , the inter c o n n e c t is att ac h e d to the I/O bus.
The me mo ry bus ope r a t e s at a highe r frequ e n c y than the I/O bus.
1. Four nodes of a clust e r are used as host node s to rec eive user s’ login
req u e s t s .
2. To log into the clust e r a stan d a r d Unix com m a n d such as “telne t
clust e r .c s . h k u . h k”, using the symbolic nam e of the clust e r syste m is
issue d.
3. The symbolic nam e is tran sl a t e d by the DNS, which ret u r n s with the
IP addr e s s 159.22 6. 4 1. 1 5 0 of the least- loade d node, which hap p e n s to
be node Host 1.
4. The user then logs in using this IP addr e s s .
| 19
5. The DNS periodic ally receive s load infor m a ti o n from the host nodes to
make load- balanci n g tra n sl a tio n decision s.
B. Sin g l e File Hier ar c h y : xFS, AFS, Solaris MC Proxy
The illusion of a single, hug e file syste m imag e that
tran s p a r e n t ly integ r a t e s local
and global disks and othe r file device s (e.g., tape s ). Files can
resid e on 3 type s of
location s in a clust e r:
Local stor a g e - disk on the local node.
Re m o t e stor a g e - disks on rem ot e node s.
Sta b l e stor a g e -
Persis t e n t - dat a, once writt e n to the stabl e stora g e , will
stay ther e at least for
a period of time (e.g., a week), even after the
clust e r shuts down.
Fault toler a n t - to som e degr e e , by using red u n d a n c y and
periodic al back u p to
tape s .
Three types of stor a g e in a single file hier a r c h y. Solid lines show wha t
proce s s P can acces s
and the dash e d line shows what P may be able to acce s s
and control the entir e clust e r and each individu al node from a single point.
Many clust e r s help
with this throu g h a syste m console that is conn e c t e d to all node s of the
clust e r
Sin g l e I/O Addr e s s Spa c e : A single I/O spac e implies that any node can
acce s s the RAIDs
A clust e r with single net wo r ki n g, single I/O spac e, single mem o ry, and
single point of cont rol
Oth e r Servi c e s
Sin g l e Job Man a g e m e n t : All clust e r jobs can be submit t e d from any node
to a single
job man a g e m e n t syste m. GlUnix, Codine, LSF, etc.
Sin g l e Us e r Int erf a c e : The user s use the clust e r throu g h a single
gra p hi c al interfa c e. Such an
inte rf a c e is available for works t a t io n s and PCs like CDE in Solaris/NT
Sin g l e pro c e s s spa c e All user proc e s s e s cre a t e d on various node s form a
single proc e s s spac e
and shar e a unifor m proc e s s identifica tio n sche m e . A proc e s s on any node
can cre a t e (e.g., throu g h a UNIX fork) or com m u n ic a t e with (e.g., thro u g h
signals, pipes, etc.) proce s s e s
on remo t e node s.
• Man a g e m e n t leve l This level handle s use r applica tion s and provide s a
job man a g e m e n t syste m
such as GLUnix, MOSIX, Load Sha rin g Facility (LSF), or Codine.
• Pro gr a m m i n g leve l This level provide s single file hiera r c h y (NFS, xFS,
AFS, Proxy) and
distrib u t e d shar e d mem o ry (Trea d M a r k , Wind Tunn el ).
A syste m’s reliability is mea s u r e d by the mea n time to failur e (MTTF), which
is the
aver a g e time of nor m al oper a tio n befor e the syste m (or a compo n e n t of the
syste m) fails. The met ric for service a bility is the mea n time to repai r
| 22
(MTTR), which is the aver a g e time it takes to rep ai r the syste m and resto r e
it to working condition after it fails.
Fail ur e is any event that preve n t s the syste m from nor m al oper a tio n
• Unp l a n n e d fail ur e s The syste m bre ak s , due to an oper a ti n g syste m
cras h, a har d w a r e
failur e, a netwo r k discon n e c tio n, hum a n oper a tio n erro r s , a powe r
outa g e , and so on. All
thes e are simply called failure s. The syste m mus t be repai r e d to
corr e c t the failure.
• Pla n n e d sh u t d o w n s The syste m is not broke n, but is periodic ally
take n off nor m al
oper a tio n for upgr a d e s , reconfigu r a t i o n, and maint e n a n c e .
Red u n d a n c y Tec h n i q u e s
• Fail ov e r clu s t e r: When a compo n e n t fails, this tech niq u e allows the
rem ai ni n g syste m to
take over the service s originally provide d by the failed compo n e n t . A
failover mec h a ni s m
must provide seve r al function s, such as failur e diag no si s, failure
notification, and failur e
| 24
recove ry. Failur e diag no sis refer s to the det e c tio n of a failure and the
location of the failed
compo n e n t that caus e d the failure. A comm o nly use d tech ni q u e is
hea r t b e a t , whe r e b y the
clust e r node s send out a stre a m of hea r t b e a t mes s a g e s to one anot h e r . If
the syste m does not
rec eive the stre a m of hea rt b e a t mes s a g e s from a node, it can conclud e
that eithe r the node or
the net wo r k conn e c tio n has failed.
Rec o v e r y Sc h e m e s
Fail ur e rec o v e r y refer s to the actions nee d e d to take over the workloa d of
a failed compo n e n t . Ther e are two types of recove ry tech niq u e s . In
bac k w a r d rec o v e r y , the proce s s e s run ni n g on a clust e r periodic ally save a
consis t e n t stat e (called a check p oin t) to a stabl e stor a g e . After a failur e, the
syste m is reconfig u r e d to isolat e the failed compo n e n t , resto r e s the previous
check p oi n t, and resu m e s nor m al oper a tio n. This is called rollback.
Backw a r d recove ry is relatively easy to imple m e n t in an applica tion-
indep e n d e n t , port a bl e fashion
If exec utio n time is crucial, such as in real- time syste m s whe r e the rollback
time canno t be toler a t e d , a forw ar d rec ov e ry sche m e should be used. With
such a sche m e , the syste m is not rolled back to the previou s check p oi n t
upon a failure. Inste a d , the syste m utilizes the failure diagno sis inform a tio n
to recon s t r u c t a valid syste m stat e and contin u e s exec utio n. Forw a r d
recove ry is applic ation- depe n d e n t and may nee d extr a hard w a r e
Checkp oin ti n g is the proce s s of periodic ally saving the stat e of an execu ti n g
prog r a m to stabl e stora g e , from which the syste m can recove r afte r a
failur e. Each progr a m stat e saved is called a check p oin t . The disk file that
cont ai n s the saved stat e is called the check p oi n t file.
Checkp oin ti n g tech ni q u e s are useful not only for availability, but also for
prog r a m deb u g gi n g ,
proc e s s migr a tio n, and load balan ci n g
Checkp oin ti n g can be realize d by the oper a ti n g syste m at the ker n e l lev e l ,
whe r e the OS tran s p a r e n t ly check poi n t s and res t a r t s proc e s s e s
A less tra n s p a r e n t app ro a c h links the use r code with a check poi n ti n g
library in th e us e r spa c e . Check pointin g and rest a r ti n g are handle d by
this runti m e suppo r t . This appr o a c h is used widely bec a u s e it has the
adva n t a g e that user applic ation s do not have to be modified.
Che c k p o i n t Overh e a d s
During a prog r a m’s execu tion, its stat e s may be saved many time s. This is
denot e d by the time cons u m e d to save one check poi n t. The stor a g e
overh e a d is the extra me mo ry and disk spac e requir e d for check poi n ti n g.
Both time and stor a g e overh e a d s depe n d on the size of the check poi n t file.
The time period betw e e n two check poi n t s is called the check p oi n t interv al.
Making the interv al
larg e r can redu c e check poi n t time overh e a d .
Wong and Fra nklin derive d an expr e s sio n for optim al check poi n t interval
MTTF is the syste m’s mea n time to failur e. This MTTF accou n t s the time
cons u m e d to
save one check p oi n t , and h is the aver a g e perc e n t a g e of nor m al
comp u t a t i o n perfor m e d in a check p oi n t inte rv al befor e the syste m fails. The
para m e t e r h is always in the ran g e . After a syste m is resto r e d , it nee d s to
spen d h × (check p oi n t inte rv al) time to reco m p u t e .
Incr e m e n t a l Che c k p o i n t
Inst e a d of saving the full stat e at eac h check p oi n t, an incre m e n t a l
check p oi n t sche m e saves only the portion of the stat e that is chan g e d from
the previous check poi n t In full- stat e check p oi n ti n g , only one check p oi n t file
nee d s to be kept on disk. Subs e q u e n t check p oi n t s simply overw rit e this file.
With incre m e n t a l check p oi n ti n g , old files nee d e d to be kept, beca u s e a stat e
may span many files. Thus, the total stor a g e req ui r e m e n t is large r
Fork e d Che c k p o i n t i n g
Most check p oin t sche m e s are blockin g in that the nor m al comp u t a ti o n is
stopp e d while check p oi n tin g is in prog r e s s . With enou g h me mo ry,
check p oi n t overh e a d can be red uc e d by making a copy of the progr a m stat e
in me mo ry and invoking anoth e r asynch r o n o u s thre a d to perfor m the
check p oi n ti n g concu r r e n t ly. A simple way to overla p check p oin ti n g with
comp u t a t i o n is to use the UNIX fork( ) syste m call. The forke d child proce s s
duplic at e s the pare n t proce s s’s addr e s s spac e and check poi n t s it.
Mea n w hile, the par e n t proce s s contin u e s execu tio n. Overla p pi n g is achieve d
since check p oi n ti n g is disk- I/O inten sive
Us e r- Dire c t e d Che c k p o i n t i n g
The check p oi n t overh e a d s can som e ti m e s be subs t a n t i ally red uc e d if the
user inser t s code (e.g.,
libra ry or syste m calls) to tell the syste m when to save, what to save, and
what not to save. What
| 26
should be the exact conte n t s of a check poi n t ? It should cont ai n just enou g h
infor m a ti o n to allow a syste m to recove r. The stat e of a proce s s includ e s its
dat a stat e and cont rol stat e
JMS Admi n i s t r a t i o n
JMS should be able to dyna mic ally reconfigu r e the cluste r with
minim al impac t on the runnin g jobs.
The admi nis t r a t o r’s prolog u e and epilogu e script s should be able to
run before and after each job for secu ri ty checki n g, accou n ti n g , and
clean u p .
Users should be able to cleanly kill their own jobs.
The adminis t r a t o r or the JMS should be able to clea nly susp e n d or kill
any job.
Clean mea n s that when a job is susp e n d e d or killed, all its
proc e s s e s must be includ e d.
| 27
Sc h e d u l i n g Mod e s
Ded i c a t e d Mod e :
Only one job runs in the cluste r at a time, and at most one proce s s of
the job is assign e d to a node at a time.
The single job runs until comple tion before it relea s e s the clust e r to
run othe r jobs.
Spa c e Sh ari n g :
Multiple jobs can run on disjoint partition s (grou p s) of node s
simult a n e o u s ly.
At most one proc e s s is assig n e d to a node at a time.
Althoug h a partition of node s is dedic a t e d to a job, the inter c o n n e c t
and the I/O subsys t e m may be shar e d by all jobs.
Tim e sh ari n g :
Multiple user proce s s e s are assign e d to the sam e node.
Time- sha ri n g introd u c e s the following parallel sche d uli n g policies:
Ind e p e n d e n t
Sch e d u l i n g (Ind e p e n d e n t ) : Uses the oper a ti n g
syste m of each clust e r node to sche d ul e differ e n t proc e s s e s as in
a tradition al works t a ti o n.
Gan g Sc h e d u l i n g : Sche d ul e s all proce s s e s of a parallel job
toget h e r . When one proce s s is active, all proce s s e s are active.
| 29
A tradition al comp u t e r runs with a host oper a ti n g syste m speci ally tailor e d
for its hard w a r e arc hit e c t u r e , as show n in Figur e ( a). After virtu aliza tio n,
differe n t use r applica tio n s man a g e d by their own oper a ti n g syste m s (gues t
OS) can run on the sam e har d w a r e , indep e n d e n t of the host OS. This is
often done by adding addition al softw a r e , called a virtualiza tion layer as
show n in Figur e( b). This virtualiza tion layer is known as hype rviso r or
virtu al mac hi n e monitor (VMM). The VMs are in the uppe r boxes, whe r e
applica tion s run with their own gues t OS over the virtualize d CPU, me mo ry,
and I/O resou r c e s . The main function of the softw a r e layer for virtu aliza tio n
is to virtu alize the physical hard w a r e of a host mac hi n e into virtu al
reso u r c e s to be use d by the VMs
e.g, MIPS bina ry code can run on an x-86- bas e d host mac hi n e with
the help of ISA emula tion. Typical syste m s : Bochs, Cruso e, Que m u, BIRD,
Dynam o
Advan t a g e :
Limit a t i o n :
• Very expen s ive to imple m e n t (complexity)
Limit a t i o n :
• Poor applica tion flexibility and isolation
Limit a t i o n :
| 33
Oper a ti n g syste m virtu alization inse rt s a virtu aliza tion layer inside an
oper a ti n g syste m to
partition a machin e’s physical resou r c e s . It ena bl e s multiple isolat e d VMs
within a single oper a ti n g syste m kern el. This kind of VM is often called a
virtu al execu tion enviro n m e n t (VE), Virtual Privat e Syste m (VPS), or simply
cont ai n e r . From the user’s point of view, VEs look like real serve r s . This
mea n s a VE has its own set of proc e s s e s , file syste m , use r accou n t s , netw o r k
inte rf a c e s with IP add r e s s e s , routin g tables, firew all rules, and othe r
pers o n al settin g s . Althoug h VEs can be custo mize d for differ e n t people, they
sha r e the sam e oper a ti n g syste m kern el. Therefo r e, OS- level virtualiza tio n is
also called single- OS imag e virtu alization
| 34
The hype rviso r suppo r t s hard w a r e- level virtualiza tio n (see Figur e 3.1(b)) on
bare met al devices like CPU, memo ry, disk and netw o rk interfa c e s . The
hype rviso r softwa r e sits directly betw e e n the physical har d w a r e and its OS.
This virtualiza tion layer is refer r e d to as eithe r the VMM or the hype rviso r.
The hype rviso r provide s hype r c alls for the gues t OSes and applic ation s.
| 35
Depe n di n g on the function ality, a hype rvis or can assu m e a micro- kern el
arc hit e c t u r e like the Microsoft Hype r- V. Or it can assu m e a monolithic
hype rviso r archit e c t u r e like the VMwar e ESX for serve r virtualiza tion
The Xen arch i t e c t u r e ’ s spe c i a l do m a i n 0 for con t r o l and I/O, and sev e r a l
gu e s t do m a i n s for us er appli c a t i o n s
The core compo n e n t s of a Xen syste m are the hype rviso r, kern el, and
applica tion s The gues t OS, which has cont rol ability, is called Domain 0, and
the othe r s are called Domain U. Domain 0 is a privileg e d gues t OS of Xen. It
is first loade d whe n Xen boots withou t any file syste m drive rs being
availa ble. Domain 0 is desig n e d to acces s har d w a r e direc tly and man a g e
devices. Ther efor e , one of the res po n si bilities of Domain 0 is to alloca t e and
map har d w a r e resou r c e s for the gues t domai n s (the Domain U domai n s).
Full virtualiza tion does not nee d to modify the host OS. It relies on bina ry
tran sl a tio n to trap and to virtu alize the execu tion of cert ai n sensitive, non
virtu aliza bl e instr u c tio n s . The gues t OS es and their applica tio n s consist of
nonc ritic al and critical instr u c tio n s . In a host- bas e d syste m , both a host OS
and a gues t OS are use d. A virtu aliza tio n softwa r e layer is built betw e e n the
host OS and gues t OS
| 36
Full Virtu al i z a t i o n
With full virtu aliz ation, noncritic al instr u c tio n s run on the har d w a r e directly
while critical
instr u c tio n s are discove r e d and repla c e d with trap s into the VMM to be
emul at e d by softw a r e . Both the hyperviso r and VMM app ro a c h e s are
consid e r e d full virtualiza tion. Why are only critical instr u c tio n s tra p p e d into
the VMM? This is beca u s e bina ry tran sl a tio n can incur a larg e perfor m a n c e
overh e a d . Nonc ritic al instr u c tio n s do not cont rol har d w a r e or thre a t e n the
secu ri ty of the syste m , but critical instr u c tio n s do. Ther efo r e , run nin g
nonc ritic al instr u c tio n s on hard w a r e not only can promo t e efficiency, but
also can ens u r e syste m secu ri ty.
Ho s t - Bas e d Virtu a l i z a t i o n
This host bas e d archit e c t u r e has som e distinct adva n t a g e s . First, the use r
can install this VM archit e c t u r e witho u t modifying the host OS. The
| 37
virtu alizing softw a r e can rely on the host OS to provide device driver s and
othe r low- level service s. This will simplify the VM desig n and eas e its
deploym e n t . Secon d, the host- bas e d app ro a c h app e al s to many host
machin e configu r a ti o n s .
Comp a r e d to the hype rviso r/VM M archit e c t u r e , the perfor m a n c e of the host-
bas e d arc hit e c t u r e may also be low. When an applica tio n requ e s t s hard w a r e
acce s s , it involves four layers of mappi n g which downg r a d e s perfor m a n c e
significa n tly. When the ISA of a gues t OS is differe n t from the ISA of the
und e rlyin g hard w a r e , bina ry tra n sl a tio n must be adopt e d . Althou g h the
host- bas e d arc hit e c t u r e has flexibility, the perfor m a n c e is too low to be
useful in prac tic e.
Para- Virtu a l i z a t i o n with Com p i l e r Sup p o r t
Para- virtualiza tio n nee d s to modify the gues t oper a ti n g syste m s . A para-
virtu alize d VM provide s
speci al APIs req ui ri n g subs t a n t i al OS modificatio n s in user applica tio n s .
Para- virtualiza tio n atte m p t s to redu c e the virtualiza tio n overh e a d , and thus
improve perfor m a n c e by modifying only the gues t OS kern el.
The gues t oper a ti n g syste m s are para- virtualize d. They are assist e d by an
intellige n t compile r to repla c e the non virtualiza ble OS instr u c tio n s by
hype r c alls The lower the ring num b e r , the highe r the privileg e of instr u c tio n
being execu t e d . The OS is res po n si ble for man a gi n g the har d w a r e and the
privileg e d instru c tio n s to execu t e at Ring 0, while use r- level applica tio n s
run at Ring 3.
The bes t exam pl e of para- virtualiza tion is the KVM
To suppo r t virtu aliza tio n, proc e s s o r s such as the x86 employ a special
run ni n g mode and instr u c tio n s , known as hard w a r e- assist e d virtualiza tio n.
In this way, the VMM and gues t OS run in differe n t mode s and all sensitive
instr u c tio n s of the gues t OS and its applica tion s are trap p e d in the VMM. To
save proce s s o r stat e s , mode switc hin g is compl et e d by har d w a r e .
| 38
Mode r n oper a ti n g syste m s and proce s s o r s per mit multiple proc e s s e s to run
simult a n e o u s ly. If ther e is no prote c tio n mech a ni s m in a proc e s s o r , all
instr u c tio n s from differe n t proce s s e s will acce s s the hard w a r e direc tly and
caus e a syste m cras h. Ther efor e , all proc e s s o r s have at least two mode s,
user mode and supe rviso r mode, to ensu r e controlle d acce s s of critical
har d w a r e . Instr u c tio n s run ni n g in supe rviso r mode are called privileg e d
instr u c tio n s . Othe r instr u c tio n s are unprivile g e d instr u c tio n s . In a
virtu alize d environ m e n t , it is more difficult to make OSes and applic ation s
run corr e c tly beca u s e ther e are more layers in the mac hi n e stack. The
VMwar e
Workst a tio n is a VM softw a r e suite for x86 and x86- 64 comp u t e r s . This
softw a r e suite allows use r s to set up multiple x86 and x86- 64 virtual
comp u t e r s and to use one or more of thes e VMs simult a n e o u s ly with the
host oper a ti n g syste m . The VMwar e Workst a tio n assu m e s the host- bas e d
virtu aliza tio n. Xen is a hyperviso r for use in IA-32, x86- 64, Itaniu m , and
Powe r P C 970 hosts.
CPU Virtu al i z a t i o n
A VM is a duplica t e of an existing comp u t e r syste m in which a majority of
the VM instr u c tio n s are execu t e d on the host proc e s s o r in native mode.
Thus, unp rivileg e d instr u c tio n s of VMs run direc tly on the host mac hin e for
highe r efficiency. Othe r critic al instr u c tio n s should be han dl e d carefully for
corr e c t n e s s and stability. The critical instr u c tio n s are divided into thre e
cate g o ri e s : privileg e d instr u c tio n s , control sensitive instr u c tio n s, and
beh avior- sensitive instr u c tio n s . Privileg e d instr u c tio n s exec u t e in a
privileg e d mode and will be tra p p e d if execu t e d outsid e this mode. Cont rol-
sensitive instr u c tio n s att e m p t to chan g e the configu r a t io n of reso u r c e s
used. Behavior- sensitive instr u c tio n s have differe n t behavior s depe n di n g on
the configu r a ti o n of resou r c e s , includin g the load and store oper a tio n s over
the virtual memo ry. A CPU archit e c t u r e is virtualiza bl e if it suppo r t s the
ability to run the VM’s privileg e d and unp rivileg e d instr u c tio n s in the CPU’s
user mode while the VMM runs in supe rviso r mode.
Me m o r y Virtu a l i z a t i o n
Virtual me mo ry virtu aliza tio n is similar to the virtu al me mo ry suppo r t
provide d by mode r n oper a ti n g syste m s . In a tradition al executio n
environ m e n t , the ope r a ti n g syste m maint ai n s map pi n g s of virtual me mo ry to
machin e me mo ry using page tables, which is a one- stag e mappi n g from
virtu al mem o ry to mac hin e memo ry. All mode r n x86 CPUs includ e a mem o ry
man a g e m e n t unit (MMU) and a tran sl a tio n lookasid e buffer (TLB) to
optimize virtu al memo ry perfor m a n c e . Howeve r , in a virtu al exec ution
environ m e n t , virtual me mo ry virtu aliza tio n involves sha ri n g the physical
syste m memo ry in RAM and dyna mi c ally allocatin g it to the physical
mem o ry of the VMs. a two- stag e mappi n g proc e s s should be maint ai n e d by
the gues t OS and the VMM, res p e c tiv ely: virtual mem o ry to physical
mem o ry and physic al mem o ry to machin e me mo ry. The gues t OS contin u e s
to cont rol the map pin g of virtu al addr e s s e s to the physical me mo ry
| 39
addr e s s e s of VMs. But the gues t OS canno t directly acce s s the actu al
machin e me mo ry.
The VMM is res po n si ble for map pi n g the gues t physical mem o ry to the
actu al mac hi n e mem o ry
I/O Virtu a l i z a t i o n
I/O virtu aliza tio n involves man a gi n g the routin g of I/O requ e s t s betw e e n
virtu al devices and
the sha r e d physical har d w a r eAll the function s of a device or bus
infras t r u c t u r e , such as device enu m e r a t i o n , identifica tio n, inter r u p t s , and
DMA, are replic a t e d in softw a r e . This softw a r e is locat e d in the VMM and
acts as a virtu al device. The I/O acces s requ e s t s of the gues t OS are tra p p e d
in the VMM which inter a c t s with the I/O device sA single har d w a r e device
can be shar e d by multiple VMs that run conc u r r e n t ly
Therefor e , it is com m o n that most serve r s in dat a cent e r s are und e r u t ilize d.
A larg e amou n t of
har d w a r e , spac e, pow e r, and man a g e m e n t cost of thes e serve r s is wast e d.
Serve r consolida tio n is an appr o a c h to improve the low utility ratio of
har d w a r e reso u r c e s by redu ci n g the num b e r of physical serve r s . Among
seve r al serve r consolid a tio n tech niq u e s such as cent r a liz e d and physical
| 41
consolid a tio n, virtu alization- bas e d serve r consolida tio n is the most powe rf ul.
Data cent e r s need to optimiz e their resou r c e man a g e m e n t the use of VMs
incre a s e s reso u r c e man a g e m e n t complexity. This caus e s a challen g e in
ter m s of how to improve reso u r c e utilization as well as gua r a n t e e QoS in
dat a cent e r s .
Advan t a g e s
“Clouds are a large pool of easily usable and acce s si ble virtu alize d
reso u r c e s (such as hard w a r e , develop m e n t platfor m s and/or service s).
Thes e resou r c e s can be dyna mic ally reconfig u r e d to adjus t to a varia bl e
load (scale), allowing also for an optim u m reso u r c e utilization”
| 42
Key ch ar a c t e r i s t i c s of clo u d co m p u t i n g
(1) the illusion of infinite comp u ti n g resou r c e s ;
(2) the elimin a tio n of an up- front com mit m e n t by cloud use r s;
(3) the ability to pay for use…as nee d e d
Figur e 1.1 shows the conve r g e n c e oftec h n olo gy fields that significa n tly
adva n c e d and contrib u t e d to the adve n t of cloud comp u ti n g .
| 43
The IT world is curr e n tly expe rie n ci n g a switch from in- hous e g e n e r a t e d
comp u ti n g powe r into utility- supplied comp u ti n g reso u r c e s deliver e d ov e r
the Inte r n e t as Web service s .
The eme r g e n c e of Web service s (WS) open stan d a r d s has significa n tly
cont rib u t e d t o adva n c e s in the dom ai n of softwa r e integ r a t io n. Web
servic es c a n combin e toget h e r applica tio n s run ni n g on differe n t mess a gi n g
prod u c t platfor m s , e n a b li n g infor m a tio n from one applica tio n to be mad e
availa ble tooth e r s , and ena blin g inter n al applica tion s to be mad e available
over theInt e r n e t .
function al Web applica tion into prac tic e just by gluing piece s with few
linesof code.
Grid Com p u t i n g
A key aspe c t of the grid vision realization has bee n building stan d a r d
Webs e rvic e s- bas e d protocols that allow distrib u t e d resou r c e s to be
“discove r e d , a c c e s s e d , allocat e d, monito r e d , accou n t e d for, and billed for,
etc., and ingen e r a l man a g e d as a single virtual syste m .” The Open Grid
Servic es Archite c t u r e ( OG SA) add r e s s e s this nee d for stan d a r d i z a ti o n by
defining a set of core c a p a b ilities and beh avior s that addr e s s key conce r n s in
grid syste m s .
Anoth e r issue that has lead to frustr a t io n when using grids is the
availa bilityof reso u r c e s with divers e softw a r e configu r a tio n s , includin g
dispa r a t e oper a ti n g s y s t e m s , librari e s, compiler s , runti m e environ m e n t s , and
so forth. At the sam e ti m e , use r applica tio n s would often run only on
speci ally custo mize d environ m e n t s . C o n s e q u e n t ly, a port a bility bar ri e r has
often bee n pres e n t on most g ri d infras t r u c t u r e s , inhibitin g users of adoptin g
grids as utility comp u ti n g e n vi ro n m e n t s
| 46
Utility Com p u t i n g
Hard w a r e Virtu a l i z a t i o n
The adve n t of sever al innovative tech nolo gi e s— m ul ti- core chips, par a-
virtu aliza tio n,
| 47
Perc eive d ben efits were improve m e n t s on shari n g and utilization, bett e r
man a g e a b ility, and highe r reliability.
A num b e r of VMM platfor m s exist that are the basis of many utility orclou d
comp u ti n g enviro n m e n t s . The most nota bl e ones are VMWare, Xen,
andKVM.
Xen :The Xen hype rviso r star t e d as an open- sourc e projec t and has serve d
as a bas e to othe r virtu aliza tio n prod u c t s , both com m e r ci al and open- sourc e.
It has pione e r e d the para- virtu alization conc e p t , on which the gues t
oper a ti n g syste m , by mea n s of a speci alize d kern el, can inte r a c t with the
hype rviso r, thus significa n tly improving perfor m a n c e .
KVM : The kern el- bas e d virtual mac hi n e (KVM) is a Linux virtu aliza tio n
subsys t e m . It has been part of the mainline Linux kern el since version
2.6.20, thus being natively suppo r t e d by sever al distrib u tio n s . In addition,
activities such as memo ry man a g e m e n t and sche d ulin g are carrie d out by
| 48
existing kern el feat u r e s , thus makin g KVM simple r and smalle r than
hype rviso r s that take cont rol of the entir e mac hin e
Auto n o m i c Com p u t i n g
inspire s oft w a r e tech n olo gie s for data cent e r auto m a t io n, which may
perfor m tasks s u c h as: man a g e m e n t of service levels of run ni n g applica tio n s ;
man a g e m e n t ofdat a cent e r capa city; proac tive disas t e r recove ry; and
auto m a t io n of VMprovisionin g
Figur e 1.3 depict s the layer e d orga niz a tio n of the cloud stackfro m physical
infras t r u c t u r e to applica tio n s .
Infra s t r u c t u r e as a Servi c e
Amazon Web Servic e s mainly offers IaaS, which in the case of its
EC2s e rvic e mea n s offering VMs with a softw a r e stack that can be
custo miz e d s i mil a r to how an ordina r y physical serve r would be custo miz e d.
Users are give n privileg e s to perfor m num e r o u s activities to the serve r, such
as: star ti n g a n d stoppi n g it, custo mizin g it by inst alling softw a r e packa g e s ,
atta c hi n g vi r t u a l disks to it, and configu ri n g acce s s per mi s sio n s and firew alls
rules.
Platf or m as a Servi c e
Dep l o y m e n t Mod e l s
Pu bli c clo u d : “cloud mad e availabl e in a pay- as- you- go man n e r to the
gen e r al public”
Privat e clo u d : “inter n a l dat a cent e r of a busin e s s or othe r orga niza tio n,
not mad e availabl e to the gen e r al public.”
Com m u n i t y clo u d: “sha r e d by sever al orga niz a tio n s and suppo r t s a specific
com m u n i ty that has sha r e d conce r n s (e.g., mission, secu ri ty requir e m e n t s ,
policy, and complia n c e conside r a t i o n s )
Hybrid clo u d takes sha p e when a privat e cloud is supple m e n t e d with
comp u ti n g capa city from public clouds.
The appr o a c h of temp o r a r ily renti n g capa ci ty to handl e spike s in load is
known as “clo u d - bur s t i n g ”
Des ir e d Feat u r e s of a Clou d
Self- Servi c e : clouds must allow self- service acce s s so that custo m e r s can
req u e s t , custo miz e, pay, and use servic e s withou t interv e n t io n of hum a n
oper a t o r s
Per- Usa g e Met e r i n g and Billi n g : Cloud comp u ti n g elimina t e s up- front
com mit m e n t by user s, allowing the m to requ e s t and use only the nece s s a r y
amou n t . Servic es must be price d on a short ter m basis (e.g., by the hour),
allowing user s to relea s e (and not pay for) reso u r c e s as soon as they are not
nee d e d
Feat u r e s
Virtu a l i z a t i o n Sup p o r t : The multi- ten a n cy aspec t of clouds req ui r e s
multiple custo m e r s with dispa r a t e requir e m e n t s to be serve d by a single
har d w a r e infras t r u c t u r e . Virtualize d resou r c e s (CPUs, me mo ry, etc.) can be
sized and resize d with cert ai n flexibility. Thes e featu r e s make har d w a r e
virtu aliza tio n, the ideal tech n ology to cre a t e a virtual infras t r u c t u r e that
partition s a dat a cent e r amon g multiple ten a n t s .
Self- Servi c e , On- De m a n d Res o u r c e Provi s i o n i n g : Self- service acce s s to
reso u r c e s has bee n perc eive d as one the most attr a c tive featu r e s of clouds.
This feat u r e ena bl e s user s to direc tly obtain service s from clouds, such as
spaw ni n g the cre a tio n of a serve r and tailoring its softw a r e , configu r a ti o n s ,
and secu ri ty policies, withou t inte r a c t i n g with a hum a n syste m
adminis t r a t o r . This capa bility “elimina t e s the nee d for more time-
cons u mi n g, labor- inte n sive, hum a n driven
proc u r e m e n t proc e s s e s familiar to many in IT”.
Multi p l e Bac k e n d Hyp e rvi s o r s : Differe n t virtu aliza tion models and tools
offer differ e n t ben efits, draw b a c k s , and limitatio n s. Thus, some VI man a g e r s
provide a unifor m man a g e m e n t layer reg a r dl e s s of the virtu alization
tech nology use d. This char a c t e r i s ti c is more visible in open- sourc e VI
man a g e r s , which usually provide plugg a bl e driver s to inter a c t with multiple
hype rviso r s . In this direc tion, the aim of libvirt is to provide a unifor m API
that VI man a g e r s can use to man a g e dom ain s (a VM or contai n e r runnin g an
insta n c e of an oper a ti n g syste m ) in virtualize d nodes using stan d a r d
oper a tio n s that abst r a c t hype rviso r specific calls.
| 53
Int erf a c e to Pub li c Clou d s : Exten di n g the capa city of a local in- hous e
comp u ti n g infras t r u c t u r e by borro wi n g reso u r c e s from public clouds is
adva n t a g e o u s . In this fashion, institu tio n s can make good use of their
availa ble reso u r c e s and, in case of spikes in dem a n d , extr a load can be
offload e d to rent e d reso u r c e s . A VI man a g e r can be used in a hybrid cloud
setu p if it offers a drive r to man a g e the life cycle of virtualize d reso u r c e s
obtain e d from exter n a l cloud provide r s . To the applic ation s, the use of
leas e d reso u r c e s must ideally be tran s p a r e n t .
This feat u r e is illust r a t e d by the case in which an AR requ e s t for a given slot
canno t be satisfie d, but the provide r can offer a distinc t slot that is still
satisfa ct o ry to the user. This proble m has been addr e s s e d in OpenP EX,
which incorp o r a t e s a bilat e r al negotia tio n protoc ol that allows user s and
provide r s to come to an alter n a t iv e agr e e m e n t by exch a n gi n g offers and
count e r offers.
Cas e Stu di e s
Apac h e VCL : The Virtual Compu ti n g Lab projec t has bee n incep t e d in 2004
by res e a r c h e r s at the North Carolina Stat e University as a way to provide
custo miz e d enviro n m e n t s to comp u t e r lab user s. Apach e VCL provide s the
following feat u r e s : (i) multi- platfor m controlle r, base d on Apach e/P H P ; (ii)
Web port al and XML-RPC interf ac e s ; (iii) suppo r t for VMwar e hype rviso r s
(ESX, ESXi, and Serve r); (iv) virtual net wo r k s ; (v) virtu al clust e r s ; and (vi)
adva n c e res e rv a tio n of capa city.
Citrix Ess e n t i a l s : The Citrix Esse n ti als suite is one the most feat u r e
comple t e VI man a g e m e n t softw a r e availabl e, focusing on man a g e m e n t and
| 55
Eno m a l y ECP: The Enom aly Elastic Comp u ti n g Platfor m , in its most
comple t e edition, offers most featu r e s a service provide r nee d s to build an
IaaS cloud. Enom aly ECP provide s the following feat u r e s : Linux- bas e d
cont rolle r; Web port al and Web service s (REST) interf ac e s ; Xen back- end;
inte rf a c e to the Amazon EC2 public cloud; virtu al net wo r k s; virtu al clust e r s
(Elas ticValet) Eucalypt u s .
The Euc a ly p t u s : fram e w o r k was one of the first open- sourc e project s to
focus on building IaaS clouds. It has been develop e d with the inten t of
providing an open- sourc e imple m e n t a t i o n nearly identic al in function ality to
Amazon Web Servic e s APIs. Eucalypt u s provide s the following featu r e s :
Linux- bas e d cont rolle r with adminis t r a t i o n Web port al; EC2- comp a ti bl e
(SOAP, Query) and S3- comp a ti bl e (SOAP, REST) CLI and Web port al
inte rf a c e s ; Xen, KVM, and VMWare backe n d s ; Amazon EBS- comp a ti bl e
virtu al stora g e devices; interf ac e to the Amazon
EC2 public cloud; virtual netwo r k s .
featu r e s : multi- platfor m (Java) cont rolle r; Web port al and Web service s
(REST) interfa c e s ; Citrix XenSe rv e r backe n d; adva n c e res e rv a tio n of
capacity with negoti a tio n.
VMWar e vSp h e r e and vClou d: vSph e r e is VMwar e’s suite of tools aime d
at tra n sfo r m i n g IT infras t r u c t u r e s into privat e clouds. In the vSphe r e
arc hit e c t u r e , serve r s run on the ESXi platfor m . A sepa r a t e serve r runs
vCent e r Serve r, which cent r alize s cont rol over the entir e virtual
infras t r u c t u r e . Throu g h the vSphe r e Client softw a r e , adminis t r a t o r s conn e c t
to vCent e r Serve r to perfor m various tasks. The Distribu t e d Resou r c e
Sche d ul e r (DRS) make s allocation decision s base d on pre d efin e d rules and
policies. It continu o u sly monitor s the amou n t of reso u r c e s availa ble to VMs
and, if nece s s a r y, mak e s alloca tion chan g e s to meet VM req ui r e m e n t s . In
the stor a g e virtualiza tio n real m, vStor a g e VMFS is a clust e r file syste m to
provide agg r e g a t e sever al disks in a single volum e. VMFS is especi ally
optimize d to store VM image s and virtu al disks. It suppo r t s stora g e
equip m e n t that use Fibre Chan n el or iSCSI SAN. vSph e r e provide s the
following feat u r e s : Window s- bas e d cont rolle r (vCent e r Serve r); CLI, GUI,
Web port al, and Web service s interfa c e s ; VMwar e ESX, ESXi backe n d;
VMwar e vStor a g e VMFS stora g e virtu aliza tio n; inte rfac e to exte r n a l clouds
(VMwa r e vCloud part n e r s ); virtu al net wo r k s (VMWar e Distribu t e d Switch);
dyna mic reso u r c e allocation (VMwa r e DRM); high availability; data
prot e c tio n (VMWar e Consolid a t e d Backu p).
Feat u r e s
cost_be n efit ratio to be expe ri e n c e d by user applic ation s whe n moved to the
cloud. The most releva n t feat u r e s are: (i) geog r a p hi c distrib u tio n of data
cent e r s ; (ii) variety of use r interfa c e s and APIs to acce s s the syste m ; (iii)
speci alize d compo n e n t s and service s that aid particul a r applica tio n s (e.g.,
loadb al a n c e r s , firew alls); (iv) choice of virtualiza tion platfor m and oper a ti n g
syste m s ; and (v) differe n t billing met ho d s and period (e.g., pre p ai d vs. post-
paid, hourly
vs. mont hly).
Hyp e rvi s o r and Oper a t i n g Sys t e m Choi c e : Tradition ally, IaaS offerin g s
have bee n bas e d on heavily custo mize d open- sourc e Xen deploym e n t s . IaaS
provide r s need e d expe r tis e in Linux, netw o r ki n g, virtu aliza tio n, mete ri n g,
reso u r c e man a g e m e n t , and many othe r low- level aspe c t s to succ e s sfully
deploy and maint ai n their cloud offerin g s.
Cas e Stu di e s
Amaz o n Web Servi c e s : Amazon WS4 (AWS) is one of the major playe r s in
the cloud comp u ti n g mark e t. It pione e r e d the intro d u c tio n of IaaS clouds in
2006. It offers a variety cloud service s, most nota bly: S3 (stor a g e ), EC2
(virtu al serve r s ), Cloudfro n t (cont e n t delivery), Cloudfro n t Stre a m i n g (video
stre a m i n g), SimpleDB (stru c t u r e d data s t o r e ), RDS (Relation al Data b a s e ),
SQS (reliabl e mes s a gi n g), and Elastic MapRe d u c e (data proc e s s in g). The
ElasticCo m p u t e Cloud (EC2) offers Xen- bas e d virtu al serve r s (insta n c e s )
that can be insta n ti a t e d from Amazon Machin e Imag e s (AMIs). Inst a n c e s
are available in a variety of sizes, oper a ti n g syste m s , arc hit e c t u r e s , and
price. CPU capa city of insta n c e s is mea s u r e d in Amazon Comp u t e Units and,
altho u g h fixed for each insta n c e , vary amon g insta n c e types from 1 (small
insta n c e ) to 20 (high CPU insta n c e ). Each insta n c e provide s a cert ai n
amou n t of nonp e r si s t e n t disk spac e; a persis t e n c e disk service (Elastic Block
Stor a g e ) allows atta c hi n g virtu al disks to insta n c e s with spac e up to 1TB.
Elasticity can be achieve d by combinin g the CloudW a t c h , Auto Scaling, and
Elastic Load Balancin g feat u r e s , which allow the num b e r of insta n c e s to
scale up and down auto m a ti c ally bas e d on a set of custo miz a bl e rules, and
traffic to be distrib u t e d acros s available insta n c e s . Fixed IP add r e s s (Elastic
IPs) are not availabl e by default, but can be obtain e d at an addition al cost.
mea n s a virtu al serve r can make use of addition al CPUs auto m a ti c ally up to
the maxim u m num b e r of cores availabl e in the physical host.
The Joyent public cloud offers the following feat u r e s : multiple geogr a p h i c
location s in the United Stat e s ; Web- bas e d use r interfa c e ; acces s to virtu al
serve r via SSH and Web- bas e d administ r a ti o n tool; 100% availability SLA;
per mont h pricing; OS- level virtu aliza tio n Solaris cont ain e r s ; Open- Solaris
oper a ti n g syste m s ; auto m a t i c scaling (vertic al).
GoGrid : GoGrid, like many othe r IaaS provide r s , allows its custo m e r s to
utilize a ran g e of pre- mad e Window s and Linux imag e s , in a rang e of fixed
insta n c e sizes. GoGrid also offers “value- add e d” stacks on top for
applica tion s such as high- volum e Web serving, e- Comm e r c e , and dat a b a s e
store s . It offers som e nota ble feat u r e s , such as a “hybrid hosting” facility,
which combin e s tradition al dedic a t e d hosts with auto- scaling cloud serve r
infras t r u c t u r e . As part of its core IaaS offering s, GoGrid also provide s free
har d w a r e load balanci n g, auto- scaling capa bilities, and persis t e n t stor a g e ,
featu r e s that typically add an addition al cost for most othe r IaaS provide r s .
Feat u r e s
Cas e Stu di e s
App Engi n e : Google App Engine lets you run your Python and Java Web
applica tion s on elastic infras t r u c t u r e supplie d by Google. App Engine allows
your applica tio n s to scale dyna mic ally as your traffic and data stor a g e
req ui r e m e n t s incre a s e or decr e a s e . It gives develop e r s a choice betw e e n a
Python stack and Java. The App Engine serving archit e c t u r e is nota bl e in
that it allows real- time auto- scaling withou t virtu aliza tio n for many com m o n
types of Web applic ation s. Howeve r , such auto- scaling is dep e n d e n t on the
applica tion develop e r using a limited subs e t of the native APIs on each
platfor m , and in some insta n c e s you nee d to use specific Google APIs such
as URLFet c h, Datas t o r e , and mem c a c h e in plac e of cert ain native API calls.
| 61
For exa m pl e, a deploye d App Engine applica tio n canno t write to the file
syste m direc tly (you must use the Google Datas t o r e ) or open a socke t or
acce s s anot h e r host directly (you mus t use Google URL fetch servic e). A
Java applica tio n canno t crea t e a new Thre a d eithe r.
Forc e . c o m : In conjun c tio n with the Salesforc e. c o m service, the Forc e.co m
PaaS allows develop e r s to cre a t e add- on function ality that integ r a t e s into
main Salesfor c e CRM SaaS applica tion. Force.co m offers develop e r s two
appr o a c h e s to cre a t e applica tion s that can be deploye d on its SaaS plafor m:
a host e d Apex or Visualforc e applica tio n. Apex is a prop ri e t a r y Java- like
lang u a g e that can be used to crea t e Salesfor c e applica tio n s . Visualforc e is
an XML-like syntax for building UIs in HTML, AJAX, or Flex to overlay over
the Salesforc e hoste d CRM syste m . An applica tio n store called
AppExch a n g e is also provide d, which offers a paid &
free applic a tion direc t o ry.
Despit e the initial succ e s s and popul a rity of the cloud comp u ti n g par a di g m
and the exten sive availa bility of provide r s and tools, a significa n t num b e r of
challen g e s and risks are inhe r e n t to this new model of comp u ti n g .
Provide r s , develop e r s , and end user s mus t consid e r thes e challe n g e s and
risks to take good adva n t a g e of cloud comp u ti n g . Issue s to be face d includ e
user privacy, data secu ri ty, dat a lock- in, availability of service, disas t e r
recove ry, perfor m a n c e , scala bility, ene r gy- efficiency, and prog r a m m a b i lity.
Se c u r i t y, Priva cy, and Trus t: Secu rity and privacy affect the entire cloud
comp u ti n g stack, since ther e is a mas sive use of third- party service s and
infras t r u c t u r e s that are use d to host import a n t dat a or to perfor m critical
oper a tio n s . In this scen a rio, the trus t towa r d provide r s is funda m e n t a l to
ens u r e the desir e d level of privacy for applica tio n s hoste d in the cloud.
| 62
Legal and regul a t o r y issue s also nee d atte n tio n. When data are moved into
the Cloud, provide r s may choos e to locat e the m anyw h e r e on the plane t.
The physical location of dat a cent e r s dete r m i n e s the set of laws that can be
applie d to the man a g e m e n t of dat a. For exam pl e, specific crypto g r a p h y
tech ni q u e s could not be used bec a u s e they are not allowe d in som e
count ri e s . Similarly, count ry laws can impos e that sensitive dat a, such as
patie n t healt h reco r d s , are to be store d within nation al bord e r s .
Mi gr a t i n g int o a Clou d
As show n in Figur e 2.1, the promis e of the cloud both on the busin e s s front
(the attr a c tiv e cloudo no m ic s) and the tech nolo gy front widely aided the
CxOs to spaw n out sever al non- mission critical IT need s from the ambit of
their captive tra dition al data cent e r s to the app ro p ri a t e cloud service.
Invaria bly, thes e IT nee d s had som e com m o n featu r e s : They wer e typically
Web- orient e d; they repr e s e n t e d seaso n al IT dem a n d s ; they were ame n a bl e
to par allel batc h proce s si n g; they were non- mission critical and the r efo r e
did not have high secu ri ty dem a n d s . They includ e d scientific applic a tion s
too. Seve r al small and mediu m busin e s s ente r p r i s e s , how ev e r, levera g e d the
cloud much beyond the cautiou s user. Many star t u p s open e d their IT
dep a r t m e n t s exclusively using cloud service s—ve ry succe s sf ully and with
high ROI. Having obse rv e d thes e succe s s e s , seve r al larg e ente r p ri s e s have
start e d succe s sf ully run ni n g pilots for lever a gi n g the cloud. Many large
ente r p r i s e s run SAP to man a g e their oper a tio n s .
| 65
Why Migr a t e
• Busine s s Reas on s
• Technologic al Reaso n s
What can be Mi gr a t e d
• Applicatio n
• Code
• Design
• Archite c t u r e
• Usag e
Apart from thes e costs, othe r factors that play a major role in the
cloudo no m ic s of migr a tio n are
• the licensin g issue s (for per h a p s part s of the ente r p r i s e applic ation)
• the SLA complian c e s , and the pricing of the cloud service offerin g s.
Migr a tio n initiative s into the cloud are imple m e n t e d in phas e s or in stag e s
| 66
• Opti m i z e : Thes e test res ult s could be positive or mixed. In the latte r case,
we iter a t e and optimize as appr o p ri a t e . After sever al such optimizin g
itera tio n s , the migra tio n is dee m e d succe s sf ul.
| 67
Mi gr a t i o n risk s
Migr a tio n risks for migr a ti n g into the cloud fall unde r two broa d
cate g o ri e s:
The ge n e r a l mi gr a t i o n risk s :
• perfor m a n c e monito ri n g and tunin g,
• the port a bility and intero p e r a b ility issues which could help mitigat e
pote n ti al vendo r lock- ins
Why Inte g r a t i o n ?
• Incre a s i n gly busine s s applic ation s are deploye d in clouds to rea p
the busin e s s and tech nic al benefits.
• On the othe r hand, ther e are still innu m e r a b l e applica tio n s and
dat a sourc e s locally station e d and sust ai n e d prima rily due to the
secu ri ty reas o n.
• The ques tio n her e is how to crea t e a sea ml e s s conn e c tivity
betw e e n those host e d and on- pre mis e applica tion s to empow e r
the m to work toge t h e r .
Saa S INTEGRATIO N
• Cloud- cent ric integ r a t io n solutions are being develop e d and
demo n s t r a t e d for showc a s i n g their capa bilities for integ r a ti n g
ente r p r i s e and cloud applica tion s .
• Now with the arrival and adoptio n of the tran sfo r m a t iv e and
disru p tiv e par a di g m of cloud comp u ti n g , every ICT produ c t s are
being conve r t e d into a collection of servic es to be deliver e d via
the open Inte r n e t
• In that line, the stan d a r d s- complia n t integ r a ti o n suites are being
tran sition e d into service s so that any integ r a ti o n nee d of any
one from any part of the world , can be easily, chea ply and
rapidly met.
| 69
For service inte g r a t io n, it is ente r p r i s e service bus (ESB) and for data
integ r a t io n, it is ente r p r i s e dat a bus (EDB).
Ther e are Mess a g e orien t e d middlew a r e (MOM) and mes s a g e broke r s for
integ r a t i n g decou pl e d applica tion s throu g h mes s a g e passin g and pick up.
Events are comin g up fast and ther e are compl ex event proce s si n g (CEP)
engin e s that rec eive a stre a m of divers e event s from divers e sourc e s ,
proce s s the m at real- time to extr a c t and figur e out the enc a p s u l a t e d
knowled g e , and accor di n gly selec t and activat e one or more targ e t
applica tio n s .
• Cloud infras t r u c t u r e is not very useful witho u t SaaS applic a tion s that
run on top of the m, and SaaS applica tion s are not very valua bl e
witho u t acces s to the critical corpo r a t e dat a that is typically locked
away in various corpo r a t e syste m s .
• So, for cloud applic ation s to offer maxim u m value to their user s, they
nee d to provide a simple mec h a n i s m to import or load exte r n a l dat a,
export or replica t e their data for repor tin g or analysis purpo s e s , and
finally keep their data synch r o niz e d with on- pre mis e applic ation s.
Rea s o n s :
Limit e d Acc e s s : Access to cloud resou r c e s (SaaS, PaaS, and the
infras t r u c t u r e s ) is more limite d than local applica tio n s . Once
applica tion s move to the cloud, custo m applic ation s must be
design e d to suppo r t integ r a t io n beca u s e ther e is no longe r that low
level of acces s. Ente r p ri s e s puttin g their applica tion s in the cloud or
thos e subsc ri b e r s of cloud- base d busin e s s service s are dep e n d e n t on
the vendor to provide the integ r a ti o n hooks and APIs.
The Perva s i v e DataCl o u d : Perva sive Data Cloud is the first multi-
tena n t platfor m for deliverin g the following
Oth er Prod u c t s
• Bluewolf
• Online MQ
• CloudM Q
• Linxter
Ent e r p ri s e clo u d co m p u t i n g
| 74
• reso u r c e pooling,
• rapid elasticity
• meas u r e d service
These ques tio n s are addr e s s e d from two stra t e gi c pers p e c tive s :
(1) Adoption (2) Consu m p tio n
• Mark e t - Drive n Stra t e g y : This stra t e g y is more attr a c tive and viable
for small, agile orga niz a tio n s that do not have (or wish to have)
mas sive invest m e n t s in their IT infra s t r u c t u r e . The objective her e is
to identify and acquir e the “bes t deals” for IT capa bilities as dem a n d
and supply chan g e , ena bling ongoing redu c tio n s in OPEX and CAPEX.
• Conv e n i e n c e - Drive n Strat e g y : The objective is to redu c e the load and
nee d for dedic a t e d syste m admi nist r a t o r s and to make acce s s to IT
capa bilities by user s easie r, reg a r dl e s s of their location and conn e c tivity
(e.g. over the Inte r n e t ). The expec t a t io n is that the cost of obtainin g IT
capa bilities from a CDC and makin g the m acce s si ble to user s is
significa n tly lower than the cost of having a dedic a t e d admi nis t r a t o r
Con s u m p t i o n Stra t e g y:
The cons u m p t i o n strat e gi e s make a distinc tio n betw e e n data and
applica tio n logic beca u s e ther e are que s tio n s of progr a m m i n g models
used, dat a sensitivity, softw a r e licensin g and expe ct e d res po n s e times that
nee d to be consid e r e d .
Ther e are four cons u m p t i o n s stra t e gi e s identified, whe r e the differe n c e s in
objectives, conditions and actions reflect the decision of an orga niz a tio n to
tra d e- off hosting costs, cont rolla bility and resou r c e elas ticity of IT
resou r c e s for softw a r e and dat a
• Soft w a r e Provi s i o n . This stra t e g y is releva n t whe n the elasticity
req ui r e m e n t is high for softw a r e and low for dat a, the cont rolla bility
conc e r n s are low for softw a r e and high for dat a, and the cost redu c tio n
conc e r n s for softw a r e are high, while cost redu c tio n is not a priority for
dat a, given the high cont rolla bility conc e r n s for dat a, that is, dat a are
highly sensitive
• Stor a g e Provi s i o n . This stra t e g y is releva n t whe n the elasticity
req ui r e m e n t s is high for data and low for softw a r e , while the
cont rolla bility of softw a r e is more critical than for data. This can be the
case for dat a inten sive applica tio n s , whe r e the res ult s from proce s si n g in
the applica tion are more critical and sensitive than the data itself.
Furt h e r m o r e , the cost redu c tio n for dat a reso u r c e s is a high conce r n,
here a s cost for softw a r e , given its criticality, is not an issue for the
orga niza tio n within rea so n a bl e mea n s .
• Sol u t i o n Provi s i o n . This stra t e g y is releva n t whe n the elas ticity and cost
red uc tio n req ui r e m e n t s are high for softw a r e and dat a, but the
cont rolla bility requir e m e n t s can be entr u s t e d to the CDC. It is not the case
that cont rolla bility is an insignifica n t req uir e m e n t ; it is rath e r the cas e that
the orga niz a tio n trus t s the CDC sufficien tly to man a g e acce s s and usag e
cont rol of its softwa r e and data
• Red u n d a n c y Servi c e s . This stra t e g y can be consid e r e d as a hybrid
ente r p r i s e cloud stra t e g y, wher e the orga niz a tio n switc h e s betw e e n
| 77
UNIT – 3
Virtu a l Mac h i n e s Provi s i o n i n g and Migr a t i o n Servi c e s
Intro d u c t i o n an d Ins p ir a t i o n
Cloud comp u ti n g builds on
service- orient e d arc hit e c t u r e (SOA)
grid comp u ti n g and
virtualiza tion tech n olo gy
Offers Infras t r u c t u r e as a service(Ia a S ) to the end user s as a public utility
servic e base d on pay- as- you- use and on- dem a n d comp u ti n g models
The provisionin g of the cloud infras t r u c t u r e in data cent e r s is a
pre r e q u i s it e
The provisionin g for syste m s and applica tio n s on a large num b e r of
physical machin e s is time- cons u mi n g proce s s with low ass u r a n c e on
deploym e n t’s time and cost
| 78
Two core servic es ena bl e the user s to get the best out of the IaaS model in
public and privat e cloud setu p s
Virtual mac hi n e provisionin g
Migra tion servic es
When installing a new serve r for a cert ai n workloa d to provide a service for
a client, the following step s are requir e d
Check the invent o ry for a new mac hin e
Get one, form a t, install OS req ui r e d , install servic es
A serve r is need e d along with lots of secu ri ty batc h e s and applian c e s
With the eme r g e n c e of virtualiza tion tech nolo gy and the cloud comp u ti n g
IaaS model, the sam e task can be achieve d in few minut e s .
To provision a virtu al serve r throu g h a self- service inte rfac e with small
step s to get what we desire with the requir e d specifica tion s
Provisionin g this mac hi n e in a public cloud like Amazon Elastic
Comp u t e Cloud (EC2)
With the adva n c e of the revolutionize d virtu alization tech n olo gy and
migr a tio n service s associa t e d with hype rvis o r s’ capa bilities thes e tasks
(maint e n a n c e , upgr a d e s , patc h e s , etc.) nee d no time to accom plis h
Provisionin g a new virtual machin e is a matt e r of minut e s
Migra tion s of a virtual machin e is a matt e r of milliseco n d s
• The virtualiza tio n layer will partition the physical resou r c e of the
und e rlyin g physical serve r into multiple virtual mac hin e s with differe n t
workloa d s
Exam pl e s for vendo r s and fram e w o r k s that provide Iaas in privat e setu p s
▫ Eucalypt u s (elastic utility comp u ti n g archit e c t u r e linking your
prog r a m s to useful syste m s )
▫ Open Neb ul a
A third type of cloud setu p nam e d Hybrid clo u d
▫ A combin a tio n of privat e/int e r n a l and exte r n a l cloud reso u r c e s
existing toge t h e r by ena blin g outsou r ci n g of nonc ritic al service s and
functions in public cloud and keepin g the critical ones inte r n al
Main function of Hybrid cloud is to rele a s e reso u r c e s from a public cloud
and han dle sudd e n dem a n d usag e called cloud burs tin g
OCCI an d OGF
Anoth e r stan d a r d i z a tio n effort has been initiat e d by Ope n Grid Foru m
(OGF) to deliver a stan d a r d API for cloud IaaS
Ope n Clou d Com p u t i n g Inte rf a c e Worki n g Grou p (OCCI- WG)
Dedica t e d for deliverin g an API specifica tion for the remot e
man a g e m e n t of cloud comp u ti n g’s infras t r u c t u r e for allowing the
develop m e n t of intero p e r a b l e tools for com m o n tasks includin g
deploym e n t , auto no mi c scaling, and monitorin g.
Coverin g a high- level function ality requi r e d for man a gi n g the lifecycle
of virtu al mac hi n e s / w o r kl o a d s , runnin g on virtu alization
tech nologi e s/ co n t a i n e r s and suppo r ti n g service elasticity
The new API for interf aci n g IaaS cloud comp u ti n g facilities will allow
| 83
VM Provisioning Process
The com m o n and norm al step s of provisionin g a virtual serve r
Select a serve r from a pool of availabl e serve r s (Physical serve r s with
enou g h capa city) along with the appro p ri a t e OS tem pla t e
Load the appro p ri a t e softwa r e oper a ti n g syste m , device driver s,
middle w a r e , and the nee d e d applic ation s for the service req ui r e d
Custo miz e and configu r e the mac hi n e to configu r e an associa t e d
netw o r k and stor a g e reso u r c e s e.g., IP addr e s s , Gate w a y
The virtual serve r is rea dy to star t with its newly loade d softw a r e
Serve r provisionin g is defining serve r’s configu r a ti o n base d on the
orga niza tio n req uir e m e n t s , a hard w a r e , and softw a r e compo n e n t ,
| 84
VM Provi s i o n i n g Proc e s s
Migra tion service is the proc e s s of moving a virtu al mac hi n e from one
host serve r or stor a g e location to anoth e r
Mi gr a t i o n s Tec h n i q u e s
• Live migr a tio n also called hot or real- time migr a tio n: The move m e n t
of a virtual machin e from one physic al host to anot h e r while being powe r e d
on witho u t any notice a bl e effect from the end user’s point of view (a mat t e r
of milliseco n d s )
Live Migr a t i o n
Live mi g r a t i o n ’ s me c h a n i s m
Sta g e 1: Res e r v a t i o n
During the first itera tion, all page s are tran sfe r r e d from A to B
Subs e q u e n t itera tio n s copy only thos e page s dirtie d during the
previous tran sf e r pha s e
Sta g e 4: Com m i t m e n t
Sta g e 5: Activat i o n
Post- migr a tio n code runs to reat t a c h the device’s driver s to the new
machin e and adver tis e move d IP add r e s s e s
• VMware Vmotion
– Allows users to automatically optimize and allocate an entire pool of resources for maximum
hardware utilization, flexibility, and availability
– To perform hardware’s maintenance without scheduled downtime along with migrating virtual
machines away from failing or underperforming servers
| 87
Differe n c e s betw e e n Hot (Live) migra tio n and Cold migr a tio n
Live migr a tio n nee d s a sha r e d stora g e for virtual mac hin e s in the
serve r’s pool, but cold migr a tio n does not
In live migr a tio n for a virtual mac hin e betw e e n two host s, ther e would be
cert ai n CPU comp a ti bility checks to be applied, in cold migr a tio n this
checks do not apply
• Res e r v e d ins t a n c e s : which allow you to pay a low, one- time fee and in turn
receive a significa n t discou n t on the hourly usag e char g e for that insta n c e .
It ens u r e s that any rese rv e d insta n c e you launc h is guar a n t e e d to succ e e d
(provide d that you have booke d the m in adva n c e). This mea n s that use r s of
thes e insta n c e s should not be affect e d by any tran si e n t limitatio n s in EC2
capacity.
• Spo t ins t a n c e s : which ena bl e you to bid what e v e r price you want for
insta n c e capa city, providing for even gre a t e r saving s, if your applica tio n s
have flexible start and end time s.
• Suppo r t for most Linux distrib u tio n s (sourc e and bina ry pack a g e s ).
• Suppo r t for runnin g VMs that run atop the Xen hype rviso r or KVM.
• Suppo r t for othe r kinds of VMs, such as VMwar e, is targ e t e d for futur e
relea s e s .
• Secu r e inte r n a l com m u ni c a ti o n using SOAP with WS security.
• Cloud administ r a t o r’s tool for syste m’s man a g e m e n t and use r’s accou n ti n g .
• The ability to configu r e multiple clust e r s eac h with privat e inter n a l netw o rk
addr e s s e s into a single cloud.
| 89
Euc al y p t u s Archi t e c t u r e
Privat e and hybrid clouds shar e thes e sam e char a c t e ri s ti c s but, inste a d
of selling capa city over publicly acce s si bl e interfa c e s , focus on providing
capa city to an orga niza tio n’s inter n a l user s.
Virtualiza tio n tech nolo gi e s have bee n the key enabl e r of many of thes e
salient char a c t e r i s ti c s of IaaS clouds by giving provide r s a more flexible
and gene ri c way of man a gi n g their resou r c e s .
Service owne r s will ent e r into SLAs with their end user s, coverin g
gua r a n t e e s such as the timelin e s s .
Cloud provide r s are not directly expos e d to the service sem a n t i c s or the
SLAs that service owne r s may cont r a c t with their end users .
The cloud provide r’s task is to make sure that resou r c e alloca tion
requ e s t s are satisfie d with specific prob a bility and timelin e s s .
These req ui r e m e n t s are formalize d in infras t r u c t u r e SLAs betw e e n the
service owne r and cloud provide r, sep a r a t e from the high- level SLAs
betw e e n the service owne r and its end use r s.
VM Man a g e m e n t
Open Neb ul a uses the following conc e p t s for its imag e man a g e m e n t
model
| 93
Virtualiza tio n tech n ologi e s are a key enabl e r of many featu r e s found in
IaaS clouds. Virtual mac hi n e s are also an app e alin g vehicle for
imple m e n t i n g efficient rese rv a tio n of resou r c e s due to their ability to be
susp e n d e d , pote n ti ally migra t e d , and res u m e d withou t modifying any of
the applica tion s run nin g inside the VM.
over the SLA evalua tio n period. The lower the availability perc e n t ile,
the che a p e r the cost of reso u r c e cons u m p t i o n
• Det e r m i n i s t i c SLAs . Thes e are, in fact, prob a bilistic SLAs whe r e
reso u r c e availability perc e n tile is 100%. These SLAs are most string e n t
and difficult to gua r a n t e e . From the provide r’s point of view, they do
not admit capa ci ty multiplexin g. Ther efo r e this is the most costly option
for service provide r s , which may be applied for critical service s
Elas t i c i t y rul e s are scaling and de- scaling policies that guide
tran sition of the servic e from one configu r a t io n to anoth e r to matc h
cha n g e s in the environ m e n t . The main motivatio n for defining thes e
policies ste m s from the pay- as yougo billing model of IaaS clouds.
The servic e owne r is inter e s t e d in paying only for wha t is really
req ui r e d to satisfy workloa d dem a n d s minimizin g the over-
provisionin g overh e a d
Clu s t e r as a Servi c e ( C a a S )
RVWS Des i g n
• While Web service s have simplified resou r c e acce s s and man a g e m e n t , it is
not possible to know if the resou r c e ( s) behin d the Web service is (are) rea dy
for requ e s t s . Clients need to excha n g e num e r o u s mess a g e s with req ui r e d
Web service s to learn the curr e n t activity of reso u r c e s and thus face
significa n t overh e a d loss if most of the Web service s prove ineffective
• client s still have to locat e the service s the m s elv e s.
• Finally, the Web service s have to be stat eful so that they are able to best
reflec t the curr e n t stat e of their resou r c e s .
| 97
This was the motivatio n for cre a ti n g the RVWS fram e w o r k (Resou r c e s Via
Web Service s)
RVWS combin e s dyna mi c attrib u t e s , stat eful Web service s (awa r e of their
past activity), stat ef ul and dyna mi c WSDL docu m e n t s , and broke ri n g into a
single, effective, servic e- bas e d fram e w o r k
Ther e are two cate go ri e s of dyna mic attrib u t e s addr e s s e d in the RVWS
fram e w o r k
• Stat e
• Char a c t e r i s tic
Stat e att rib u t e s cover the curr e n t activity of the service and its reso u r c e s ,
thus indicati n g rea di n e s s .
Exa m p l e : A Web service that expos e s a clust e r (itself a complex
reso u r c e ) would most likely have a dyna mic stat e attrib u t e
that indicat e s how many node s in the clust e r are busy and
how many are idle.
Char a c t e r i s ti c attrib u t e s cover the oper a tio n al feat u r e s of the service, the
resou r c e s behind it, the quality of servic e (QoS), price and provide r
inform a tio n.
Exa m p l e : A possible char a c t e r i s tic is an array of suppo r t softw a r e within
the clust e r.
| 98
• Decision
• Notification.
The Detec tio n module routin ely que ri e s the reso u r c e for attrib u t e
inform a tio n (1- 2). Any cha n g e s in the attrib u t e s are pass e d to the Decision
modul e (3) that decide s if the attrib u t e chan g e is large enoug h to warr a n t a
notifica tio n. This preve n t s exces sive com m u ni c a ti o n with the Web service.
Upda t e d attrib u t e s are pass e d on to the Notification modul e (4), which
inform s the stat eful Web service (5) that upd a t e s its inter n al stat e. When
client s requ e s t s the stat ef ul WSDL docu m e n t (6), the Web service retu r n s
the WSDL docu m e n t with the values of all attrib u t e s (7) at the requ e s t time.
• The Web Service Descri ption Lang u a g e (WSDL) gover n s a sche m a that
desc ri b e s a Web service and a docu m e n t writt e n in the sche m a
• All inform a t io n of servic e resou r c e s is kept in a new WSDL section called
Resou r c e s .
• For each resou r c e behind the Web service, a Resou r c eI nfo section exists.
• Each Resou r c e I nfo section has a resou r c e- id attrib u t e and two child
sections: stat e and char a c t e r i s ti c.
• All reso u r c e s behind the Web service have uniqu e identifier s. When the
Conne c t o r lear n s of the reso u r c e for the first time, it publish e s the resou r c e
to the Web service.
Pu bli c a t i o n in RVWS
To help eas e the publica tio n and discove ry of req ui r e d service s with stat ef ul
WSDL docu m e n t s , a Dyna mic Broke r was propos e d .
The goal of the Dyna mic Broke r is to provide an effective public ation and
discove ry servic e base d on service, resou r c e , and provide r dyna mic
att rib u t e s .
| 100
When publishi n g to the Broke r (1), the provide r sends attrib u t e s of the Web
service to the Dyna mic Broke r.
The dyna mi c attrib u t e s indicat e the function ality, cost, QoS, and any othe r
att rib u t e s the provide r wishe s to have publish e d abou t the service.
Furt h e r m o r e , the provide r is able to publish inform a t io n about itself, such
as the provide r’s conta c t details and rep u t a ti o n.
After publica tio n (1), the Broke r gets the stat ef ul WSDL docu m e n t from the
Web service (2).
After gettin g the stat ef ul WSDL docu m e n t , the Dyna mic Broke r extra c t s all
resou r c e dyna mic attri b u t e s from the stat ef ul WSDL docu m e n t s and stor e s
the resou r c e attrib u t e s in the reso u r c e s stor e.
The Dyna mic Broker then store s the (stat el e s s) WSDL docu m e n t and service
att rib u t e s from (1) in the service stor e. Finally, all attrib u t e s abou t the
provide r are plac e d in the provide r s stor e.
As the Web servic e chan g e s , it is able to send a notifica tio n to the Broke r (3)
which then upda t e s the releva n t attrib u t e in the releva n t store.
When discove ri n g servic es , the client sub mi t s to the Dyna mic Broker
thre e grou p s of requir e m e n t s
• servic e
• reso u r c e
• provide r.
The Dyna mic Broke r comp a r e s each requir e m e n t group on the relat e d
dat a store (2). Then, after gettin g matc h e s , the Broke r applies filterin g
(3). As the client using the Broker could vary from hum a n oper a t o r s to
othe r softwa r e units, the resultin g matc h e s have to be filtere d to suit
the client. Finally, the filter e d result s are ret u r n e d to the client (4).
Clu s t e r as a Servi c e
The purpo s e of the CaaS Technolo gy is to eas e the publica tio n, discove ry,
selection, and use of existin g comp u t a ti o n a l clust e r s .
• Middle w a r e .
| 102
The middlew a r e virtualize s the clust e r into a single syste m imag e; thus
reso u r c e s such as the CPU can be used withou t knowing the
orga niza tio n of the clust e r.
• Sc h e d u l e r : man a g e s the allocatio n of jobs to node s
• Pu bli s h e r Web servi c e was to expos e the dyna mic attrib u t e s of the clust e r
The CaaS Servic e com m u ni c a t e s with the clust e r’s sche d ul e r , thus freeing
the client from nee din g to know how the sche d ul e r is invoked whe n
submit ti n g and monitori n g jobs.
• file man a g e m e n t .
Each modul e in the CaaS Service enca p s u l a t e s one of the tasks and is
able to com m u ni c a t e with othe r modul e s to exte n d its function ality .
• The modul e s inside the CaaS Web service are only acces s e d throu g h an
inte rf a c e
• Invoking an oper a tio n on the CaaS Service Inte rf ac e (discove ry, etc.)
invoke s oper a tio n s on various modul e s.
• It resolves the tran sf e r of all files to the cluste r by invoking the File
Mana g e r (3) that make s a conn e c tio n to the clust e r stora g e and com m e n c e s
the tra n sf e r of all files (4).
• Upon compl etio n of the tran sfe r , the outco m e is repo r t e d back to the Job
Mana g e r (5).
• On failur e, a repor t is sent and the client can decid e on the appr o p ri a t e
action to take. If the file tran sf e r was succe s sful, the Job Man a g e r invokes
the sche d ul e r on the clust e r (6).
• If the outco m e of the sche d ul e r (6) is succe s sful, the client is then inform e d
(7- 8). The outco m e includ e s the respo n s e from the sche d ul e r, the job
| 105
identifier the sche d ul e r gave to the job, and any othe r inform a tio n the
sche d ul e r provides .
• During executio n, client s should be able to view the exec utio n prog r e s s of
their jobs.
• The client conta c t s the CaaS servic e interfa c e (1) that invokes the Job
Mana g e r modul e (2).
• No matt e r what the oper a tio n is (check, paus e, or ter mi n a t e ), the Job
Mana g e r only has to com m u ni c a t e with the sche d ul e r (3) and
• repo r t s back a succe s sful outco m e to the client (4- 5).
Res u l t Coll e c t i o n :
• Clients start the error or res ult file tra n sf e r by cont a c ti n g the CaaS Servic e
Inte rf ac e (1) that then invokes the File Mana g e r (2) to retri ev e the files from
the clust e r’s data stora g e (3).
• If ther e is a tran sfe r error, the File Mana g e r atte m p t s to resolve the issue
first befor e informin g the client.
| 106
• If the tra n sf e r of files (3) is succe s sf ul, the files are ret u r n e d to the CaaS
Servic e Inte rf a c e (4) and the n the client (5).
• When ret u r ni n g the files, URL link or a FTP add r e s s is provide d so the client
can ret rieve the files.
Goo g l e App En gi n e
| 108
• The SDC const r u c t s an encryp t e d conn e c tio n betw e e n the dat a sourc e
and Google Apps. As long as the dat a sourc e is in the Google Apps
domai n to the Google tunn el protocol serve r s , whe n the use r want s to
get the data, he/sh e will first send an autho riz e d data requ e s t s to
Google Apps, which forwa r d s the req u e s t to the tunn el serve r.
• The tunn el serve r s validat e the req u e s t identity. If the identity is valid,
the tunn el protoc ol allows the SDC to set up a conn e c tio n,
aut h e n t i c a t e , and encry p t the data that flows acros s the Inter n e t . At
the sam e time, the SDC uses resou r c e rules to validat e whet h e r a user
is autho riz e d to acces s a specified resou r c e .
• When the req u e s t is valid, the SDC perfor m s a netw o r k req u e s t . The
serve r validat e s the signe d req u e s t , checks the cred e n ti al s, and
retu r n s the dat a if the use r is autho rize d
• From the pers p e c tiv e of cloud stora g e service s, data integ ri ty dep e n d s
on the secu rity of oper a tio n s while in stor a g e in addition to the
secu ri ty of the uploa di n g and downlo a di n g session s.
• The uploa di n g session can only ens u r e that the dat a receive d by the
cloud stora g e is the dat a that the user uploa d e d;
• The downloa di n g session can gua r a n t e e the dat a that the user
retriev e d is the dat a cloud stor a g e recor d e d .
• Unfort u n a t e l y, this proce d u r e applie d on cloud stora g e service s
canno t gua r a n t e e dat a integ ri ty
• First, assu m e that Alice, a comp a ny CFO, store s the comp a n y financi al
dat a at a cloud stora g e servic e provide d by Eve.
• And then Bob, the comp a ny adminis t r a t i o n chair m a n , downlo a d s the
dat a from the cloud.
Four Sol u t i o n s
• Neith e r TAC nor SKS
• With SKS but withou t TAC
• With TAC but withou t SKS
• With Both TAC and SKS
| 110
Uniqu e issue s of the cloud dat a stor a g e platfor m from a few differe n t
pers p e c t iv e s
– Data b a s e Outs o u r c i n g and Query Int e g r i t y Ass ur a n c e
• Storin g dat a into and fetchin g dat a from devices and
machin e s behin d a cloud are esse n ti ally a novel form of
dat a b a s e outso u r ci n g
– Data Int e g r i t y in Untr u s t w o r t h y Stor a g e
• The fear of losing dat a or data corr u p tio n
• Relieve the user s’ fear by providin g tech nolo gi e s that
ena bl e use r s to check the inte g rity of their dat a
– Web- Appli c a t i o n - Bas e d Se c u r i t y
• Once the data s e t is store d remot ely, a Web brows e r is one
of the most conve nie n t appr o a c h e s that end user s can use
to acces s their data on remo t e servic es
• Web secu rity plays a more impor t a n t role for cloud
comp u ti n g
– Multi m e d i a Data Se c u r i ty
• With the develop m e n t of high- spee d netwo r k tech n ologi e s
and large band w id t h conn e c tio n s , more and more
multim e di a data are being store d and shar e d in cyber
spac e
• The secu ri ty requir e m e n t s for video, audio, pictu r e s , or
imag e s are differ e n t from othe r applica tio n s
• Or deny the vulne r a bility in the syste m afte r the dat a have bee n
stolen by an adve r s a r y
Before cloud comp u ti n g, seve r al remo t e dat a stora g e checkin g protocols
have bee n sugg e s t e d
In practic e , a rem ot e data poss e s sio n checki n g protocol has to satisfy the
following five requir e m e n t s
Req u ir e m e n t # 1
• The verifier has to poss e s s a comple t e copy of the dat a to
be check e d
• In practic e , it does not make sens e for a verifier to keep a
duplic at e d copy of the cont e n t to be veri fie d
• Storin g a more concis e cont e n t s diges t of the dat a at the
verifier should be enou g h
Req u ir e m e n t # 2
• The protoc ol has to be very robu s t consid e ri n g the
untr u s t w o r t h y prove r
• A malicious prove r is motivat e d to hide the violation of
dat a integ ri ty
• The protoc ol should be robu s t enou g h that such a prove r
ough t to fail in convincin g the verifier
Req u ir e m e n t # 3
• The amou n t of inform a t io n excha n g e d durin g the
verificatio n oper a tio n should not lead to high
com m u n ic a t io n overh e a d
Req u ir e m e n t # 4
• The protoc ol should be comp u t a t i o n a lly efficient
Req u ir e m e n t # 5
• It ought to be possible to run the verificatio n an unlimite d
num b e r of times
But, Privat e clouds canno t easily scale out in the case of peak dem a n d
The integ r a t i o n with public clouds could be a solution to the incre a s e d load
Hybrid clo u d s are the res ult of a privat e cloud growin g and provisionin g
reso u r c e s from a public cloud
• Best option for the futur e in many case s
Res o u r c e Provi s i o n i n g Sc e n a r i o
– Res o u r c e Pool
• A cont ain e r of virtual reso u r c e s that mostly come from the
sam e resou r c e provide r
• In cha r g e of man a gi n g the virtu al reso u r c e s it contai n s
• Finally relea si n g the m when they are no longe r in use
• Each vendo r expos e s its own specific interf ac e s
• Enca p s ul a t e s the specific imple m e n t a t i o n of the
com m u n ic a t io n protocol req ui r e d to inter a c t with it
• Provides the pool man a g e r with a unified interf ac e for
acquiri n g, ter mi n a ti n g, and monitori n g virtu al resou r c e s
• The req u e s t for addition al reso u r c e s is gene r a lly trigg e r e d by a
sche d ul e r
– Detect s that the curr e n t capa city is not sufficient to satisfy
the expe ct e d
quality of service s ensu r e d for specific applica tio n s
• In this case, a provisionin g requ e s t is mad e to the Resou r c e
Provisionin g Servic e
– Accordin g to specific policies, the pool man a g e r det e r m i n e s the
pool insta n c e ( s) that will be used to provision resou r c e s and will
forw a r d the req u e s t to the select e d pools
| 128
– Res o u r c e Ide n t i t y
• The identifie r of a public or a privat e Amazon Machin e
Imag e (AMI) used as tem pl a t e from which to cre a t e virtual
machin e insta n c e s
– Res o u r c e Capa c i ty
• Specifies the differe n t type of insta n c e that will be
deploye d by Amazon EC2
| 129
• Clouds typically have highly dyna mic dem a n d s for reso u r c e s with
highly hete r o g e n e o u s and dyna mi c workloa d s
• Differe n t applic ation s may have very differe n t and dyna mic quality of
servic e (QoS) requir e m e n t s
– One applic ation may req uir e high throu g h p u t
– Anoth e r may be const r ai n e d by a budg e t
– A third may have to balanc e both thro u g h p u t and budg e t
• To suppo r t on- dem a n d scale- up, scale- down, and scale- out
– Combinin g public cloud platfor m s and integ r a t i n g the m with
existing grids and data cent e r s
– Users may want to use resou r c e s in their privat e cloud or data
cent e r or grid first before scaling out onto a public cloud
– Inte g r a t i n g thes e public cloud platfor m s with exiting
comp u t a t i o n a l grids to provide oppor t u n i ti e s for on- dem a n d
scale- up and scale- dow n is clo u d bur s t
COMETCLOU D ARCHITECTU RE
–
Guar a n t e e s that cont e n t- base d que ri e s are deliver e d with
boun d e d costs using flexible cont e n t descri p t o r s in the form of
keywor d s , partial keywor d s , and wildca r d s
– Come t builds a tuple- bas e d coordin a tio n spac e abs t r a c t io n using
Squid
– Can be associa tively acce s s e d by all syste m pee r s withou t
req ui ri n g the location inform a tio n of tuples and host identifier s
– Also provide s tran si e n t spac e s that ena bl e applica tion s to
explicitly exploit cont ext locality
Com e t Spa c e
• The tag, field nam e s , and value s mus t be actu al dat a for a
tuple
• Can contai n wildca r d s (“*”) for a tem pl a t e tuple
– This lightw ei g h t forma t is flexible enou g h to repr e s e n t the
infor m a ti o n for a wide rang e of applica tio n s
• Can suppo r t rich matc hi n g relation s hi p s
• The cros s- platfor m nat u r e of XML make s this form a t
suita bl e for infor m a tio n excha n g e in distrib u t e d
hete r o g e n e o u s enviro n m e n t s
• A tuple in Come t can be retrieve d
– If it exactly or app roxi m a t e ly matc h e s a tem pl a t e tuple
– Exact matc hi n g requir e s the tag and field nam e s of the tem pla t e
tuple to be specified withou t any wildca r d , as in Linda
• This strict matc hi n g patt e r n mus t be relaxe d in highly
dyna mic environ m e n t s
• Applicatio n s like service discove ry may not know exact
tuple struc t u r e s
– Suppo r t s retri ev al s with incom ple t e struc t u r e infor m a tio n using
appr oxi m a t e matc hi n g
• Only requir e s the tag of the tem pl a t e tuple be specified
using a keywor d or a partial keywor d
• Tuple (a) tagg e d “cont a c t”
– Has fields “nam e, phon e, email, dep”
– With values “Smith, 7324 4 5 1 0 0 0 , smith@ g m a il.co m , ece”
– Can be retri ev e d using tuple tem pla t e (b) or (c)
• Each tuple is associa t e d with k keywor d s select e d from its tag and
field nam e s
– The keys of a tuple
– e.g., The keys of tuple (a) can be “nam e, phon e” in a 2D stud e n t
infor m a ti o n spac e
– Tuples are local in the inform a ti o n spac e
• If their keys are lexicogr a p h i c ally close
• Or if they have com m o n keywor d s
– The selectio n of keys can be specified by the applica tio n s
• A Hilbe r t SFC is a locality pres e rvi n g continu o u s map pi n g
– From a k-dimen sio n a l (kD) spac e to a 1D spac e
– Points that are close on the curve are mapp e d from close points
in the kD spac e
– The Hilbe r t curve rea dily exte n d s to any num b e r of dime n sio n s
– Enable s the tuple spac e to maint ain cont e n t locality in the index
spac e
• In Come t, the pee r nodes form a one- dime n sio n al overlay
– Indexe d by a Hilbe r t SFC
– The tuple s are mapp e d from the multi- dime n sio n al inform a t io n
spac e to the linea r peer index spac e
– Come t uses the Hilber t SFC const r u c t s the distrib u t e hash table
(DHT)
• For tuple distrib u tio n and lookup
– If the keys of a tuple only includ e comple t e keywor d s
• The tuple is mapp e d as a point in the inform a ti o n spac e
and locat e d on at most one node
– If its keys consist of partial keywor d s , wildca r d s , or rang e s
• The tuple identifies a region in the inform a tio n spac e
• This region is map p e d to a collection of seg m e n t s on the
SFC
• Corre s p o n d s to a set of point s in the index spac e
– Each node store s the keys that map to the seg m e n t of the curve
betw e e n itself and the pre d e c e s s o r node
• Five node s are indexe d using SFC from 0 to 63
– With id show n in solid circle
– The tuple define d as the point (2, 1) is mapp e d to index 7 on the
SFC
• Corre s p o n d s to node 13
– The tuple define d as the region (23, 15) is map p e d to two
seg m e n t s on the SFC
• Corre s p o n d s to node s 13 and 32
| 138
Auto n o m i c Clou d b u r s t i n g
• The goal of auto no m i c cloudb u r s t s
– To sea ml e s sly and secu r ely integ r a t e privat e ent e r p ri s e clouds
and data cent e r s with public utility clouds on- dem a n d to provide
the abs t r a c t io n of resiza bl e comp u ti n g capacity
– Enable s the dyna mic deploym e n t of applica tio n compo n e n t s
onto a public cloud
• Typically run on inter n al orga niza tio n al comp u t e
reso u r c e s
• To addr e s s dyna mi c workloa d s , spikes in dem a n d s , and
othe r extr e m e requir e m e n t s
• Typical over- provisionin g strat e g i e s are no longe r feasible
– The incre a si n g applica tio n and infras t r u c t u r e scale s
– Their cooling, oper a tio n, and man a g e m e n t costs
– Autono mi c cloud b u r s t s can lever a g e utility clouds
• To provide on- dem a n d scale- out and scale- in capa bilities
bas e d on a rang e of met ric s
• The overall app ro a c h for suppo r ti n g auton o mi c cloudb u r s t s in
Come t Clo u d
– Come t Clo u d c o n si d e r s thre e type s of clouds
• Base d on perc eive d secu rity/t r u s t
– Assigns capa bilities accor di n gly
– The first is a highly trus t e d , robus t, and secu r e cloud
• Usually compo s e d of trus t e d/ s e c u r e node s within an
ente r p r i s e
• Typically used to host mas t e r s and othe r key
(man a g e m e n t , sche d uli n g, monito ri n g) roles
• Also use d to stor e stat e s
– In most applic ation s, the privacy and integ rity of critical dat a
mus t be maint ai n e d
| 139
This make s all the tuple s from the failed node recove r e d
– The pred e c e s s o r node make s a new replica for the local spac e of
its new succ e s s o r
• Also suppo r t fault- toler a n c e in the prog r a m m i n g layer
– Some tasks can be lost during runti m e
• Beca u s e of netw o r k c o n g e s t i o n or task gen e r a ti o n durin g
failur e
– The mast e r checks the spac e periodic ally and reg e n e r a t e s lost
tasks
• Load Balanci n g
– Execu tin g applica tio n requ e s t s on unde rlying grid reso u r c e s
consis ts of two key steps
– The first consist s of crea ti n g VM inst a n c e s to host each
applica tion requ e s t , matc hin g the specific char a c t e r i s tic s and
req ui r e m e n t s of the req u e s t
• Called VM Provisionin g
– The secon d step is map pi n g and sche d ulin g thes e req u e s t s onto
distrib u t e d physic al resou r c e s
• Called Resou r c e Provisionin g
– Most virtualize d dat a cent e r s curr e n tly provide a set of gen e r al-
purpo s e VM class e s with gene ri c reso u r c e configu r a ti o n s
| 144
UNIT – 4
Intro d u c t i o n
– The “pay as you go” billing model applies char g e s for the
actu ally used resou r c e s per unit time.
A Us e cas e
• SAP syste m s are use d for a variety of busine s s applica tion s that differ
by version and function ality
RESERVOIR Archi t e c t u r e
| 150
The fede r a ti o n can offer public IP addr e s s e s rete n tio n post cross- site
migr a tio n. With
| 152
fully virtualize d netwo r k s, this may be a direc tly suppo r t e d featu r e ; but even
if virtu alize d
netw o r k s are not availabl e, it may still be possible to maint ain public IP
addr e s s e s by
manip ul a tin g routin g inform a ti o n
Inform a ti o n disclos u r e within the fede r a ti o n has also to be take n into
accou n t . The sites in the fede r a ti o n may provide inform a ti o n to differe n t
deg r e e s (for insta n c e , the inform a tio n excha n g e betw e e n sites may be
larg e r within the sam e adminis t r a t iv e dom ain tha n outsid e it).
Infor m a ti o n rega r di n g deploye d VEEs will be prim a rily via the monitori n g
syste m, wher e a s some inform a tio n may also pote n ti ally be expos e d via the
VMI as res po n s e to a VEE deploym e n t requ e s t .
Which oper a tio n s are requi r e d may be relat e d to the amou n t of inform a tio n
that is expos e d by the remo t e sites; acces s to more inform a t io n may also
incre a s e the possibility and nee d to manip ul a t e the deploye d VEEs.
Fed e r a t i o n Sc e n a r i o s
This scen a rio offers a useful cloud comp u ti n g fede r a t io n with suppo r t for
site collabo r a tio n in ter m s of fram e w o r k agre e m e n t s withou t partic ul a rly
high tech n ologic al req ui r e m e n t s on the und e rlyin g archit e c t u r e in ter m s of
netw o r ki n g suppo r t .
An exam pl e scen a ri o wher e two web applica tion s , applica tion A and
applica tion B, are host e d on a sepa r a t e set of dedica t e d serve r s within
the ent e r p ri s e- owne d serve r room s is show n in Figur e 16.1. The
plan n e d capa city for each of the applica tion s to run succ e s sf ully is
thre e serve r s . As the num b e r of web applica tion s grew, the serve r
room s in the orga niz a tio n beca m e larg e and such serve r rooms wer e
known as data cent e r s . Thes e dat a cent e r s were owne d and man a g e d
by the ent e r p ri s e s the m s elve s
• As the num b e r and compl exity of the web applica tio n s gre w,
ente r p r i s e s realize d that it was econo mi c al to outsou r c e the
applica tion hosting activity to third- party infras t r u c t u r e provide r s
• Thes e provide r s get the requir e d har d w a r e and make it availabl e for
applica tion hosting.
| 154
Exa m p l e of SLA
• one SLA may stat e that the applica tio n’s serve r machin e will be
availa ble for 99.9% of the key busin e s s hours of the applica tio n’s end
user s , also called core time, and 85% of the non- core time.
• Anoth e r SLA may stat e that the service provide r would res po n d to a
repo r t e d issue in less than 10 minut e s durin g the core time, but would
res po n d in one hour during non- core time.
• Thes e SLAs are know n as the infras t r u c t u r e SLAs, and the
infras t r u c t u r e service provide r s are known as Applica tion Service
Provide r s (ASPs).
• This scen a rio is depict e d in Figur e, wher e t h e ente r p ri s e applic a tion s
are host e d on the dedica t e d serve r s belon gi n g to an ASP.
This scen a rio is show n in Figur e , whe r e both applic ation A and
applica tion B shar e the sam e set of virtu alize d serve r s .
The load bala nci n g algorit h m execu t e s on a physic al mac hi n e that interfa c e s
with the client s. This physical mac hin e , also called the front- end node,
receive s the incomin g requ e s t s and distrib u t e s thes e requ e s t s to differ e n t
physical machi n e s for furth e r execu tion.
This set of physical mac hin e s is res po n si bl e for servin g the incomi n g
req u e s t s and are known as the back- end node s.
• Clas s - ag n o s t i c : The front- end node is neith e r awa r e of the type of client
from which the req u e s t origina t e s nor awa r e of the cate g o r y (e.g., brow si n g,
selling, paym e n t , etc.) to which the requ e s t belon g s to.
• Clas s - awar e : The front- end node must addition ally inspe c t the type of
client making the requ e s t and/or the type of service requ e s t e d before
deciding which back- end node should service the requ e s t .
• Admi s s i o n Contr o l: Decidin g the set of requ e s t s that should be admit t e d
into the applic ation serve r when the serve r expe ri e n c e s “very” heavy loads.
During overloa d situa tion s, since the res po n s e time for all the requ e s t s
would invaria bly degr a d e if all the arriving requ e s t s are admitt e d into the
serve r, it would be prefe r a b l e to be selective in identifying a subs e t of
req u e s t s that should be admitt e d into the syste m so that the overall pay- off
is high.
• The objective of admis sio n cont rol mec h a ni s m s , ther efo r e , is to police the
incomin g requ e s t s and identify a subs e t of incomin g req u e s t s that can be
admitt e d into the syste m when the syste m faces overloa d situa tio n s .
| 157
Request-based admission control algorithms reject new requests if the servers are running to
their capacity.
The disadv a n t a g e with this appr o a c h is that a client’s session may consistof
multiple req u e s t s that are not nec e s s a rily unrel a t e d . Cons e q u e n t ly, some
req u e s t s are reject e d even if the r e are othe r s that are honor e d .
TYPES OF SLA
• Servic e- level agre e m e n t provide s a fram e w o r k within which both
seller and buye r of a service can purs u e a profita bl e service busin e s s
relation s hi p.
• It outline s the broa d und e r s t a n d i n g betw e e n the servic e provide r and
the service consu m e r for cond u c ti n g busine s s and forms the basis for
maint ai nin g a mut u ally beneficial relation s hi p.
• From a legal pers p e c tiv e, the nece s s a r y ter m s and condition s that
bind the servic e provide r to provide service s contin u a lly to the service
cons u m e r are form ally define d in SLA.
• SLA can be modele d using web service- level agre e m e n t (WSLA)
lang u a g e Specifica tio n. Service- level para m e t e r , met ric, function,
| 158
meas u r e m e n t direc tive, service- level objective, and penalty are som e of the
impor t a n t compo n e n t s of WSLA
There are two type s of SLAs from the pers p e c tive of applic ation hosting
Ente r p ri s e s man a g e the m s elv e s , their applic ation s that are deploye d on
thes e serve r mac hin e s . The mac hin e s are leas e d to the custo m e r s and are
isolat e d from mac hi n e s of othe r custo m e r s .
Appli c a t i o n SLA : In the applic a tion co- location hosting model, the serve r
capacity is availabl e to the applica tion s bas e d solely on their reso u r c e
dem a n d s .
Henc e, the servic e provide r s are flexible in alloca tin g and de-
allocatin g c o m p u t i n g reso u r c e s amon g the co- locat e d applic ation s.
Therefor e , the servic e provide r s are also res po n si ble for ens u ri n g to mee t
their custo m e r’s applica tion SLOs
The service provide r nee d s to analyze the applic ation’s beh avior with
res p e c t to
scala bility and perfor m a n c e before agre ei n g on the specifica tion of SLA. At
the end of
this pha s e , the SLA is mut u ally agre e d by both custo m e r and provide r and is
event u a lly
signe d off.
II . On- Boar d i n g of Appli c a t i o n : Once the custo m e r and the MSP agre e in
principle to host the applica tio n bas e d on the findings of the feasibility
study, the applica tion is moved from the custo m e r serve r s to the hosting
platfor m . Moving an applica tio n to the MSP’s hostin g platfor m is called on-
boar d i n g
As part of the on- boar di n g activity, the MSP unde r s t a n d s the applic ation
runti m e char a c t e r i s ti c s using runtim e profilers.
This helps the MSP to identify the possible SLAs that can be offere d to the
custo m e r for that applica tio n.
This also helps in crea tio n of the nece s s a r y policies (also called rule sets)
req ui r e d to guar a n t e e the SLOs mention e d in the applica tio n SLA.
The applica tion is acce s si bl e to its end user s only after the onboa r di n g
activity is comple t e d
perfor m a n c e char a c t e r i s tic s like the applic ation’s ability to scale (out and
up) and perfor m a n c e boun d s (minim u m and maxim u m perfor m a n c e ) are
note d
d. Based on the meas u r e d perfor m a n c e char a c t e r i s ti c s, differ e n t possible
SLAs are identified. The reso u r c e s req ui r e d and the costs involved for eac h
SLA are also comp u t e d
e. Once the custo m e r agre e s to the set of SLOs and the cost, the MSP
start s crea ti n g differe n t policies requir e d by the data cent e r for auto m a t e d
man a g e m e n t of the applic ation. Thes e policies are of thre e types:
(1) Busine s s policies help prioritize acce s s to the reso u r c e s in case of
cont e n tio n s . Busine s s policies are in the form of weigh t s for differe n t
custo m e r s or group of custo m e r s .
(2) Oper a tio n a l policies are the actions to be take n when differe n t
thre s h ol d s/ co n di tio n s are reac h e d . Also, the actions whe n thre s h ol d s/
conditions /t ri g g e r s on service- level para m e t e r s are brea c h e d or about to be
bre a c h e d are define d.
(3) Provisionin g: The corr e c tiv e action could be differe n t types of
provisionin g such as scale- up, scale- down, scale- out, scale- in, and so on, of a
partic ul a r tier of an applica tio n. Addition ally, notifica tio n and loggin g action
are also define d.
Oper a tio n al policies (OP) are rep r e s e n t e d in the following form a t:
OP = collection of <Co n di tio n, Action >
Ex: OP < av e r a g e laten cy of web serve r > 0.8 sec, scale- out the web- serve r
tier >
It mea n s , if aver a g e laten cy of the web serve r is more than 0.8 sec then
auto m a t i c ally scale out the web- serve r tier. On reac hin g this thre s h ol d, MSP
should incre a s e the num b e r of insta n c e s of the web serve r.
Additionally, custo m e r may req u e s t the MSP for inclusion of new ter m s and
conditions in the SLA. If the applic ation SLA is bre a c h e d frequ e n tly or if the
custo m e r requ e s t s for a new non- agre e d SLA, the on- boar din g proce s s is
perfor m e d again.
In the case of the form e r , on- boar di n g activity is repe a t e d to analyze the
applica tion and its policies with res p e c t to SLA fulfillme n t.
In case of the latte r, a new set of policies are form ul a t e d to meet the fresh
ter m s and condition s of the SLA.
V. Ter m i n a t i o n : When the custo m e r wishes to withd r a w the host e d
applica tion and does not wish to contin u e to avail the servic es of the MSP
for man a gi n g the hosting of its applica tion, the ter mi n a tio n activity is
initiat e d.
On initiation of termin a tio n, all dat a relat e d to the applica tio n are
tran sfe r r e d to the custo m e r and only the ess e n ti al inform a ti o n is retain e d
for legal complia n c e .
This ends the hostin g relation s hi p betw e e n the two partie s for that
applica tion, and the custo m e r sign- off is obtain e d
| 165
HPC on Clou d
• But, IaaS lets user s run applica tio n s on fast pay- per- use mac hi n e s
they don’t want to buy, to man a g e , or to maint ai n.
• For the HPC use r, this solution is undou b t e dly attr a c tive:
• HPC users usu ally exploit parallel har d w a r e , and so they would like to
get par allel har d w a r e to execu t e their explicitly- par allel applica tio n s .
• Stat e d anot h e r way, they exploit the cloud as a provide r of clust e r on-
dem a n d (CoD) syste m s .
The adoptio n of the cloud par a di g m in HPC is relat e d to the evalua tio n
(and, possibly, to the redu c tio n) of possible perfor m a n c e losses comp a r e d
to physical HPC hard w a r e .
The actu al hard w a r e use d in the cloud, along with the losses at the VE
and CE
levels, will dete r mi n e the actu al perfor m a n c e of applica tion s runnin g in
the
cloud
• The main proble m with physic al clust e r s is that all jobs runnin g on the
clust e r , whet h e r concu r r e n t or non- conc u r r e n t , have to shar e the
sam e oper a ti n g syste m (OS), the syste m and applica tio n libra ri e s , and
the oper a ti n g enviro n m e n t .
• The freq u e n tly recu r ri n g req ui r e m e n t s for mut u ally exclusive or
incom p a t i bl e libra ri e s and suppo r t softw a r e make physical cluste r
man a g e m e n t a night m a r e for syste m adminis t r a t o r s .
Virtual clust e r may have an exec ution environ m e n t of its own (OS,
libra ri e s, tools, etc.) that is loade d and initialize d whe n the clust e r is
cre a t e d .
Given a physical node provide d with n CPUs, ther e are two possibilities
to exploit all the comp u ti n g resou r c e s available:
Using n VMs (each running its OS instance) with one, or even several, VCPUs;
Using a single VM with at least n VCPU
Grid vs Clou d
On the othe r hand, cloud tech nolo gy was desig n e d using a top- down
appr o a c h .
• It aims at providin g its use rs with a specific high- level function ality:
a stor a g e , a comp u ti n g platfor m , a specialize d servic e. They get virtual
reso u r c e s from
the cloud.
• The und e rlyin g har d w a r e / s oft w a r e infras t r u c t u r e is not expos e d. The
only inform a ti o n the use r nee d s to know is the quality of service (QoS)
of the servic es he is paying for. Band wid t h, comp u ti n g powe r, and
stora g e rep r e s e n t para m e t e r s that are used for specifying the QoS
and for billing.
• Cloud users ask for a high- level function ality (service, platfor m ,
infras t r u c t u r e ) , pay for it, and beco m e owne r s of a virtu al machi n e .
• A single ente r p ri s e is the owne r of the cloud platfor m (softw a r e and
und e rlyin g hard w a r e ), whe r e a s custo m e r s beco m e owne r s of the
virtu al resou r c e s they pay for.
• Cloud suppo r t e r s claim that the cloud is easy to be use d, is scala ble,
and always gives user s exactly what they want.
• On the othe r hand, grid is difficult to be use d, does not give
perfor m a n c e gua r a n t e e s , is used by narr o w comm u ni tie s of scientis t s
| 170
• Grid suppo r t e r s answ e r that grid user s do not nee d a credit card, that
arou n d the world ther e are many exam pl e s of succe s sf ul project s, and
that a gre a t num b e r of comp u ti n g nodes conn e c t e d acros s the net
execu t e larg e scale scientific applica tion s , add r e s s i n g proble m s that
could not be solved othe r wi s e
The adoption of the cloud paradigm for HPC is a flexible way to deploy (virtual) clusters dedicated to
execute HPC applications
The first and well-known difference between HPC and cloud environmentsis the different economic
approach:
In the latter, every time that a task is started, the user will be charged for the used resources. But it
is very hard to know in advance which will be the resource usage and hence the cost. On the other
hand, even if the global expense for a physical cluster is higher, once the system has been acquired,
all the costs are fixed and predictable
| 171
In clouds, performance counts two times. Low performance means not only long waiting times, but
also high costs
The use of alternative cost factors (e.g., the RAM memory allocated, as for GoGrid, leads to
completely different considerations and requires different application optimizations to reduce the
final cost of execution.
The typical HPC user would like to know how long his application will run on the target cluster and
which configuration has the highest performance/cost ratio.
The advanced user, on the other hand, would also know if there is a way to optimize its application
so as to reduce the cost of its run without sacrificing performance.
The high-end user, who cares more for performance than for the cost to be sustained, would like
instead to know how to choose the best configuration to maximize the performance of his
application.
The system dimensioning is the choice of the system configuration fit for the
technology and to be able to sustain the highest system usage that may eventually be required. This
can be measured in terms of GFLOPS, in terms of number of runnable jobs, or by other indexes
depending on the HPC applications that will be actually executed.
In other words, the dimensioning is made by considering the peak system usage. It takes place at
system acquisition time, by examining the machine specifications or by assembling it using hardware
components of known performance.
| 172
In clouds, instead, the system must be dimensioned by finding out an optimal trade-off between
application performance and used resources.
The optimality is a concept that is fairly different, depending on the class of users.
Someone would like to obtain high performance at any cost, whereas others would privilege
economic factors
To support HPC applications, a fundamental requirement from a cloud provider is that an adequate
service-level agreement (SLA) is granted.
For HPC applications, the SLA should be different from the ones currently offered for the most
common uses of cloud systems, oriented at transactional Web applications.
The SLA should offer guarantees useful for the HPC user to predict his application performance
behavior and hence to give formal (or semiformal) statements about the parameters involved.
At the state of the art, cloud providers offer their SLAs in the form of a contract (hence in natural
language, with no formal specification).
| 173
There are some clear busine s s ben efits to building applic ation s in the cloud:
• Just- in- Tim e Infra s t r u c t u r e : By deploying applic ation s in- the- cloud
with just- in- time self- provisionin g, we do not have to worry about pre-
proc u ri n g capa city for large- scale syste m s . This incre a s e s agility,
lower s risk, and lower s oper a tio n al cost beca u s e you scale only as you
grow and only pay for what you use.
• Usa g e - Bas e d Costi n g : With utility- style pricing, you are billed only
for the infras t r u c t u r e that has bee n used
• Reduced Time to Market: Parallelization is one of the great ways to speed up processing. Having
available an elastic infrastructure provides the application with the ability to exploit
parallelization in a cost-effective manner reducing time to market
• The Amazon Web Service s (AWS) cloud provide s a highly reliable and
scala bl e infras t r u c t u r e for deploying Web- scale solution s, with
minim al suppo r t and administ r a t io n costs, and more flexibility than we
expe ct from our own infra s t r u c t u r e , eithe r on- pre mis e or at a
dat ac e n t e r facility
dem a n d insta n c e or spot insta n c e s whe r e they can bid for unus e d
capacity and furth e r red uc e cost
• An aut o- scali n g gro u p can be cre a t e d using the auto- scaling featu r e
to auto m a ti c ally scale capacity on cert ain conditions base d on met ric
that Amazon CloudW a t c h collects
• AWS also offers various paym e n t and billing service s that levera g e s
Amazon’s paym e n t infra s t r u c t u r e . All AWS infras t r u c t u r e service s
offer utility- style pricing that requir e no long ter m commit m e n t s or
cont r a c t s .
The following best practices will help to build an application in the cloud
Rule of Thumb: Be a pessimist when designing architectures in the cloud; assume things will fail.
In other words, always design, implement, and deploy for automated recovery from failure.
1. Have a coherent backup and restore strategy for your data and automate it.
3. Allow the state of the system to re-sync by reloading messages from queues.
| 178
5. Avoid in-memory sessions or stateful user context; move that to data stores
1. Failover gracefully using Elastic IPs: Elastic IP is a static IP that is dynamically remappable
2. Utilize multiple availability zones: Availability zones are conceptually like logical
datacenters.
3. Maintain an Amazon Machine Image so that you can restore and clone environments very
easily in a different availability zone; maintain multiple database slaves across availability zones
and set up hot replication.
4. Utilize Amazon Cloud Watch to get more visibility and take appropriate actions in case of
hardware failure or performance degradation.
5. Utilize Amazon EBS and set up jobs so that incremental snapshots are automatically
uploaded to Amazon S3 and data are persisted independent of instances.
6. Utilize Amazon RDS and set the retention period for backups, so that it can perform
automated backups.
• Decouple your Components: The cloud reinforces the SOA design principle that the more
loosely coupled the components of the system, the bigger and better it scales.
• The key is to build components that do not have tight dependencies on each other, so that if one
component were to die (fail), sleep (not respond), or remain busy (slow to respond) for some
reason, the other components in the system are built so as to continue to work as if no failure is
happening.
• Decoupling components, building asynchronous systems, and scaling horizontally become very
important in the context of the cloud. It will not only allow scaling out by adding more instances
of same component but will also allow to design innovative hybrid models in which a few
components continue to run in on-premise while other components can take advantage of the
cloud scale and use the cloud for additional compute-power and bandwidth.
One can build a loosely couple d syste m using mes s a gi n g queu e s . If a queu e/
buffer is used to conn e c t any two compo n e n t s toge t h e r (Loose Coupling), it
can suppo r t concu r r e n c y, high availability, and load spike s.
As a res ult, the over all syste m continu e s to perfor m even if part s of
compo n e n t s are mom e n t a r ily unavaila bl e
2. Proa c t i v e Even t- Bas e d Scal i n g : Scaling just when expec tin g a big
surg e of traffic requ e s t s due to a sche d ul e d busin e s s event (new produ c t
launc h, mark e ti n g cam p ai g n s ).
1. Define auto- scaling group s for differ e n t clust e r s using the Amazon
auto- scaling featu r e in Amazon EC2.
2. Monito r syste m met rics (CPU, me mo ry, disk I/O, netw o r k I/O) using
Amazon CloudW a t c h and take appro p ri a t e actions (launc hi n g new AMIs
dyna mic ally using the auto- scaling service) or send notification s .
3. Store and ret ri eve machin e configu r a ti o n inform a tio n dyna mi c ally:
Utilize Amazon SimpleDB to fetch config data durin g the boo t-time of
an insta n c e (e.g., dat a b a s e conn e c tio n string s ). SimpleDB may also be used
to stor e inform a tio n about an insta n c e such as its IP addr e s s , mac hi n e
nam e, and role. \
4. Design a build proc e s s such that it dum p s the lates t builds to a bucke t
in Amazon S3;
| 180
7. Reduc e bundling and launc h time by booting from Amazon EBS volum e s
and atta c hi n g multiple Amazon EBS volum e s to an insta n c e . Crea t e
snap s h o t s of com m o n volum e s and shar e sna ps h o t s amon g accou n t s
whe r e v e r app ro p ri a t e .
A gen e r al best prac tic e, in the case of a Web applic ation, is to distrib u t e the
incomin g req u e s t s acros s multiple Web serve r s using load bala nc e r
• 4. Use the Elastic Load Balanci n g servic e and spre a d load acros s
multiple Web app serve r s dyna mic ally.
• Conve r s e ly, if the dat a are static and not going to chan g e often, it is
advisa bl e to take adva n t a g e of a conte n t delivery service so that the
static data are cach e d at an edge location close r to the end user
(req u e s t e r ), ther e b y loweri n g the acces s laten cy.
1. Ship dat a drives to Amazon using the Impor t/Exp o r t service. It may be
che a p e r and faste r to move large amou n t s of data than to uploa d using the
Inte r n e t .
Custo m e r s are char g e d only for their utilization of stor a g e and tran sf e r of
cont e n t
Custo m e r s are char g e d only for their utilization of stor a g e and tran sf e r of
cont e n t
• Nirva nix launc h e d its Amazon S3 comp e ti to r, the Nirva nix Stor a g e
Delivery Netw o r k (SDN), on Sept e m b e r 2007.
• Nirva nix is price d slightly highe r than Amazon’s servic e, and they do
not publish their pricing rate s for larg e r custo m e r s (2 TB/mont h).
• Racks p a c e (form e rly Mosso) Cloud Files provides a self- serve stor a g e
and delivery service in a fashion similar to that of the Amazon and
Nirva nix offering s .
• Rath e r tha n building their own CDN exte n sio n to the Cloud Files
platfor m as Amazon has done for S3, Racks p a c e has part n e r e d with a
tradition al CDN service, Limelight, to distrib u t e files store d on the
Cloud Files platfor m to edge nodes oper a t e d by Limeligh t.
• This CDN exte n sio n is still unde r testin g and is curr e n tly being offere d
to custo m e r s as a Comm u ni ty Techno logy Preview (CTP) at no
cha r g e .
• Most “stor a g e cloud” provide r s are mer ely basic file stora g e and
delivery service s and do not offer the capa bilities of a fully featu r e d
CDN such as auto m a ti c replica tio n, fail- over, geog r a p h i c al load
redir e c tio n, and load bala nci n g.
• The MetaCDN service is presented to end users in two ways. First, it can be presented as a Web
portal, which was developed using
(a) Java Enterprise and Java Server Faces (JSF) technologies, with a
The Web portal acts as the entry point to the system and also functions as an application-level
load balancer for end users that wish to download content that has been deployed by MetaCDN.
| 187
Using the Web portal, users can sign up for an account on the MetaCDN system and enter
credentials for any cloud storage or other provider they have an account with.
Once this simple step has been performed, they can utilize the MetaCDN system to intelligently
deploy content onto storage providers according to their performance requirements and budget
limitations.
The Web portal is most suited for small or ad hoc deployments and is especially useful for less
technically inclined content creators.
• The second method of accessing the MetaCDN service is via RESTful Web Services.
• These Web Services expose all of the functionality of the MetaCDN system.
• This access method is most suited for customers with more complex and frequently changing
content delivery needs, allowing them to integrate the MetaCDN service in their own origin Web
sites and content creation workflows.
• The MetaCDN system works by integrating with each storage provider via connectors that
provides an abstraction to hide the complexity arising from the differences in how each provider
allows access to their systems.
• An abstract class, DefaultConnector, prescribes the basic functionality that each provider could
be expected to support, and it must be implemented for all existing and future connectors.
• These include basic operations like creation, deletion, and renaming of replicated files and
folders.
• If an operation is not supported on a particular service, then the connector for that service
throws a FeatureNotSupportedException.
• The MetaCDN service has a number of core components that contain the logic and management
layers required to encapsulate the functionality of different upstream storage providers and
present a consistent, unified view of the services available to end users.
• These components include the MetaCDN Allocator, which (a) selects the optimal providers to
deploy content to and (b) performs the actual physical deployment.
| 188
• The MetaCDNQoS monitor tracks the current and historical performance of participating
storage providers
• The MetaCDN Manager tracks each user’s current deployment and performs various
housekeeping tasks.
• The MetaCDN Database stores crucial information needed by the MetaCDN portal, ensuring
reliable and persistent operation of the system. It also stores information needed by the
MetaCDNsystem, such as MetaCDN user details, their credentials for various storage cloud and
other providers, and information tracking their (origin) content and any replicas made of such
content
• The MetaCDN Load Redirector is responsible for directing MetaCDN end users (i.e., content
consumers) to the most appropriate file replica, ensuring good performance at all times
• The MetaCDN Allocator allows users to deploy files either directly (uploading a file from their
local file system) or from an already publicly accessible origin Web site (sideloading the file,
where the backend storage provider pulls the file).
• MetaCDN users are given a number of different deployment options depending on their needs,
regardless of whether they access the service via the Web portal or via Web services.
• Maximize coverage and performance, where MetaCDN deploys as many replicas as possible to
all available locations. The MetaCDN Load Redirector directs end users to the closest physical
replica.
• Deploy content in specific locations, where a user nominates regions and MetaCDN matches the
requested regions with providers that service those areas. The MetaCDNLoad Redirector directs
end users to the closest physical replica.
• After MetaCDN deploys replicas using one of the above options, it stores pertinent details such
as the provider used, the URL of the replica, the desired lifetime of the replica, and the physical
location (latitude and longitude) of that deployment in the MetaCDN Database.
• A geolocation service is used to find the latitude and longitude of where the file is stored.
• The MetaCDNQoS Monitor tracks the performance of participating providers (and their
available storage and delivery locations) periodically, monitoring and recording performance and
reliability metrics from a variety of locations, which is used for QoS-optimized deployment
matching.
• This component also ensures that upstream providers are meeting their service-level
agreements (SLAs), and it provides a logging audit trail to allow end users to claim credit in the
event that the SLA is broken
• First, it ensures that all current deployments are meeting QoS targets of users that have made
QoS optimized deployments.
• Second, it ensures that replicas are removed when no longer required (i.e., the “deploy until”
date set by the user has expired), ensuring that storage costs are minimized at all times.
| 190
• Third, for users that have made cost-optimized deployments, it ensures that a user’s budget has
not been exceeded, by tracking usage (i.e., storage and downloads) from auditing information
provided by upstream providers.
The initial cloud provide r s simply open e d their existing infras t r u c t u r e to the
custo m e r s and thus exploite d their resp e c tiv e propri e t a r y solution s.
Implicitly, the offere d service s and henc e the accor di n g API are specific to
the servic e provide r and canno t be used in othe r environ m e n t s .
This, how ev e r, poses major issue s for custo m e r s , as well as for futur e
provide r s .
Platfor m as a Service (PaaS) provide r s often offer speci alize d capa bilities to
their users via a dedica t e d API, such as Google App Engine providing
Additional featu r e s for handling (Google) docu m e n t s , and MSAzur e is
focusin g partic ul a rly on deploym e n t and provisionin g of Web service s, and
so on.
The Cloud Comp u ti n g Expe r t Working Group refe rs to such integ r a t e d cloud
syste m s with agg r e g a t e d capa bilitie s acros s the individu al infra s t r u c t u r e s as
Met a- Clouds and Met a- Servic es
Be n e f i t s of Clou d Mas h u p s
As clouds develop into applic ation platfor m s , cont ext such as user
device prop e r t i e s or location beco m e s more and more releva n t : Device types
design a t e the execu tio n capa bilities (even if remo t e ), their conn e c tivity
req ui r e m e n t s and rest ric tio n s , and the location . Each of thes e aspe c t s has a
direct impa c t on how the cloud nee d s to handl e data and applica tio n
location, com m u ni c a ti o n
• A comp a ti bl e virtual machin e, resp e c tiv ely an imag e form a t that all
accor di n g cloud infras t r u c t u r e s can host (IaaS).
| 192
• for exam pl e, when too many user s acce s s the syste m concu r r e n t ly and
execu t e a load balan c e betw e e n the two locations. Simila rly, an ideal
syste m would dow n- scale the replic a t e d units once the resou r c e load
is redu c e d again
for exam pl e, it would not be sensible to dest r oy the whole imag e if only one
user (of many) logs out from the mac hi n e .
(2) the req ui r e m e n t s of user s/ a p plic a tio n s into accou n t (i.e., dat a shall be
distrib u t e d accor di n g to the inter e s t in the data/infor m a ti o n).
For this rea so n, use r s, devices, and applica tion s nee d to be model e d by
capt u ri n g releva n t context para m e t e r s (e.g., the actu al position and netw o r k
prop e r ti e s) as well as analyzing applica tio n stat e s with res p e c t to upcomin g
dat a ret riev al and/or proce s si n g nee d s In addition, stor a g e reso u r c e s ,
platfor m s , and infras t r u c t u r e s (i.e., entir e virtual imag e s ) shall also be
contin u o u sly monitor e d , so as to reac t on sudd e n bottle n e c k s imme di a t e ly
we can disting uis h betw e e n the base imag e set consis tin g of
(a) the setu p environ m e n t and any engin e (if requir e d),
| 194
(b) the bas e data s e t that may be custo m e r- specific (but not user-
specific), such as gen e r al data that are provide d to the use r, but also and
more import a n t ly the apple t or servic e bas e that is provide d to each user
equ ally, and
(c) the user- specific inform a ti o n which may differ per acces s and which
may only be availabl e on a single mac hi n e .
• In othe r words, mac hin e s with the res p e c tive comp a ti bl e bas e imag e
can host the replica t e d service insta n c e s , rath e r than having to
duplic at e the full imag e all the time.
UNIT - 5
Chan g e can be challe n gin g; it brings out the fear of having to deal with
unce r t a i n ti e s . This is the FUD syn dr o m e : Fear, Unc e r t a i n t y , an d
Dou b t Driver s For Cha n g e s : A Fra m e w o r k To Com p r e h e n d th e
Com p e t i t i v e Enviro n m e n t :
The five driving factor s for chan g e enc a p s ul a t e d by the fram e w o r k are:
The res ult s will help the busin e s s to make bett e r decisions, and it will also
help shap e the short- and long- term stra t e gi e s of that busine s s .
It is this proce s s that helps reveal the import a n t factor s for the
orga niza tio n’s desir a bl e futur e stat e, and it helps the orga niz a tio n to
comp r e h e n d which driving force s will chan g e the comp e ti tive landsc a p e in
the indus t ry the busine s s is in, identify critical unc e r t ai n ti e s , and recog nize
what part of the futur e is pred e t e r m i n e d such that it will happ e n rega r dl e s s
how the futur e will play out
Followin g are sam pl e ques tio n s that could help in und e r s t a n d i n g Econo mic
factor s
• What will the econo my looks like in 1 year, 2 years, 3 years, 5 years,
and so on?
• What are some of the factors that will influen c e the futur e econo mi c
outlook?
• How does this tech nolo gy tra n s c e n d the existing busin e s s model?
• How can IT initiatives help and suppo r t orga niz a tio n a l initiative s to
red uc e carbo n footprin t ?
• Can orga niz a tio n s and corpo r a t io n s lever a g e inform a ti o n techn ology,
includin g cloud comp u ti n g to purs u e sustai n a bl e develop m e n t
• How does this eme r gi n g tech nolo gy (cloud comp u ti n g) open up new
are a s for innova tio n?
• How can an applica tion be built once so it can configu r e dyna mi c ally
in real time to oper a t e most effectively, bas e d on the situa tio n al
const r a i n t (e.g., out in the cloud some w h e r e , you might have
band wi d t h cons t r a i n t to tran sf e r need e d dat a)?
| 198
• Lewin obse rv e d that ther e are thre e stag e s of chan g e , which are:
Unfre e z e , Transition, and Refre e z e .
| 199
• The tran sition phas e is whe n the chan g e (plan) is execu t e d and actu al
cha n g e is being imple m e n t e d . Since thes e “activities” take time to be
comple t e d , the proce s s and orga niz a tio n al struc t u r e may also nee d to
cha n g e , specific jobs may also cha n g e . The most resist a n c e to chan g e
may be expe ri e n c e d during this tran sitio n period. This is whe n
lead e r s h i p is critical for the chan g e proc e s s to succ e e d , and
motivation al factor s are par a m o u n t to projec t succe s s .
• The last phas e is Refre ez e; this is the stag e when the orga niza tio n
once again beco m e s unch a n g i n g /froz e n until the next time a cha n g e is
initiat e d
De m i n g ’ s PDCA cycl e
CROPS worki n g mo d e l :
Cultu r e , Rewa r d s , Orga niz a tio n and Struc t u r e s , Proc e s s, Skills and
Comp e t e n c i e s (CROPS) fram e w o r k.
• Storie s and myths about the history of the orga niz a tio n
• Nor m s—t h e feelings evoke d by the way mem b e r s inter a c t with each
othe r, with outsid e r s , and with their environ m e n t
• Orga niz a tio n and Struc t u r e s . How the orga niz a tio n is struc t u r e d is
larg ely influen c e d by what the jobs are and how the jobs are
perfor m e d . The desig n of the busine s s proc e s s e s gover n wha t the jobs
are, and whe n and wher e they get done.
Busine s s proce s s e s nee d to align with orga niz a tio n a l vision, mission, and
stra t e g i e s in orde r to cre a t e custo m e r and sha r e h ol d e r values. Therefo r e , all
the compo n e n t s of the CROPS fram e w o r k are inter r el a t e d .
| 202
A proce s s is wher e the work gets done, and value crea tion occur s throu g h
tran sfor m i n g input into outpu t
CROPS fra m e w o r k
• A Chan g e Man a g e m e n t Matu rity Model (CMMM) helps orga niza tio n s
to
• The working model is bas e d on CMM (Cap a bility Matu rity Model),
originally develop e d by Americ a n Softw a r e Engine e ri n g Instit ut e (SEI)
in coope r a t io n with Mitre Corpo r a tio n.
| 203
for exam pl e
whe r e
• CMMM invest m e n t
Level 5
De s c r i p t i o n : Optimize d
Level 4
Des c r i p t i o n : Mana g e d
Level 3
Des c r i p t i o n : Define d
Level 2
Des c r i p t i o n : Repe a t a b l e
Level 1
• It also define s the road m a p (stra t e g i e s and tactics) req ui r e d to fill the
gap and to get the orga niz a tio n moving towa r d wher e it want s to go
(futur e stat e) from its curr e n t stat e.
| 206
• The proce s s implies that the orga niza tio n nee d s to compl et e the
stra t e g y analysis proce s s first and to form ul a t e the futur e goals and
objectives that suppo r t the futur e direc tion of the busine s s
orga niza tio n.
The asse s s m e n t should provide insight into challe n g e s and help dete r m i n e
som e of thes e key ques tio n s :
• Does orga niz a tio n have the capa ci ty to execu t e and imple m e n t
cha n g e s ?
• Are all employe e s in the orga niz a tio n rea dy to adopt chan g e s that
help realize the vision?
The Jeric h o Foru m was found e d in 2004 beca u s e of the incre a si n g nee d for
dat a excha n g e betw e e n comp a ni e s and exter n a l partie s—
The idea of cre a ti n g, ess e n ti ally, de- cent r a liz e d perim e t e r s , whe r e the
perim e t e r s are cre a t e d by the dat a object itself, allows the secu ri ty to move
with the dat a, as oppos e d to retainin g the dat a within a secu r e d and static
wall
The most obvious risk in this scen a ri o is that associa t e d with the stora g e of
that dat a. A use r uploa di n g or crea ti n g cloud- bas e d data includ e thos e data
that are stor e d and maint ai n e d by a third- party cloud provide r such as
Google, Amazon, Microsoft, and so on.
• Firstly, it is nece s s a r y to prot e c t the dat a during uploa d into the dat a
cent e r to ensu r e that the dat a do not get hijack e d on the way into the
dat a b a s e .
• Secon dly, it is nece s s a r y to the store s the dat a in the data cent e r to
ens u r e that they are encryp t e d at all times.
| 208
• Thirdly, and perh a p s less obvious, the acce s s to thos e data nee d to be
cont rolle d; this cont rol should also be applied to the hosting comp a n y,
includin g the admi nis t r a t o r s of the data cent e r.
Data secu ri ty risks are compo u n d e d by the open natu r e of cloud comp u ti n g .
Acces s control beco m e s a muc h more fund a m e n t a l issue in cloud- bas e d
syste m s beca u s e of the acce s si bility of the data
Inform a ti o n- cent ric acces s cont rol (as oppos e d to acces s cont rol lists) can
help to balanc e improv e d acce s si bility with risk, by associa tin g acces s rules
with differe n t dat a object s within an open and acces si bl e platfor m , witho u t
losing the Inher e n t usability of that platfor m .
A furth e r are a of risk associa t e d not only with cloud comp u ti n g , but also
with tradition al net wo r k comp u ti n g, is the use of cont e n t afte r acce s s.
The risk is pote n ti ally high e r in a cloud netwo r k, for the simple rea so n that
the inform a tio n is outsid e of your corpo r a t e walls.
Encryp tion is a vital compo n e n t of the prot e c tio n policy, but furth e r controls
over the acce s s of that data and on the use of the data must be met. In the
case of mas h u p s , the cont rolling of acces s to data reso u r c e s , can help
alleviat e the secu rity conce r n s by ens u ri n g that mas h u p acces s is
aut h e n t i c a t e d . Linking secu rity policies, as applie d to the use of conte n t , to
the acce s s cont rol met ho d offer a way of contin ui n g prot e c tio n of data, post
acce s s and thro u g h o u t the life cycle; this type of dat a secu rity philosop hy
mus t be incor po r a t e d into the use of cloud comp u ti n g to alleviat e secu rity
risks.
Digi t a l ide n t i t y holds the key to flexible dat a security within a cloud
Environ m e n t . A digital identity rep r e s e n t s who we are and how we inte r a c t
with othe r s on- line.
| 209
Acc e s s , ide n t i t y, and risk are thre e variabl e s that can beco m e inhe r e n t ly
conn e c t e d when applie d to the secu ri ty of dat a, bec a u s e acces s and risk are
directly propo r tio n a l: As acce s s incre a s e s , so then risk to the secu rity of the
dat a incre a s e s . Access cont rolle d by identifying the actor atte m p ti n g the
acce s s is the most logical man n e r of perfor mi n g this oper a tio n. Ultim a t e ly,
digital identity holds the key to secu ri n g data, if that digital identity can be
prog r a m m a t i c a lly linked to secu ri ty policies contr olling the post- acce s s
usag e of dat a.
Our basic societ al comm u ni c a tio n struc t u r e is built upon the idea of
rep u t a t io n and trust.
Reput a tio n and its count e r value, trus t, is easily tran sfe r a bl e to a digital
realm:
eBay, for exam pl e, having partly built a succe s sf ul busin e s s model on the
stre n g t h of a rating s syste m, builds up the rep u t a t io n of its buye r s and
sellers thro u g h succ e s sf ul (or uns uc c e s sf ul) tran s a c ti o n s . Thes e type s of
rep u t a t io n syste m s can be extre m e ly useful whe n use d with a digital
identity. They can be use d to associa t e varying levels of trus t with that
identity, which in turn can be used to define the level (gra n ul a r variation s)
of secu rity policy applied to dat a resou r c e s that the individu al wishe s to
acce s s .
This reve r s al of owne r s hi p away from cent r ally man a g e d identity platfor m s
(ente r p r i s e- cent ric) has many adva n t a g e s . This includ e s the pote n ti al to
improve the privacy aspe c t s of a digital identity, by giving an individu al the
ability to apply per mi s sio n policies bas e d on their identity and to control
which aspe c t s of that identity are divulge d. An identity may be cont rolla bl e
by the end user, to the exte n t that the user can then decide wha t
infor m a ti o n is given to the party relying on the identity.
Infor m a t i o n Card: Infor m a tio n cards per mit a use r to pre s e n t to a Web
site or othe r service (relying party) one or more claims, in the form of a
softw a r e token, which may be used to uniqu ely identify that user. They can
| 210
Inform a ti o n card s are part of an identity met a- syste m consis tin g of:
2. Us e r s who own and utilize the cards to gain acces s to Web sites
and othe r resou r c e s that suppo r t inform a ti o n cards.
4. Relyi n g parti e s . These are the applica tio n s , servic e s, and so on,
that can use an infor m a tio n card to auth e n ti c a t e a pers o n and to then
aut ho riz e an action such as loggin g onto a Web site, acce s sin g a
docu m e n t , signing cont e n t , and so on
Each inform a tio n card is associa t e d with a set of claims which can be use d
to identify the user. Thes e claims includ e identifiers such as nam e, email
addr e s s , post code. Only the claim types are stor e d in cards issue d by an
identity provide r; The claim values are stor e d by the provide r, cre a ti n g a
more secu r e and privacy- rich syste m . One of the most positive aspec t s of an
infor m a ti o n card is the user- cent ric natu r e of the card. An inform a tio n card
IdP can be set up so that the end user s the m s elv e s can self- issue a card,
bas e d on the req uir e d claims that they the m s elve s input—th e claims being
validat e d if nee d e d . Altern a tively, the claims can be prog r a m m a t i c a lly input
by the IdP via a Web service or similar, allowing the end user to simply
ente r the inform a ti o n card site and downloa d the card.
Claims are the building blocks of the card and are dyna mi c in that they can
be chan g e d eithe r man u ally or progr a m m a t i c ally. A secu rity policy could be
applie d to a dat a reso u r c e that will be enac t e d whe n a specific inform a tio n
card claim is pres e n t e d to it: If this claim cha n g e s , the policy can
subs e q u e n t ly chan g e . For exa m pl e, a policy could be applied to a Google
Apps docu m e n t specifying that acce s s is allowe d for user A whe n they
pres e n t their infor m a ti o n card with claim “secu ri ty clea r a n c e level = 3” and
that post acce s s, this user will be able to view this docu m e n t for 5 days and
be allowe d to edit it. The sam e policy could also reflect a differ e n t secu rity
settin g if the claim cha n g e s , say to a secu rity clear a n c e level = 1; in this
insta n c e the user could be disallow e d acce s s or allowe d acce s s with very
limite d usag e right s.
The dyna mi c natu r e of inform a t io n card s is the stre n g t h of the syste m , but
the weak n e s s of inform a ti o n card s lies in the auth e n ti c a tio n. The curr e n t
infor m a ti o n card identity provisionin g servic es on offer includ e Micros oft
Genev a, Parity, Azigo, Higgins Project, Bandit, and Avoco Secu r e . Each
offers varying levels of card auth e n ti c a ti o n and are chos e n from User n a m e
and pass w o r d , Kerbe r o s token, x509 digital certificat e, and pers o n al card.
• X509 digital certificat e s can be difficult for less tech nic al user s to
install and use
nonp rofit orga niz a tio n that is striving to esta blis h open stan d a r d s for IT, has
form e d a working com mit t e e to ena bl e the use of inform a ti o n cards to
unive rs ally man a g e perso n al digital identitie s.
Data bre a c h e s can be expe n sive. A bre a c h gene r a lly res ult s in a comp a ny
notification of pers o n s acros s the count ry whe n their inform a ti o n has bee n
comp r o mi s e d . For purpo s e s of dat a bre a c h law, dat a in the cloud are
| 213
tre a t e d no differe n tly than any othe r elect r o ni c ally store d infor m a t io n.
Cloud provide r s that have had their syste m s comp r o mi s e d will be requir e d
to notify affect e d pers o n s and will have to coordin a t e with the cloud user s
who provide d the dat a in orde r to do so
Gra m m Leac h Blil e y Act : Fina n c i a l Priva cy Rul e . The Gram m Leach
Bliley Act (GLB) req ui r e s that financial instit ution s imple m e n t proc e d u r e s to
ens u r e the confide n ti ality of perso n al inform a t io n and to prot e c t agains t
unau t h o riz e d acce s s to the inform a ti o n.
• The Rol e of th e FTC: Saf e g u a r d s Rul e and Red Fla g s Rul e . At the
Unite d Stat e s feder al level, the Fede r a l Trade Commis sion (FTC) working
und e r the aus pic e s of the FTC Act has been given autho ri ty to prot e c t
cons u m e r s and their perso n al infor m a ti o n. The Safeg u a r d s Rule man d a t e d
by GLB and enforc e d by the FTC requir e s that all busin e s s e s significa n tly
involved in the provision of financi al service s and prod u c t s have a writt e n
secu ri ty plan to prot e c t custo m e r infor m a ti o n.
releva n t are a of the comp a n y’s oper a tio n, and evalu a tio n of the
effective n e s s of the curr e n t safeg u a r d s for cont rolling thes e risks;
In 2007, as part of the Fair and Accura t e Credit Trans a c ti o n Act of 2003
(FACT), the FTC prom ul g a t e d the Red Flag Rul e s . Thes e rules are
inte n d e d to curb identity theft by having financial instit utio n s identify
pote n ti al “red flags” for activities cond u c t e d thro u g h the orga niza tio n’s
syste m s that could lead to identity theft.
The rules apply to financial institution s or those that hold credit accou n t s
• USA PATRIOT Act . Shortly after Sept e m b e r 11, 2001, the Unite d
Stat e s Cong r e s s pass e d the “Uniting and Stre n g t h e n i n g America by
Providing
The Act also exte n d s the U.S. gover n m e n t’s ability to gain acce s s to
pers o n al financi al inform a t io n and stud e n t inform a tio n store d in elect ro nic
syste m s withou t any suspicion of wrong d oi n g of the pers o n whos e
infor m a ti o n it seeks.
The Direc tive cover s writt e n , oral, elect ro nic, and Inter n e t- bas e d data that
resid e in the EU.
| 215
• China’s constit u tio n refe rs to privacy indirec tly, but the count ry has
very few specific laws.
• On the othe r hand, Hong Kong has a Perso n al Data Ordin a n c e that
covers public and privat e data proce s s o r s and both electr o ni c and
non- elect ro nic recor d s .
• Orga niz a tio n s are held accou n t a b l e for the prot e c tio n of pers o n al
infor m a ti o n it tran sfe r s to third parti e s, whet h e r such partie s are
inside or outsid e of Cana d a .
(b) stops paying the fee the licens e e cha r g e s for the licens e. In the case
of infring e m e n t the licens e agre e m e n t provide s a mech a ni s m for the
licenso r to rep ai r, repla c e , or rem ov e the softw a r e from the licens e e’s
poss e s sio n
It is prim a rily design e d to provide the ter m s unde r which a servic e can
be acces s e d or used by a custo m e r .
The service agre e m e n t may also set forth quality par a m e t e r s arou n d
which the service will be provide d to the user s.
The first, the on- lin e agr e e m e n t , is a click wrap agre e m e n t with which
a cloud user will be pres e n t e d befor e initially acce s si n g the service. A click
wrap is the agre e m e n t the user ente r s into whe n he/sh e checks an “I Agre e”
box, or some t hi n g similar at the initiation of the service relation s hi p.
The agre e m e n t is not subjec t to negoti a tio n and is gene r a lly thou g h t to
be a cont r a c t of adhe sio n
• As large r comp a ni e s move to the cloud (esp eci ally the public cloud), or
more mission- critical applica tio n s or data move to the cloud, the cloud
user will most likely req ui r e the option or a more robu s t and user-
friendly agre e m e n t . The cloud user will push for a negotia t e d
agre e m e n t .
One of the benefits of cloud comp u ti n g from the cloud provide r’s
pers p e c t iv e is the ability of the cloud provide r to move dat a amon g its
availa ble dat a cent e r reso u r c e s as nece s s a r y to maximize the
efficiencie s of it over all syste m. From a tech n ology pers p e c tive, this
ability to move data is a reas o n a b ly good solution to the proble m of
und e r utilized machin e s .
Data Prote c tio n. In fact, in the cloud environ m e n t it is possible that the
sam e data may be store d in multiple location s at the sam e time. For
exam pl e, real time- tran s a c ti o n data may be in one geog r a p h i c location while
the back u p or disas t e r recove ry syste m s may be elsew h e r e . It is also likely
that the agre e m e n t gover ni n g the servic e s says nothin g abou t dat a location.
From a legal pers p e c tive, flexibility of dat a location pote n ti ally challen g e s
the gover ni n g law provision in the cont r a c t. If the law specified in the
cont r a c t (e.g., the cont r a c t says that laws of Thailan d will gover n this
| 218
agre e m e n t ) req uir e s a cert ai n tre a t m e n t of the dat a, but the law of the
jurisdic tion whe r e the dat a resid e s (e.g., data cent e r in Polan d) req ui r e s
anot h e r tre a t m e n t , ther e is an inhe r e n t conflict that must be resolved. This
conflict exists rega r dl e s s of whet h e r the stor a g e is tem p o r a l, and as part of
the proce s si n g of the dat a, or long- term stor a g e that might be a service in
itself (i.e., infras t r u c t u r e as a service), or part of a softw a r e or platfor m as a
servic e offering.
cont r a c t s mad e and actions carrie d out within its bord e r s . In a cloud
environ m e n t , the conflicts of laws issues make the cloud provide r’s
decision s reg a r di n g cross- geog r a p h y virtualiza tio n and multi- tena n cy,
the cloud user’s lack of inform a t io n rega r di n g data location, and the
pote n ti al issue s with geog r a p h i c ally divers e subco n t r a c t o r s highly
releva n t .