Академический Документы
Профессиональный Документы
Культура Документы
http://www.researchgate.net/publication/2830193
74
DOWNLOADS 183
VIEWS 134
CITATIONS
5 AUTHORS, INCLUDING:
William Hamish Bell
University of Geneva
137 PUBLICATIONS 2,125 CITATIONS
SEE PROFILE
A. Paul Millar
Deutsches Elektronen-Synchrotron
38 PUBLICATIONS 1,021 CITATIONS
SEE PROFILE
Kurt Stockinger
Zurich University of Applied Sciences
115 PUBLICATIONS 3,026 CITATIONS
SEE PROFILE
Available from: Kurt Stockinger
Retrieved on: 20 September 2015
Sim ulation of Dynami Grid Repli ation Strategies in OptorSim William H. Bell1 ,
Da vid G. Cameron1 , Luigi Cap ozza2 , A. P aul Millar1 , Kurt Sto kinger3 ,
Floriano Zini2 1 Univ ersit y of Glasgo w, Glasgo w, G12 8QQ, S otland 2 ITC-irst,
Via Sommariv e 18, 38050 P o v o (T ren to), Italy 3 CERN, Europ ean
Organization for Nu lear Resear h, 1211 Genev a, Switzerland Abstra t.
Computational Grids normally deal with large omputationally in tensiv e problems
on small data sets. In on trast, Data Grids mostly deal with large omputational
problems that in turn require ev aluating and mining large amoun ts of data. Repli
ation is regarded as one of the ma jor optimisation te hniques for pro viding fast
data a ess. Within this pap er, sev eral repli ation algorithms are studied. This is a
hiev ed using the Grid sim ulator: Opto rSim. Opto rSim pro vides a mo dular
framew ork within whi h optimisation strategies an b e studied under di
eren t Grid ongurations. The goal is to explore the stabilit y and transien t b eha
viour of sele ted optimisation te hniques. 1 In tro du tion Within the Grid omm
unit y m u h w ork has b een done on pro viding the basi infrastru ture for a t ypi
al Grid en vironmen t. Globus [4, Condor [1 and re en tly the EU DataGrid [3
ha v e on tributed substan tially to ore Grid middlew are servi es soft w are that
are a v ailable as the basis for further appli ation dev elopmen t. Ho w ev er, little
e
ort has b een made so far to optimise the use of Grid resour es. T o use a Data
Grid, users t ypi ally submit job s. In order for a job to b e exe uted, three t yp es
of resour es are required: omputing fa ilities, data a ess and storage, and net w
ork onne tivit y . The Grid m ust mak e s heduling de isions for ea h job based on
the urren t state of these resour es (w orkload and features of Computing Elemen
ts, lo ation of data, net w ork load). Complete optimisation is a hiev ed when the
om bined resour e impa t of all jobs is minimised, allo wing jobs to run as fast as
p ossible. File repli ation (i.e. spread of m ultiple opies of les a ross the Grid) is an
e
e tiv e te hnique for redu ing data a ess o v erhead. Main taining an optimal
distribution of repli as implies that the Grid optimisation servi e [7 m ust b e
able to mo dify the geographi lo ation of data les. This is a hiev ed b y triggering
b oth repli ation and deletion of data les. By re e ting the dynami load on the
Grid, su h repli a managemen t will a
e t the migration of parti ular les to w ard sites that sho w in reased frequen y of
le-a ess requests. In order to study the omplex nature of a t ypi al Grid en
vironmen t and ev aluate v arious repli a optimisation algorithms, a Grid sim
ulator ( alled Opto rSim) w as dev elop ed. In this pap er the design on epts of
Opto rSim are dis ussed and preliminary results based on sele ted repli ation
algorithms are rep orted. The pap er is stru tured as follo ws. Se tion 2 des rib es
the design of the sim ulator Opto rSim. V arious repli ation algorithms are dis
ussed in Se tion 3. After setting the sim ulation onguration in Se tion 4, Se tion 5
is dedi ated to a des ription of sim ulation results. Se tion 6 highligh ts related w
ork. Finally , Se tion 7 on ludes the pap er and rep orts on future w ork. 2 Sim
ulation Design Opto rSim [2 is a sim ulation pa k age written in Ja v a. It w as
dev elop ed to study the e
e tiv eness of repli a optimisation algorithms within a Data Grid en vironmen t.
2.1 Ar hite ture One of the main design onsiderations for Opto rSim is to mo del
the in tera tions of the individual Grid omp onen ts of a running Data Grid as
realisti ally as p ossible. Therefore, the sim ulation is based on the ar hite ture of
the EU DataGrid pro je t [14 as illustrated in Figure 1.
Storage
Element
Replica
Optimiser
Computing
Element
Replica Manager
Storage
Element
Replica
Optimiser
Computing
Element
Replica Manager
Storage
Element
Replica
Optimiser
Computing
Element
Replica Manager
User Interface
Resource Broker
Elemen ts. Sites without Storage or Computing Elemen ts a t as net w ork no des
or routers. The de ision ab out data mo v emen t asso iated with jobs b et w een
sites is p erformed b y a omp onen t alled the R epli a Manager. Within the Repli a
Manager the de ision to reate or delete repli as is on trolled b y a R epli a
Optimiser alled Opto r. A t the heart of Opto r is a repli a optimisation algorithm,
the prop erties of whi h are dis ussed in Se tion 3. 2.2 In ternals In the sim ulation
ea h Computing Elemen t is represen ted b y a thread. Job submission to the
Computing Elemen ts is managed b y another thread: the Resour e Brok er. The
Resour e Brok er ensures ev ery Computing Elemen t on tin uously runs jobs b y
frequen tly attempting to distribute jobs to all the Computing Elemen ts. When
the Resour e Brok er nds an idle Computing Elemen t, it sele ts a job to run on it a
ording to the p oli y of the Computing Elemen t, i.e. whi h t yp e of jobs it will run
and ho w often it will run ea h job. A t an y time, a Computing Elemen t will b e
running at most one job. As so on as the job nishes, another is assigned b y the
Resour e Brok er. So, although there is no expli it job s heduling algorithm, all
Computing Elemen ts pro ess jobs for the duration of the sim ulation but are nev
er o v erloaded. Curren tly , optimisation only o urs after a job has b een s
heduled to a Computing Elemen t. The more omplex s enario of optimising b oth
job s heduling and data a ess will b e part of future w ork. Ea h job has a set of les
it ma y request. Tw o t yp es of referen e ma y b e used for a le: a logi al le name
(LFN) and a ph ysi al le name (PFN). An LFN is an abstra t referen e to a le that is
indep enden t of b oth where the le is stored and ho w man y repli as exist. A PFN
refers to a sp e i repli a of some LFN, lo ated at a denite site. Ea h LFN will ha v e
one PFN for ea h repli a in the Grid. A job will t ypi ally request a set of LFNs for
data a ess. The order in whi h those les are requested is determined b y the a ess
pattern. The follo wing a ess patterns w ere onsidered: se quential (the set of
LFNs is ordered, forming a list of su essiv e requests), r andom (les are sele ted
randomly from set with a at probabilit y distribution), unitary r andom walk (set is
ordered and su essiv e le requests are exa tly one elemen t a w a y from previous
le request, dire tion is random) and Gaussian r andom walk (as with unitary
random w alk, but les are sele ted from a Gaussian distribution en tred on the
previous le request). When a le is required b y a job, the le's LFN is used to lo ate
the b est repli a via the Repli a Optimiser fun tion getBestFile(LFN,
destinationStora ge - Element), where destinationStora ge El eme nt is the
Storage Elemen t to whi h the repli a ma y b e opied. It is assumed the
Computing Elemen t on whi h the job is running and requested Storage Elemen t
are lo ated at the same site. getBestFile() he ks the R epli a Catalo gue for opies
of the le. The Repli a Catalogue is a Grid middlew are servi e urren tly implemen
ted within the simulation as a table of LFNs and all orresp onding PFNs. By
examining the a v ailable bandwidth b et w een destinationStora ge Ele me nt
and all sites on whi h a repli a of the le is stored, getBestFile() an ho ose the PFN
that will b e a essed fastest and hen e de rease the job running time. The sim
ulated v ersion of getBestFile() partially fulls the fun tionalit y as des rib ed in [7.
It is a blo king all that ma y ause repli ation to a Storage Elemen t lo ated in the
site where the job is running. After an y repli ation has ompleted, the PFN of the b
est a v ailable repli a is returned to the job. If repli ation has not o urred, the b est
repli a is lo ated on a remote site and is a essed b y the job using remote I/O.
Both the repli ation time (if repli ation o urs) and the le a ess time (if from a
remote site) are dep enden t on the net w ork hara teristi s o v er the duration of
the onne tion. A t an y time, the bandwidth a v ailable to a transfer is limited b y
the lo w est bandwidth along the transfer path. F or transfers utilising a ommon
b y the Repli a Optimiser (sa y , for le f ), the predi tion fun tion E is al ulated for f
and ev ery le in the storage. If there is no le in the Storage Elemen t that has a v
alue less than the v alue of f then no repli ation o urs. Otherwise, the least v
aluable le is sele ted for deletion and a new repli a of f is reated on the Storage
Elemen t. If m ultiple les on the Storage Elemen t share the minim um v alue, the
le ha ving the earliest last a ess time is deleted. The ev aluation fun tion E (f ; r ;
n) is dened b y the equation E (f ; r ; n) = i X n =1 pi (f ); (1) with the follo wing
argumen t. Assuming that requests for les on taining similar data are lustered in
spatial and time lo alit y , the request history an b e des rib ed as a random w alk
in the spa e of in teger le iden tiers1 . In the random w alk, the iden tier of the
next requested le is obtained from the urren t iden tier b y the addition of a step,
the v alue of whi h is giv en b y some probabilit y distribution. Assuming a
binomial distribution of the steps, the probabilit y of re eiving a request for le f at
step i of the random w alk is giv en b y the equation pi (f ) = 1 22iS id(f ) 2 iS s +
iS ; jid(f ) s j iS (2) where s is the mean v alue of the binomial distribution , S is
the maxim um v alue for the step, and id(f ) is a unique le iden tier (for instan e,
the LFN). Then, the most pr ob able numb er of times le f wil l b e r e queste d
during the next n r e quests is giv en b y (1). A time in terv al t des rib es ho w
far ba k the history go es and th us determines the n um b er r of previous
requests whi h are onsidered in the predi tion fun tion. W e assume that the mean
arriv al rate of requests is onstan t. On e t has b een de ided, n is obtained b y
n = r t0 t (3) where t0 is the future in terv al for whi h w e in tend to do
the predi tion. The v alue for S in (2) dep ends on the v alue of r . The mean v
alue s is obtained from the re en t v alues of the step in the random w alk. In
parti ular, s is al ulated as the w eigh ted a v erage of the last r steps, where w
eigh ts de rease o v er past time. 4 Sim ulation Conguration 4.1 Grid Conguration
The study of optimisation algorithms w as arried out using a mo del of EU
DataGrid T estBed 1 sites and their asso iated net w ork geometry as illustrated 1
W e assume a mapping b et w een le names and iden tiers that preserv e le on
ten t similarit y . in Figure 2. Within this mo del, ea h site w as allo ated storage
resour es prop ortional to their a tual hardw are allo ations. Ea h T estBed site, ex
luding CERN, w as assigned a Computing and Storage Elemen t. CERN w as allo
ated a Storage Elemen t to hold all of the master les but w as not assigned a
Computing Elemen t. Routers, as previously stated, w ere des rib ed b y reating a
site without Computing or Storage Elemen ts. The size of the Storage Elemen ts
for ea h T estBed site are giv en in T able 1.
10G
10G
45M
45M
1G
155M
2.5G
Imperial College
2.5G
2.5G
Lyon
155M
10G 2.5G
622M
10G
NIKHEF
RAL
NorduGrid
10G
155M
2.5G
10G
Padova
Bologna
155M
10M
Catania
Torino
Milano 10M
CERN
Testbed site
Fig. 2. The EU DataGrid T estBed 1 sites and the appro ximate net w ork
geometry . The n um b ers indi ate the bandwidth b et w een t w o sites. Site
Name Bologna Catania CERN Imp erial College Ly on Storage Elemen t (GBytes)
30 30 10000 80 50 Site Name Milano NIKHEF NorduGrid P ado v a RAL T orino
Storage Elemen t (GBytes) 50 70 63 50 50 50 T able 1. A list of resour es allo
ated to the T estBed 1 sites, from whi h the results in this pap er w ere
generated. 4.2 Job Conguration Initially , all les w ere pla ed on the CERN Storage
Elemen t. Jobs w ere based on the CDF use- ase as des rib ed in [12. There w ere
six job t yp es, with no o v erlap b et w een the set of les ea h job a essed. The
total size of the le a essed b y an y job t yp e w ere estimated in [12 and are
summarised in T able 2. Ea h set of les w as assumed to b e omp osed of
10GByte les. There will b e some distribution of jobs ea h site p erforms. In the
sim ulation, w e mo delled this distribution su h that ea h site ran an equal n um b
er of jobs of ea h t yp e ex ept for a preferred job t yp e, whi h ran t wi e as often.
This job t yp e w as hosen for ea h site based on storage onsiderations; for the
repli ation algorithms to b e e
e tiv e, the lo al storage on ea h site had to b e able to hold all the les for the
preferred job t yp e. Data Sample T otal Size (GBytes) Cen tral J = 1200 High pt
leptons 200 In lusiv e ele trons 5000 In lusiv e m uons 1400 High Et photons 5800
Z 0 ! bb 600 T able 2. Estimated sizes of CDF se ondary data sets (from [12). 5
Results Fig. 3. A histogram of job duration (left) and the progression of job
duration o v er ourse of the sim ulation (righ t). The left histogram in Figure 3 sho
ws a t ypi al spread of job duration for a single job t yp e at a sele ted Computing
Elemen t o v er the ourse of a sim ulation run. The large spik e near zero is due to
the job requesting les that are a v ailable on the lo al site, hen e no timeonsuming le transfers need to tak e pla e. The longer durations are due to the job
requesting some les not presen t at the lo al site. The spread is due to the net w
ork load, whi h an v ary o v er time, a
e ting the le transfer times. The v ariation of job duration o v er the sim ulation is
sho wn in the righ t histogram in Figure 3 for the same job t yp e and Computing
Elemen t as ab o v e. There is learly a large v ariation in the job duration due to
the fa tors already men tioned, but the general trend is for jobs to b e exe uted
more qui kly o v er time, indi ating the mo v emen t to w ard a more optimal repli
a onguration. Fig. 4. In tegrated running times for 10000 jobs using ea h a ess
pattern and repli a optimisation algorithm. F urther tests w ere ondu ted sim
ulating 10000 jobs using ea h of the four algorithms: 1. No r epli ation 2. Un
onditional r epli ation, oldest le delete d 3. Un onditional r epli ation, le ast a esse
d le delete d 4. E onomi Mo del F or ea h repli ation algorithm, ea h of the follo
wing four le a ess patterns (as dened in Se tion 2.2) w as tested. 1. Se quential 2.
R andom 3. Unitary r andom walk 4. Gaussian r andom walk Figure 4 sho ws the
total time to omplete 10000 jobs for ea h of the four a ess patterns using the four
optimisation algorithms. With no optimisation, the jobs tak e m u h longer than ev
Router
en the simplest optimisation algorithm as all the les for ev ery job ha v e to b e
transferred from CERN ev ery time a job is run. The three algorithms where repli
ation is ondu ted all sho w a mark ed redu tion in the time to exe ute 10000 jobs.
This is not surprising as with no repli ation, all le requests from all jobs m ust ome
from CERN. The three optimisation algorithms that implemen t repli ation sho w
similar p erforman e for Random, Unitary random w alk and Gaussian random w
alk. F or sequen tial a ess patterns, the running time is at least 10% faster using
the E onomi Mo del optimiser than the other optimisers. These results w ere exp
e ted as the E onomi Mo del assumes a sequen tial a ess pattern. Ho w ev er, this
an b e adjusted to mat h the observ ed distribution, if needed. 6 Related W ork Re
en tly there has b een great in terest in mo delling Data Grid en vironmen ts. A
sim ulator for mo delling omplex data a ess patterns of on urren t users in a
distributed system is found in [13. These studies w ere mainly ondu ted within
the setting of s ien ti exp erimen ts su h as the LHC, whi h nally resulted in the
reation of the EU DataGrid pro je t [3. Mi roGrid [18 is a sim ulation to ol for
designing and ev aluating Grid middlew are, appli ations and net w ork servi es
for the omputational Grid. Curren tly , this sim ulator do es not tak e data
managemen t issues in to onsideration. F urther Grid sim ulators are presen ted
in [11, 6 In [15 an approa h is prop osed for automati ally reating repli as in a t
ypi al de en tralised P eer-to-P eer net w ork. The goal is to reate a ertain n um b
er of repli as on a giv en site in order to guaran tee some minimal a v ailabilit y
requiremen ts. In Nimro d-G [8, 5 an e onomi mo del for job s heduling is in tro
du ed in where \Grid redits" are assigned to users that are prop ortional to their
lev el of priorit y . In this mo del, optimisation is a hiev ed at the s heduling stage
of a job. Ho w ev er, our approa h di
ers b y in luding b oth optimal repli a sele tion and automated repli a reation in
addition to s heduling-stage optimisation. V arious repli ation and a hing
strategies within a sim ulated Grid en vironmen t are dis ussed in [16 and their
om bination with s heduling algorithms is studied in [17. The repli ation
algorithms prop osed are based on the assumption that p opular les in one site
are also p opular in other sites. Repli ation from one site to another is triggered
when the p opularit y of a le o v er omes a threshold and the destination site is
hosen either randomly or b y sele ting the least loaded site. W e tak e a
omplemen tary approa h. Our repli ation algorithms are used b y Grid sites when
they need data lo ally and are based on the assumption that in omputational
Grids there are areas (so alled \data hot-sp ots") where parti ular sets of data are
highly requested. Our algorithms ha v e b een designed to mo v e data les to w
ard \data hot-sp ots". 7 Con lusions and F uture W ork In this pap er w e des rib
ed the design of the Grid sim ulator Opto rSim. In parti ular, Opto rSim allo ws the
analysis of v arious repli ation algorithms. The goal is to ev aluate the impa t of
the hoi e of an algorithm on the throughput of t ypi al Grid jobs. W e ha v e hosen
t w o traditional a he managemen t algorithms (oldest le deletion and least a
essed le deletion) and ompared them to a no v el algorithm based on an e onomi
mo del. W e based our analysis on sev eral Grid s enarios with v arious w ork
loads. Results obtained from Opto rSim suggest that the e onomi mo del p
erforms at least as w ell as traditional metho ds. In addition, there are sp e i
realisti ases where the e onomi mo del sho ws mark ed p erforman e impro v
emen ts. Our future w ork will extending the sim ulator b y in luding the au tion
proto ol prop osed in [10. This is motiv ated b y the additional fun tionalit y of
automati repli ation to third part y sites, allo wing le migration to a urately mat h
demand. A kno wledgemen ts The authors thank Erwin Laure, Heinz Sto kinger
and Ek o w Oto o for v aluable dis ussions during the preparation of this pap er.
William Bell and Da vid Cameron thank PP AR C for funding as part of the
GridPP(EDG) pro je t and as an e-S ien e studen t resp e tiv ely; P aul Millar thanks
SHEF C for funding under the S otGRID pro je t. Referen es 1. The ondor pro je t.
http://www. s.wis .edu/ ondor/ . 2. OptorSim - A Repli a Optimiser Sim ulation.
http://grid- data- management . web. ern. h/grid- data- managem ent/o ptim isat i
%on/ opto r/. 3. The DataGrid Pro je t. http://www.eu- datagrid.org. 4. The Globus
Pro je t. http://www.globus.org. 5. D. Abramson, R. Buuy a, and J. Giddy . A
Computational E onom y for Grid Computing and its Implemen tation in the Nimro
d-G Resour e Brok er. In F utur e Gener ation Computer Systems, to app ear. 6. K.
Aida, A. T ak efusa, H. Nak ak a, S. Matsuok a, S. Sekigu hi, and U. Nagashima. P
erforman e Ev aluation Mo del for S heduling in a Global Computing System.
International Journal of High Performan e Appli ations, 14(3), 2000. 7. W. H. Bell,
D. G. Cameron, L. Cap ozza, P . Millar, K. Sto kinger, and F. Zini. Design of a Query
Optimisation Servi e. T e hni al rep ort, CERN, 2002. WP2 - Data Managemen t, EU
DataGrid Pro je t. http://edms. ern. h/do ument/3 37977 . 8. R. Buyy a, H. Sto
kinger, J. Giddy , and D. Abramson. E onomi Mo dels for Managemen t of Resour
es in P eer-to-P eer and Grid Computing. In Commer ial Appli ations for HighPerforman e Computing, SPIE's International Symp osium on the Conver gen e of
Information T e hnolo gies and Communi ations (ITCom 2001), Den v er, Colorado,
USA, August 2001. 9. L. Cap ozza, K. Sto kinger, and F. Zini. Preliminary Ev
aluation of Rev en ue Predi tion F un tions for E onomi ally-E
e tiv e File Repli ation, June 2002. 10. M. Carman, F. Zini, L. Serani, and K. Sto
kinger. T o w ards an E onom y-Based Optimisation of File A ess and Repli ation on
a Data Grid. In International Workshop on A gent b ase d Cluster and Grid
Computing at International Symp osium on Cluster Computing and the Grid
(CCGrid 2002), Berlin, German y , Ma y 2002. IEEE Computer So iet y Press. Also
app ears as IRST T e hni al Rep ort 0112-04, Istituto T ren tino di Cultura, De em b
er 2001. 11. H. Casano v a, G. Ob ertelli an F. Berman, and R. W olski. The AppLeS
P arameter Sw eep T emplate: User-Lev el Middlew are for the Grid. In Pr o . of
Sup er Computing 2002, Dallas, T exas, USA, No v em b er 2002. 12. B. T. Hu
man et al. The CDF/D0 UK GridPP Pro je t. CDF In ternal Note. 5858. 13. I. C.
Legrand. Multi-Threaded, Dis rete Ev en t Sim ulation of Distributed Computing
Systems. In Pr o . of Computing in High Ener gy Physi s (CHEP 2000), P ado v a,
Italy , F ebruary 2000. 14. EU DataGrid Pro je t. The DataGrid Ar hite ture, 2001.
15. K. Ranganathan, A. Iamnit hi, and I. F oster. Impro ving Data Av ailabilit y
through Dynami Mo del-Driv en Repli ation in Large P eer-to-P eer Comm unities.
In Glob al and Pe er-to-Pe er Computing on L ar ge S ale Distribute d Systems
Workshop, Berlin, German y , Ma y 2002. 16. K. Ranganathana and I. F oster. Iden
tifying Dynami Repli ation Strategies for a High P erforman e Data Grid. In Pr o .
of the International Grid Computing Workshop, Den v er, Colorado, USA, No v em
b er 2001. 17. K. Ranganathana and I. F oster. De oupling Computation and Data
S heduling in Distributed Data-In tensiv e Appli ations. In International Symp
osium of High Performan e Distribute d Computing, Edin burgh, S otland, July
2002. T o app ear. 18. H. J. Song, X. Liu, D. Jak obsen, R. Bhagw an, X. Zhang, K. T
aura, and A. Chien. The Mi roGrid: a S ien ti T o ol for Mo deling Computational
Grids. S ienti Pr o gr amming, 8(3):127{141, 2000.