Скачиваний:
50
Добавлен:
20.06.2019
Размер:
50.48 Mб
Скачать

234

B. Bethwaite et al.

among multiple instances. It is important to note that no fully virtualised infrastructure services, such as EC2, provide QoS guarantees on the performance of hardware sharing.

With this information, we can project the lower bound cost of running the full-quality experiment in a similar situation, again using EC2 as the overflow resource. We scale MCMC iterations and hence job run-time by a factor of 100, and similarly, we scale the deadline to 50 days. We assume that the local resource pool has not changed. In 50 days, East can deliver us 50 × 24 × 120 = 144,000 CPU hours, the equivalent of 945 jobs. We need EC2 to complete the remaining 486 jobs, which requires at least 54,197 EC2-core-hours at US$0.10 an hour, resulting in a potential charge of US$5,420. Clearly, investing that amount of money into capital expenditure (e.g. to extend the capacity of East) would have made little difference to the overall completion time.

13.4.5  Scientific Results

Using Nimrod/G and associated computer clusters and clouds enabled us to explore a greater variety of SPM in a shorter, feasible, time frame without compromising the quality of the results. For instance, we tested the variants of the original model [26], which use different ways of measuring distances between museum exhibits [30]. This led to insights regarding the suitability of the different model variants for certain application scenarios.

In the future, we intend to investigate other ways of incorporating exhibit features into SPM. We also plan to extend our model to fit non-Gaussian data.

13.5  Conclusions

This chapter demonstrates the potential for cloud computing in high-throughput science. We showed that Nimrod/G scales to both freely available resources as well as commercial services. This is significant because it allows users to balance deadlines and budgets in ways that were not previously possible.

We discussed the additions to Nimrod/G required for it to use Amazon’s EC2 as an execution mechanism, and showed that the Nimrod/G architecture is well suited to computational clouds. As a result, Nimrod/G’s cloud actuator allows higher level tools to exploit clouds as a computational resource. Hence, Nimrod/G can be classified as providing a ‘Platform as a Service’ to job producers/schedulers, and becomes both a cloud client and cloud service in its own right.

The case study showed that computational clouds provide ideal platforms for high-throughput science. Using a mix of grid and cloud resources provided timely results within the budget for the research under discussion.

13  Mixing Grids and Clouds: High-Throughput Science Using the Nimrod Tool Family

235

The economies of scale employed by commercial cloud computing providers make them an attractive platform for HTC. However, questions of national interest and policy issues in spending public research funding with commercial providers remain, especially when they are based overseas. Commercial offerings are motivated by profit, and hence it should be possible to provide a non-profit equivalent more cheaply to better utilise government and university funding, while ensuring the prioritisation of researcher requirements. There is clearly scope for the adoption of similar operational techniques in order to provide HTC resources to the research community.

Commercial computing infrastructure requirements also deviate somewhat from typical HTC requirements1. The commercial cloud provider must have sufficient data-centre capacity to meet fluctuating demand, while providing high QoS with regard to reliability and lead time to service. This necessitates reserving capacity at extra expense, passed on to the consumer. On the other hand, HTC workloads are typically not so sensitive. Waiting some time for a processor is of little significance when tens of thousands to millions of processor hours are required. Such considerations may enable higher utilisation and lower capital overhead for dedicated HTC clouds.

Future work will focus on providing accurate cost accounting by implementing a time-slice scheduler and considering data-transfer charges. We also plan to investigate the use of EC2 Spot Instance pricing. This could prove ideal for cost minimisation biased scheduling, given the spot price for a particular machine type is typically less than half of the standard cost.

AcknowledgementsThis work has been supported by the Australian Research Council under the Discovery grant scheme. We thank the Australian Academy of Technological Sciences and Engineering (ATSE) Working Group on Cloud Computing for discussions that were used as input to Section 1. We thank Ian Foster for his helpful discussions about the role of high-throughput science and for his contribution to Section 2.

We acknowledge the work of Benjamin Dobell, Aidan Steele, Ashley Taylor and David Warner, Monash University Faculty of I.T. students who worked on the initial Nimrod EC2 actuator prototype. We also thank Neil Soman for assistance in using the Eucalyptus Public Cloud.

References

1.Abramson D, Giddy J, Kotler L (2000) High performance parametric modeling with Nimrod/G: killer application for the global grid? In the 14th international parallel and distributed processing symposium (IPDPS 2000), pp 520–528

2.Abramson D, Buyya R, Giddy J (Oct 2002) A computational economy for grid computing and its implementation in the Nimrod-G resource broker. Future Gen Comput Syst 18:1061–1074

1 The recently released EC2 Spot Instance pricing (http://aws.amazon.com/ec2/spot-instances/) – a supply-demand-driven auctioning of excess EC2 data-centre capacity – is an early example of a scheme to bridge this gap.

236

B. Bethwaite et al.

3.Litzkow M, Livny M, Mutka M (1988) Condor – a hunter of idle workstations. In the proceedings of the 8th international conference of distributed computing systems. IEEE Press, June 1988, pp 104–111

4.Foster I, Kesselman C (1997) Globus: a metacomputing infrastructure toolkit. Int J Supercomput Appl 11:115–128

5.Erwin DW (2002) UNICORE – a grid computing environment. Concurr Comput Prac Exp 14:1395–1410

6.Laure E, Fisher SM, Frohner A, Grandi C, Kunszt P, Krenek A, Mulmo O, Pacini F, Prelz F, White J, Barroso M, Buncic P, Hemmer F, Di Meglio A, Edlund A (2006) Programming the Grid with gLite. Comput Method Sci Tech 12:33–46

7.Bethwaite B, Abramson D, Buckle A (2008) Grid interoperability: an experiment in bridging grid islands. In the IEEE Fourth International Conference on eScience 2008, pp 590–596

8.Riedel M (2009) Interoperation of world-wide production e-Science infrastructures. Concurr Comput Pract Exp 21:961–990

9.Goscinski W, Abramson D (2008) An infrastructure for the deployment of e-science applications. In: Grandinetti L (ed) High performance computing (HPC) and grids in action. IOS Press, Amsterdam, Netherlands, pp 131–148

10. Schmidberger J, Bethwaite B, Enticott C, Bate M, Androulakis S, Faux N, Reboul C, Phan J, Whisstock J, Goscinski W, Garic S, Abramson D, Buckle A (2009) High-throughput protein structure determination using grid computing. In the IEEE International Symposium on Parallel & Distributed Processing (IPDPS 2009), pp 1–8

11. Sher A, Abramson D, Enticott C, Garic S, Gavaghan D, Noble D, Noble P, Peachey T (2008) Incorporating local Ca2 + dynamics into single cell ventricular models. In: Proceedings of the 8th international conference on computational science, Part I, Springer-Verlag, Krakow, Poland, pp 66–75

12. Baldridge KK, Sudholt W, Greenberg JP, Amoreira C, Potier Y, Altintas I, Birnbaum A, Abramson D, Enticott C, Garic S (2006) Cluster and grid infrastructure for computational chemistry and biochemistry. In: Zomaya AY (ed) Parallel computing for bioinformatics and computational biology, Wiley Interscience, New York, pp 531–550

13. Lynch AH, Abramson D, Görgen K, Beringer J, Uotila P (Oct 2007) Influence of savanna fire on Australian monsoon season precipitation and circulation as simulated using a distributed computing environment, Geophysical Research Letters, 34(20):L20801

14. Abramson D, Lewis A, Peachey T, Fletcher C (2001) An automatic design optimization tool and its application to computational fluid dynamics. In: Proceedings of the 2001 ACM/IEEE conference on supercomputing (CDROM), ACM, Denver, Colorado, pp 25–25

15. Peachey T, Diamond N, Abramson D, Sudholt W, Michailova A, Amirriazi S (Jan 2008) Fractional factorial design for parameter sweep experiments using Nimrod/E. Sci Program 16:217–230

16. Abramson D, Enticott C, Altinas I (2008) Nimrod/K: towards massively parallel dynamic grid workflows. In: Proceedings of the 2008 ACM/IEEE conference on supercomputing, IEEE Press, Austin, Texas, pp 1–11

17. Buyya R, Abramson D, Giddy J, Stockinger H (2002) Economic models for resource management and scheduling in Grid computing. Concurr Comput Prac Exp 14:1507–1542

18. Zhang Y, Mandal A, Koelbel C, Cooper K (2009) Combined fault tolerance and scheduling techniques for workflow applications on computational grids. In: The proceedings of the 2009 9th IEEE/ACM international symposium on cluster computing and the grid – volume 00, IEEE Computer Society, pp 244–251

19. Nurmi D, Brevik J, Wolski R (2008) QBETS: Queue Bounds Estimation from Time Series. Job Scheduling Strategies for Parallel Processing, pp 76–101

20. Buyya R, Giddy J, Abramson D (2000) An evaluation of economy-based resource trading and scheduling on computational power grids for parameter sweep applications. Active middleware services: from the proceedings of the 2nd annual workshop on active middleware services, p 221

13  Mixing Grids and Clouds: High-Throughput Science Using the Nimrod Tool Family

237

21. Nurmi D, Wolski R, Grzegorczyk C, Obertelli G, Soman S, Youseff L, Zagorodnov D (2009) The eucalyptus open-source cloud-computing system. In: The IEEE international symposium on cluster computing and the grid, IEEE Press, Shanghai, China, p 131, 124

22. Sotomayor B, Montero RS, Llorente IM, Foster I (2008) Capacity leasing in cloud systems using the OpenNebula engine. Chicago, IL

23. Raicu I, Zhao Y, Dumitrescu C, Foster I, Wilde M (2007) Falkon: a fast and light-weight tasK executiON framework. In the proceedings of the 2007 ACM/IEEE conference on supercomputing – volume 00, ACM Press, Reno, Nevada, pp 1–12

24. Troger P, Rajic H, Haas A, Domagalski P (2007) Standardization of an API for distributed resource management systems. In the proceedings of the seventh IEEE international symposium on cluster computing and the grid, IEEE Computer Society, pp 619–626

25. Resnick P, Varian HR (1997) Recommender systems. Commun ACM 40:56–58

26. Bohnert F, Schmidt DF, Zukerman I (2009) Spatial processes for recommender systems. In: The 21st international joint conference on artificial intelligence (IJCAI-09), Pasadena, CA, pp 2022–2027

27. Bohnert F, Zukerman I, Berkovsky S, Baldwin T, Sonenberg L (2008) Using interest and transition models to predict visitor locations in museums. AI Commun 21:195–202

28. Banerjee S, Carlin BP, Gelfand AE (2004) Hierarchical modeling and analysis for spatial data. CRC Press, Boca Raton, FL

29. Neal RM (2003) Slice sampling. Annal Stat 31:705–767

30. Bohnert F, Zukerman I, Schmidt DF (2009) Using Gaussian spatial processes to model and predict interests in museum exhibits. In the seventh workshop on intelligent techniques for web personalization and recommender systems (ITWP-09), pp 13–19

Part III

Cloud Breaks

Соседние файлы в папке CLOUD COMPUTING