Difference between revisions of "Introduction Grid Computing Lab Course Overview"
m |
m |
||
Line 108: | Line 108: | ||
* [http://infnforge.cnaf.infn.it/glueinfomodel/ GLUE Schema specifications], Sergio Andreozzi, et al. 2002-2005. | * [http://infnforge.cnaf.infn.it/glueinfomodel/ GLUE Schema specifications], Sergio Andreozzi, et al. 2002-2005. | ||
* [http://www.globus.org/toolkit/mds/glueschemalink.html Glue Schema and the Globus Toolkit]. | * [http://www.globus.org/toolkit/mds/glueschemalink.html Glue Schema and the Globus Toolkit]. | ||
− | * ... | + | * [http://www.cs.northwestern.edu/~urgis/gis012.pdf A Unified Relational Approach to Grid Information Services], P. Dinda, B. Plale. Grid Forum Information Draft. |
+ | * [http://nws.cs.ucsb.edu/ Network Wheather Service]. | ||
+ | |||
+ | "Manager" style monitoring tools: | ||
+ | * [http://ganglia.sourceforge.net/ Ganglia] cluster monitoring | ||
+ | * [http://monalisa.cacr.caltech.edu/ MonAlisa] by CalTech | ||
+ | * [http://infnforge.cnaf.infn.it/gridice/ GridICE], EU DataTAG and EGEE (INFN), 2002-2005. | ||
=== Projects === | === Projects === | ||
Line 116: | Line 122: | ||
== Job Management and Clusters == | == Job Management and Clusters == | ||
− | The first resources to populate a grid were compute clusters (a computational grid). These resources were inintially supercomputers, but since we don't have one handy we stick with clusters. | + | The first resources to populate a grid were compute clusters (a computational grid). These resources were inintially supercomputers, but since we don't have one handy we stick with clusters. A cluster usually consists of a head node (called master, server, scheduler or the like), and a set of worker nodes. Jobs are submitted by users to the head node, and sent to worker node for execution. When there are no free worker nodes left, jobs are queued on the head node. |
+ | There are a lot of different batch systems around, both open source and commercial. The references list a few of them. For this course, we will stick with open source or free schedulers. | ||
[NEED DESCRIPTION AND REFS] | [NEED DESCRIPTION AND REFS] | ||
+ | === Literature === | ||
+ | |||
+ | Batch system software: | ||
+ | * [http://www.clusterresources.com/products/torque/ Torque], TORQUE (Tera-scale Open-source Resource and QUEue manager), Clusterresources Inc. | ||
+ | * [http://www.cs.wisc.edu/condor/ Condor], Condor High Throughput Computing, UWisc-Madison. | ||
+ | * [http://www.clusterresources.com/products/maui/ MAUI] MAUI Cluster Scheduler, ClusterResources Inc. | ||
+ | * [http://www.platform.com/Products/Platform.LSF.Family/ LSF], Load Share Facility, Platofrm Computing, Inc. (see also the [http://batch.web.cern.ch/ CERN] LSF installation as an example). | ||
+ | * [http://www.openpbs.org/ OpenPBS and PBSPro], by Altair, Inc. | ||
+ | * [http://www.ihep.ac.cn/~chep01/paper/1-001.pdf FBSNG - Batch System for Farm Architecture], J. Fromm, et al. (FNAL) CHEP 2001. | ||
+ | * [http://gridengine.sunsource.net/ Sun Grid Engine] SUN's (now open source) batch system. | ||
+ | |||
+ | Building batch farms: | ||
+ | * [http://www.clusterresources.com/products/torque/docs/torquequickstart.shtml Torque Quickstart Guide] | ||
+ | * [http://grid-deployment.web.cern.ch/grid-deployment/documentation/Maui-Cookbook.pdf MAUI Cookbook], Sophie Lemaitre, Steve Traylen. | ||
+ | * [http://www.gridpp.ac.uk/tb-support/faq/maui.cfg MAUI example configuration] from the RAL LCG Tier-1 site. | ||
+ | * [[PBS_Caching_Utilities]] PBS qstat/pbsnodes caching utilities (needed in case of high load on the headnode pbs server) | ||
+ | * [http://www.gridpp.ac.uk/tb-support/faq/torque.html Packaging] for torque and maui (including some nice patches to Torque) | ||
+ | |||
+ | MPI related links: | ||
+ | * [http://www-unix.mcs.anl.gov/mpi/ MPI] | ||
+ | * [http://www.globus.org/grid_software/computation/mpich-g2.php MPICH-G2], Grid-enabled implementation of the popular MPI. | ||
+ | * [http://goc.grid.sinica.edu.tw/gocwiki/MPI_Support_with_Torque MPI Support with Torque] LCG/EGEE GoC Wiki, Cal Loomis, 2005. | ||
+ | * [http://gridportal.fzk.de/distribution/crossgrid/releases/allfiles/7.3/cg/external/ CrossGrid MPICH-G2 RPMs] | ||
+ | |||
+ | Accounting: | ||
+ | * [http://icsoc.dit.unitn.it/abstracts/A081.pdf An OGSA-Based Accounting System for Allocation Enforcement across HPC Centers], T. Sandholm et al., Proceedings of the 2nd International Conference on Service Oriented Computing. New York, USA, 15-19 November, 2004. Web site: [http://www.sgas.se/ www.sgas.se] | ||
+ | * [http://doi.ieeecomputersociety.org/10.1109/GRID.2003.1261716 DGAS], An Economy-based Accounting Infrastructure for the DataGrid, R.M. Piro et al, Fourth International Workshop on Grid Computing p. 202. | ||
=== Projects === | === Projects === | ||
Line 128: | Line 162: | ||
** can you run multi-node jobs? | ** can you run multi-node jobs? | ||
** what happens if a node fails (try pulling the network plug!) | ** what happens if a node fails (try pulling the network plug!) | ||
− | * build a Condor based cluster with three nodes in total, of which one is also used for other tasks | + | * build a Condor based cluster with three nodes in total, of which one is also used for other tasks |
** can use use idle cycles on the shared node? | ** can use use idle cycles on the shared node? | ||
** what happens in case of failure? | ** what happens in case of failure? | ||
Line 138: | Line 172: | ||
== Integration == | == Integration == | ||
Scheduling and brokering in a grid. | Scheduling and brokering in a grid. | ||
+ | |||
+ | === Literature === | ||
+ | * [http://www.cs.wisc.edu/condor/condorg/ Condor-G] from UWisc-Madison. | ||
+ | * [http://www.csse.monash.edu.au/~davida/nimrod/nimrodg.htm NIMROD-G] for parameter sweeping over the grid. | ||
+ | * [http://www.globus.org/grid_software/computation/mpich-g2.php MPICH-G2], Grid-enabled implementation of the popular MPI. | ||
=== Projects === | === Projects === | ||
Line 145: | Line 184: | ||
== Portals == | == Portals == | ||
Spare assignment: | Spare assignment: | ||
+ | |||
+ | === Literature === | ||
+ | |||
=== Projects === | === Projects === | ||
* Build a web portal with MyProxy and a portal engine like GridPort (NPACI) | * Build a web portal with MyProxy and a portal engine like GridPort (NPACI) |
Revision as of 11:50, 28 August 2005
Structure
The aim of the lab courses will be to install, deploy and operate a mini-grid, with some applications and services. The entire minigrid will be build and run by the students partipating in the course (of course with some help from the tutors). At the end of the lab course you'll know what a grid is, be able to build one, and what is needed to make it useful for applications.
A grid contains a few components that you cannot do without:
- a common trust domain (authentication)
- communities of resources and users (authorization)
- an information service
and of course some services to make the grid useful, like
- a job submission service
- data movement or indexing
- workload management
- database access
- your favourice custom services ...
For each of these, literature and documentation are given below, together with one or two projects (assignments) to be picked up by a team of students (say, 2-3 students per project).
Authentication
Trust in the grid today is established via a Public Key Infrastructure (PKI). Every entity in the system is issues with a "certificate" that links an identifier (the persons name, or a DNS name) to a piece of unique cryptographic data (an RSA keypair, for instance). These certificates usually have a limited lifetime when stored in a file, or are carried on hardware tokens like smart-cards and USB keys.
Commercial providers, like Verisign, Thawte, or Entrust, operate a Certification Authority and sellX.509 public key certificates.
You can also setup an X.509 Certification Authority (CA) yourself. The simplest is to use the OpenSSL commands, that even come with a shellscript to automate the task. More complete functionality can be found in OpenCA. Recent version of the Globus Toolkit also come with a package called "globus-simple-ca".
But there is more to authentication than just issuing certificates to users and hosts. Keys can be compromised or lost, the data in the certificate may become invalid, etc. These issues must be considered, also for the course's CA service.
Literature
- Media:ITU-T-X500-series-UvA.zip ITU-T X.500 Document series, Open Systems Interconnection — The Directory: Overview of Concepts, Models, and Services, Recommendation X.500, ISO/IEC 9594-1
- Global Grid Forum CAOPS-WG
- RFC 2459, Certificate Revocation Lists
- RFC 2560, On-line Certificate Status Protocol
- RFC 3647, Internet X.509 Public Key Infrastructure Certificate Policy and Certification Practices Framework
- EUGridPMA, the European Grid Authentication Policy Management Authority in e-Science
- IGTF, the International Grid Trust Federation
- Minimum CA Requirements for Traditional X.509 Public Key Certification Authorities with secured infrastructure.
- e-IRG White Paper (Dublin), e-Infrastructure Reflection Group 2004, (specifically section 5).
- WS-Trust Web Services Trust Language, (defines extensions that build on WS-Security to provide a framework for requesting and issuing security tokens, and to broker trust relationships).
Project proposals
- Build a simple CA service, e.g. based on OpenSSL, that can be used by your fellow students to obtain certificates.
- Describe the way in which you would identify entities, and what the level of trust in your certificates should be. Describe what the limitations, vulnerabilities, and possible attack vectors.
- Build a more scalable system, incorporating Registration Authorities, and on-line checking of the status of your certificates (using an independent client program).
- Integrate on-line checks in a piece of middleware (optional)
Authorization
Users and resources in a grid are grouped in Virtual Organisations. These can be based on directories of users stored in LDAP directories, on attributes issued to the user by the VO, and embedded in the proxy certificate, like in VOMS, or by having a Community Authorization Service (CAS) issue the proxy to the user.
The proxy certificate is the basis for grid authorization today, and enables single sign-on. To access these proxy certs from web portals (and for proxy renewal for long-running jobs), a MyProxy service has been built. This MyProxy service is required for portal operations.
Literature
- A Security Architecture for Computational Grids, I. Foster, et al. 5th ACM Conference on Computer and Communications Security, 1998.
- Bridging Secure WebCom and European DataGrid Security for Multiple VOs over Multiple Grids, David W. O’Callaghan and Brian A. Coghlan, ISPDC'04, Cork (Ireland), 5-7 July, 2004.
- VO Server Information, J.A. Templon@nikhef.nl and D.Groep, EDG internal note, October 2001.
- Authentication and Authorization Mechanisms for Multi-Domain Grid Environments, L. Cornwall et al., J. Grid Comput. 2(4): 301-311 (2004).
- Managing Dynamic User Communities in a Grid of Autonomous Resources, R. Alfieri et al, CHEP 2003, CoRR cs.DC/0306004: (2003).
- Autonomic Management of Large Clusters and Their Integration into the Grid, T. Roeblitz et al., J. Grid Comput. 2(3): 247-260 (2004).
- A Community Authorization Service for Group Collaboration, L. Pearlman et al. IEEE Workshop on Policies for Distributed Systems and Networks, 2002.
- GT4 Community Authorization Service (CAS) Administrators Guide, The Globus Alliance, 2005.
- The PRIMA System for Privilege Management, Authorization and Enforcement in Grid Environments. M. Lorch et al. Grid2003.
- Grid and Globus Security, Frank Siebenlist, ISOC/GridForum Nederland Masterclass, Amsterdam, July 2005.
- An Online Credential Repository for the Grid: MyProxy, J. Novotny, et al. Proceedings of the 10th IEEE Symposium on High Performance Distributed Computing (HPDC 10), 2001.
- GGF OGSA Security WG
- e-IRG White Paper (Den Haag), e-Infrastructure Reflection Group 2005, (specifically sections 4 and 5).
Project proposals
- Provide a VO management service for the two grid clusters that will be built lateron (this can best be done with a VO-LDAP server).
- Old-style systems required the system administrators of a grid site to maintain a file (grid-mapfile) with a list of the authorized users. With VO-LDAP and VOMS, the membership list can be maintained in a central directory for the VO. What else is needed for smooth operation with a VO-LDAP, i.e. how to prevent the sysadmin from having to type something for each new member? (keywords: gridmapdir, LCMAPS, WorkSpace Service/WSS).
- Setup a CAS service (with GT4) and CAS-enable an example service.
Information Services
A grid consists of many autonomous resources, that come and go. A resource information system to find the resources available for you is therefor vitally important. The system must be stable, scalable to several hunderd sites, hunderds of queries per second, and universally understood.
Information systems have evolved significantly over the years. The Globus Toolkit shipped originally with the "Metacomputing Directory Service" (later renamed to Monitoring and Discovery Service, MDS). The information was presented via an LDAP interface with a proprietary schema. The EU DataGrid [1] and the LHC Computing Grid Project [2] evolved this system later into the Berkeley Database Information Index (BDII) for increased performance and stability.
R-GMA [3] (a relational implementation of the GGF Grid Monitoring Archirecture) uses a structured, SQL based 'virtual database' across all sites in the grid to propagate information in a producer-consumer paradigm.
The Web services based GT4 release contains a completely new version of MDS, that's based on a notification/subscription mechanism that are part of the WS-Resource Framework set of specifications.
And with Condor you get it's own monitoring system Hawkeye.
Note also the existence of UDDI [4], but that is a registry only (not an information or disvcovery service, the W3C Web Services Design Issues page by Tim BL has some details.
Essential for any information system is a common way to express the information in a schema so that others understand the content and meaning of the information contained therein. There are many schema in use. The most popular one today in production grids is the GLUE schema.
Besides there are various management presentation tools like GridICE, MapCenter, GOC Monitor &c.
Literature
- A Performance Study of Monitoring and Information Services for Distributed Systems, 12th IEEE International Symposium on High Performance Distributed Computing (HPDC-12 '03) p. 270
- Relational Grid Monitoring Architecture (R-GMA) Steve Fisher et al.
- The Relational Grid Monitoring Architecture: Mediating Information about the Grid, A.W. Coocke et al.; S.M. Fisher, Journal of Grid Computing, Vol 2, 323-339, December 2004.
- Fault tolerance in the R-GMA Information and Monitoring System, Rob Byrom et al., Proc.EGC'05, Amsterdam, February, 2005.
- The WS-Resource Framework, Globus Alliance, 2004-2005.
- Hawkeye, A Monitoring and Management Tool for Distributed Systems.
- GLUE Schema specifications, Sergio Andreozzi, et al. 2002-2005.
- Glue Schema and the Globus Toolkit.
- A Unified Relational Approach to Grid Information Services, P. Dinda, B. Plale. Grid Forum Information Draft.
- Network Wheather Service.
"Manager" style monitoring tools:
Projects
- extract resource information from a host (or a cluster when available) and express it in one of the information systems listed above.
- once both clusters are operational to some degree, make sure you get the same information fromn both systems, so that the users can decide which one is the best to use (i.e. make a brokering decision based on the information)
Job Management and Clusters
The first resources to populate a grid were compute clusters (a computational grid). These resources were inintially supercomputers, but since we don't have one handy we stick with clusters. A cluster usually consists of a head node (called master, server, scheduler or the like), and a set of worker nodes. Jobs are submitted by users to the head node, and sent to worker node for execution. When there are no free worker nodes left, jobs are queued on the head node.
There are a lot of different batch systems around, both open source and commercial. The references list a few of them. For this course, we will stick with open source or free schedulers.
[NEED DESCRIPTION AND REFS]
Literature
Batch system software:
- Torque, TORQUE (Tera-scale Open-source Resource and QUEue manager), Clusterresources Inc.
- Condor, Condor High Throughput Computing, UWisc-Madison.
- MAUI MAUI Cluster Scheduler, ClusterResources Inc.
- LSF, Load Share Facility, Platofrm Computing, Inc. (see also the CERN LSF installation as an example).
- OpenPBS and PBSPro, by Altair, Inc.
- FBSNG - Batch System for Farm Architecture, J. Fromm, et al. (FNAL) CHEP 2001.
- Sun Grid Engine SUN's (now open source) batch system.
Building batch farms:
- Torque Quickstart Guide
- MAUI Cookbook, Sophie Lemaitre, Steve Traylen.
- MAUI example configuration from the RAL LCG Tier-1 site.
- PBS_Caching_Utilities PBS qstat/pbsnodes caching utilities (needed in case of high load on the headnode pbs server)
- Packaging for torque and maui (including some nice patches to Torque)
MPI related links:
- MPI
- MPICH-G2, Grid-enabled implementation of the popular MPI.
- MPI Support with Torque LCG/EGEE GoC Wiki, Cal Loomis, 2005.
- CrossGrid MPICH-G2 RPMs
Accounting:
- An OGSA-Based Accounting System for Allocation Enforcement across HPC Centers, T. Sandholm et al., Proceedings of the 2nd International Conference on Service Oriented Computing. New York, USA, 15-19 November, 2004. Web site: www.sgas.se
- DGAS, An Economy-based Accounting Infrastructure for the DataGrid, R.M. Piro et al, Fourth International Workshop on Grid Computing p. 202.
Projects
(for two teams of ~3 students each)
- build a PBS/Torque based cluster with a single head-node a two worker nodes.
- can you run multi-node jobs?
- what happens if a node fails (try pulling the network plug!)
- build a Condor based cluster with three nodes in total, of which one is also used for other tasks
- can use use idle cycles on the shared node?
- what happens in case of failure?
- can you do job migration?
- add MPI support to both clusters
- add a GT2 or GT4 GRAM service to both clusters (use the same on both initially!)
- build an accounting data collector for each of the clusters. Provide usage data summaries on a per-user and per-VO basis.
Integration
Scheduling and brokering in a grid.
Literature
- Condor-G from UWisc-Madison.
- NIMROD-G for parameter sweeping over the grid.
- MPICH-G2, Grid-enabled implementation of the popular MPI.
Projects
- build a broker that looks at the info system and find the empty cluster (find a grid scheduler like Condor-G)
- try multi-cluster MPI with MPICH-G2 (and a GT2 GRAM on each cluster)
Portals
Spare assignment:
Literature
Projects
- Build a web portal with MyProxy and a portal engine like GridPort (NPACI)