Difference between revisions of "Introduction Grid Computing Lab Course Overview"

From PDP/Grid Wiki
Jump to navigationJump to search
m
m
Line 9: Line 9:
 
=== Important Notes and AUP ===
 
=== Important Notes and AUP ===
  
* Before you start, read the [[Introduction_Grid_Computing_Lab_Course_AUP Lab Course Acceptable Use Policy]]. You need to comply with it in order to get graded for the course. When in doubt, ask any of the tutors.
+
* Before you start, read the [[Introduction_Grid_Computing_Lab_Course_AUP|Lab Course Acceptable Use Policy]]. You need to comply with it in order to get graded for the course. When in doubt, ask any of the tutors.
 
* Keep a logbook (either electronically, on paper, or whatever). You'll appreciate it when you try to reproduce your results, or when a disk crashes. You will also need it to write your project result paper.
 
* Keep a logbook (either electronically, on paper, or whatever). You'll appreciate it when you try to reproduce your results, or when a disk crashes. You will also need it to write your project result paper.
 
* grading is integrated with the IGC lecture series and will be explained in the first IGC lecture on Monday
 
* grading is integrated with the IGC lecture series and will be explained in the first IGC lecture on Monday

Revision as of 09:39, 4 September 2005

Structure

The aim of the lab courses will be to install, deploy and operate a mini-grid, with some applications and services. The entire minigrid will be build and run by students partipating in the course (of course with some help from the tutors). At the end of the lab course you'll know what a grid is, be able to build one, and what's needed to make it useable by applications.

There are two tracks in the lab course:

  • build a mini-grid from scratch (for advanced students). The list of projects and literature in section 2 makes - when put together - a basic mini-grid. A team of 2-3 sudents can pick up a topic in the list below (items 1-6) and provide that as a service to the others. All students in this track should look at the authentication part (section 1)
  • paced grid services tutorial with Globus Toolkit 4 "GT4" (for 'new' students). Section 3 has the references to tutorial and background material. GT4 is already pre-installed on some systems for you. Based on these services, you can build a small application using the latest Web Services and Grid protocols.

Important Notes and AUP

  • Before you start, read the Lab Course Acceptable Use Policy. You need to comply with it in order to get graded for the course. When in doubt, ask any of the tutors.
  • Keep a logbook (either electronically, on paper, or whatever). You'll appreciate it when you try to reproduce your results, or when a disk crashes. You will also need it to write your project result paper.
  • grading is integrated with the IGC lecture series and will be explained in the first IGC lecture on Monday

Building the mini-grid

Authentication

Trust in the grid today is established via a Public Key Infrastructure (PKI). Every entity in the system is issues with a "certificate" that links an identifier (the persons name, or a DNS name) to a piece of unique cryptographic data (an RSA keypair, for instance). These certificates usually have a limited lifetime when stored in a file, or are carried on hardware tokens like smart-cards and USB keys.

Commercial providers, like Verisign, Thawte, or Entrust, operate a Certification Authority and sellX.509 public key certificates.

You can also setup an X.509 Certification Authority (CA) yourself. The simplest is to use the OpenSSL commands, that even come with a shellscript to automate the task. More complete functionality can be found in OpenCA. Recent version of the Globus Toolkit also come with a package called "globus-simple-ca".

But there is more to authentication than just issuing certificates to users and hosts. Keys can be compromised or lost, the data in the certificate may become invalid, etc. These issues must be considered, also for the course's CA service.

Literature

Project proposals

  • For everyone: try to setup your own mini-CA, issue a cert to yourself and a friend, and try to setup an authentication connection between the two of you (use openssl s_client and s_server)
  • Build a simple CA service, e.g. based on OpenSSL, that can be used by your fellow students to obtain certificates.
  • Describe the way in which you would identify entities, and what the level of trust in your certificates should be. Describe what the limitations, vulnerabilities, and possible attack vectors.
  • Build a more scalable system, incorporating Registration Authorities, and on-line checking of the status of your certificates (using an independent client program).
  • Integrate on-line checks in a piece of middleware (optional)


Authorization

Users and resources in a grid are grouped in Virtual Organisations. These can be based on directories of users stored in LDAP directories, on attributes issued to the user by the VO, and embedded in the proxy certificate, like in VOMS, or by having a Community Authorization Service (CAS) issue the proxy to the user.

The proxy certificate is the basis for grid authorization today, and enables single sign-on. To access these proxy certs from web portals (and for proxy renewal for long-running jobs), a MyProxy service has been built. This MyProxy service is required for portal operations.

Literature

Project proposals

  • Provide a VO management service for the two grid clusters that will be built lateron (this can best be done with a VO-LDAP server).
  • Old-style systems required the system administrators of a grid site to maintain a file (grid-mapfile) with a list of the authorized users. With VO-LDAP and VOMS, the membership list can be maintained in a central directory for the VO. What else is needed for smooth operation with a VO-LDAP, i.e. how to prevent the sysadmin from having to type something for each new member? (keywords: gridmapdir, LCMAPS, WorkSpace Service/WSS).
  • Setup a CAS service (with GT4) and CAS-enable an example service.

Information Services

A grid consists of many autonomous resources, that come and go. A resource information system to find the resources available for you is therefor vitally important. The system must be stable, scalable to several hunderd sites, hunderds of queries per second, and universally understood.

Information systems have evolved significantly over the years. The Globus Toolkit shipped originally with the "Metacomputing Directory Service" (later renamed to Monitoring and Discovery Service, MDS). The information was presented via an LDAP interface with a proprietary schema. The EU DataGrid [1] and the LHC Computing Grid Project [2] evolved this system later into the Berkeley Database Information Index (BDII) for increased performance and stability.

R-GMA [3] (a relational implementation of the GGF Grid Monitoring Archirecture) uses a structured, SQL based 'virtual database' across all sites in the grid to propagate information in a producer-consumer paradigm.

The Web services based GT4 release contains a completely new version of MDS, that's based on a notification/subscription mechanism that are part of the WS-Resource Framework set of specifications.

And with Condor you get it's own monitoring system Hawkeye.

Note also the existence of UDDI [4], but that is a registry only (not an information or disvcovery service, the W3C Web Services Design Issues page by Tim BL has some details.

Essential for any information system is a common way to express the information in a schema so that others understand the content and meaning of the information contained therein. There are many schema in use. The most popular one today in production grids is the GLUE schema.

Besides there are various management presentation tools like GridICE, MapCenter, GOC Monitor &c.

Literature

"Manager" style monitoring tools:

Projects

  • extract resource information from a host (or a cluster when available) and express it in one of the information systems listed above.
  • once both clusters are operational to some degree, make sure you get the same information fromn both systems, so that the users can decide which one is the best to use (i.e. make a brokering decision based on the information)

Job Management and Clusters

The first resources to populate a grid were compute clusters (a computational grid). These resources were inintially supercomputers, but since we don't have one handy we stick with clusters. A cluster usually consists of a head node (called master, server, scheduler or the like), and a set of worker nodes. Jobs are submitted by users to the head node, and sent to worker node for execution. When there are no free worker nodes left, jobs are queued on the head node.

There are a lot of different batch systems around, both open source and commercial. The references list a few of them. For this course, we will stick with open source or free schedulers.

When you have built the batch system, try running a some jobs through it.

Literature

Batch system software:

Building batch farms:

MPI related links:

Accounting:

Projects

(for two teams of ~3 students each)

  • build a PBS/Torque based cluster with a single head-node a two worker nodes.
    • can you run multi-node jobs?
    • what happens if a node fails (try pulling the network plug!)
    • can you influence scheduling?
    • (optional) implement policy-based scheduling with MAUI
  • build a Condor based cluster with three nodes in total, of which one is also used for other tasks
    • can use use idle cycles on the shared node?
    • what happens in case of failure?
    • can you do job migration?
  • add MPI support to both clusters
  • add a GT2 or GT4 GRAM service to both clusters (use the same on both initially!)
  • build an accounting data collector for each of the clusters. Provide usage data summaries on a per-user and per-VO basis.

Integration

Scheduling and brokering in a grid. We now have to separate administrative domains (the two clusters, one Torque and one Condor) that could benefit from some collaboration. Grid middleware is there is provide collective services across these two systems.

For this integration, it is needed that the clusters run Grid middleware (in this case a Globus Toolkit GRAM service).

Literature

  • Condor-G from UWisc-Madison.
  • NIMROD-G for parameter sweeping over the grid.
  • MPICH-G2, Grid-enabled implementation of the popular MPI.

Other Grid links and middleware:

Projects

  • build a broker that looks at the info system and find the empty cluster (find a grid scheduler like Condor-G)
  • try multi-cluster MPI with MPICH-G2 (and a GT2 GRAM on each cluster)

Portals

This is the optional spare assignment: build a web portal that allows laymen to use the grid compute service (see above), or other services such as those provided in the GT4 tutorial.

Literature

A list of portlet engines is forthcoming.

Projects

  • Build a web portal with MyProxy and a portal engine like GridPort (NPACI)

Grid Services using GT4 - Paced Tutorial

Globus Toolkit version 4 provides a hosting environment for grid services. You can build your own grid service and deploy it inside the GT4 container. The container itself is based on an evolution of the Apache AXIS system, but has been enhanced with grid security mechanisms and support for WS-RF.

Literature

Projects

  • build a grid service to provide XXX.
  • in small groups, aggregate your service with some-one else's to provide additional, higher-level services. For example, a stock quote service that notifies a client watchdog, that will buy more stock based on the rate of change in the share prices, but only if the client still has enough funds in the back. When the amount exceeds a threshold, have the higher-level service notify the person by e-mail.

But of course you're strongly encouraged to think of your own set of services!