PPDG Quarterly Report, Caltech March->June 2002
1) Eric Aslakson
Eric's activities during this period were divided between developing
software for physics analysis and designing
software for production activites
related to scheduling. In the analysis software
area, Eric completed work on a
Root-Tree to SQL database insert program. He generalized and extended
code supplied by Krzysztof Nienartowicz, of the IT
group at CERN, to use
Microsoft SqlServer as well as Oracle databases. He also extended the
modules to map arbitrary Root group names into SQL table names, enabling
conversion of (almost) arbitrary Root trees.
Due to design requirements that the code run unattended, Eric also
converted batched database inserts to ODBC function
call inserts. This allows
more robust error recovery for unattended inserts.
This ODBC layer also
facilitates pluggable SQL database usage. He then started proving ODBC
drivers on Linux for Oracle and SQLServer. He also began work on a
standard template library based objectification of
SQL result sets. This layer
will facilitate front end caching and easy tranport
of SQL results sets
optimized for high latency network connections.
This software will plug into
Clarens for use in Grid Analysis Environment
prototype.
In the production area, Eric continued to participate developments in
the PPDG monitoring area by participating in the
very few phone conferences
occuring. He continued design of the Execution
Priority Manager, a module working
in concert with the DAG processors and grid nodes
to attempt a guarantee of
throughput promises for production activities. This work will continue
in August when he will be working with the relevant
groups at Fermilab to
implement this module.
2) Edwin Soedarmadji
In this reporting period, Edwin concentrated on the
tasks : (1) reverse-engineering a portion of JetMet
analysis code produced by Pal Hidas while at FNAL
(2) writing programs that convert JetMet PAW ntuple
into raw text (3) building a SQLServer database and
(4) import the converted ntuples into the database.
These tasks were recently completed and now serve
as the foundation for the next set of tasks. The
SQLServer database will be used in conjunction with
an Oracle database at CERN (under construction) to
serve out JetMet analysis data in a more efficient
and granular way. Currently Edwin is working on a
set of web services that would allow queries to be
submitted and executed against the Jetmet database
by a client software.
3) Saima Iqbal
Saima's main task is the evaluation of Oracle9i for the
CMS Event store. Oracle9i implements an Object Relational
Model. The object type maps closely to the class mechanisms
found in C++ and Java, as a result the object based code
that is already in use by CMS would not need be changed
or rewritten. Only an SQL wrapper is required to store and
access data from Oracle database.
Saima's evaluation started by investigating the Object Relational
features of Oracle. She measured the overheads on the
objects after storing them in the Oracle9i database. For
these overhead measurements simple objects, embedded objects
and objects with REF were created in the form of Object-Tables.
These object tables were loaded with simple data from vectors
and hits. In the Object-Table each C++ object corresponds to
one row and each attribute maps to the column in the row of the table.
LOADING AND ACCESS OF TAGS DATA FROM ORACLE 9i DATABASE:
As a follow up activity Saima looked at the proposed architecture
for a CMS event store using Oracle 9i and its related tools. A
small-scale data warehouse prototype for CMS TAG data was
developed by her. To evaluate the performance of Oracle for
the persistency of Tags data the Java classes TagLoader.class
and TagAccess.class developed at Caltech for SQLServer were
converted for Oracle. Preliminary results show a data warehouse
ingestion speed of 25 MB/s. The tuning of this process will
improve over time.
ORACLE 9i WEB APPLICATION:
In the next step (loading and accessing the Tags Data) an Oracle
base Web Application was developed by Saima. After the comparative
study of web applications developed by using Microsoft FrontPage,
Perl, CGI (Common Gateway Interface) and Java, the web application
was designed in Java/JDBC (using Java Server Pages). It
is platform independent and its byte codes can easily be run
on any machine providing a JVM (Java Virtual Machine). It has
an open structure to avoid restricting users to specific hardware
and software configurations. It does not require the Oracle Client
installation. The demo of this web application was presented by Saima
during CMS Week June-2002.
ARCHITECTECTURE TO USE ORACLE 9i IN GRID ENVIRONMENT:
A proposed architecture makes use of Oracle 9i data warehousing
in a Grid environment, enabling hybrid data access with Oracle9i
and Microsoft SQL server. For this purpose Saima began work on
loading the Jetmet Ntuples version 2.06 in the Oracle Server at CERN.
4) Suresh Man Singh
Suresh's main activities during the period were focused on
upgrading hardware and software for Caltech's AMD cluster in the USCMS
grid
testbed. He upgraded the RAID software of the Network Attached Storage
(NAS)
unit of the cluster. The Linux kernels of all the cluster nodes were
upgraded to 2.4.18 allowing improved performance. Various security
related
fixes or upgrades were made to vulnerable packages. These included
the Apache web server, openssh 3.4 (secured shell), openssl, python etc.
The
time protocol NTP was upgraded on all the nodes and the head node was
synchronized to CDT - as required by the distributed Monte-carlo
Production
(MOP).
Suresh installed both VDT 1.1 server and client on the cluster's head
node (citcms.cacr.caltech.edu) which includes Globus 2.0. As part
of the host, service and user certificate migration process from
Globus signing to DOE Science Grid certificates, he obtained the
DOESG signed host certificate for the cluster's head node and installed
it.
This head node was registered with University of Florida's GIIS MDS
server
and can be queried by Globus grid-info-search command. Suresh is in the
process of upgrading the cluster's local batch scheduler package to the
most recent release of Condor to 6.4. This will be used for distributed
MOP production by providing a hook from the Globus jobmanager. At the
moment
only the "FORK" jobmanager of Globus is active.
5) Conrad Steenberg
Summary: Work on the Clarens web services infrastructure continues, with
improvements in robustness and functionality. Client functionality was
expanded, with a ROOT remote file access method and an ftp-like command
like file download client recently being the main highlights. The
Clarens
project has a new development home where a file repository (CVS), web
page,
and mailing list functionality is available in order to enable wider
collaboration. Work started on developing and deploying Virtual
Organization (VO) management tools to ease Grid site security
management.
1.Presentations
- "Clarens remote analysis enabling environment"CMS Tutorial week, May
1 - 4, UC San Diego
- "LDAP CA Authorization management" PPDG/Griphyn All hands meeting,
April 23 - 25, Argonne NL,
2.Project management
The Clarens project home page was moved to
http://clarens.sourceforge.net
which provides managed web hosting to open source projects. A project
mangement
interface can be found at
http://sourceforge.net/projects/clarens/,
with
access to bug tracking, a versioned file repository, and mailing lists.
3.Tool development and software changes
- Clarens services added
Added file download and browsing service added. Since most forms of
data are still stored in files, this service is a logical basic
compoenent of Clarens. The service streams data from a file to the
client
in raw form, or provides file listings for the directory structure of
the
file system. User files may be published from a special directory
created
in the user's login directory, and accessed using the well known tilde
notation (~user/).
- Service description publication
Automatically generated descriptions and API documentation is provided
using the system.Listmethods, system.methodHelp, and
system.methodSignature
API calls
- Clarens client side
Two clients were implemented for the file access service, a command-line
ftp-like interface, as well as enabling the ROOT analysis package to
access
remote files tranparently
- VO Management
In order to manage authorization in the Grid computing model, the
concept of
Virtual Organizations was developed (by others). This involves placing
users
in different groups or categories and managing access to certain
resources
on the Grid based on these groups, instead of on an indivudual basis,
thereby
reducing the load on site administrators. To aid this process a tool to
manage
VOs was developed that provides a graphical user interface to the user,
group
and certificate information, as opposed to the cumbersome collection of
scripts
that are currently used. This tool has a web page at
http://heppc22.hep.caltech.edu/groupman.
6) Vladimir Litvin
Vladimir organised the production of 50k events for
forward jet-tagging mechanism studies, and
25K events for calibration studies. He made a presentation at ACAT2002.
He prepared analysis software for July CMS milestone in the Egamma
group.
He analysed five Level1 collections for the DAQ TDR, and five full
datasets for the pixel isolation studies. An importer for data files
from CERN has been installed, and 0.8TBytes of data moved from CERN
and analysed.
7) Julian Bunn
Planning for demonstrations at iGrid2002 and SC2002 is advancing,
with discussions centering around which components of the Grid
Enabled Analysis Environment will be ready, and which need to
be "glued" together to make a sensible application. For iGrid2002
we are collaborating with Bill Allcock and others in the Globus
team at Argonne. The intention is to show a ROOT front end client
on the show floor, which talks to Clarens servers at Caltech, Argonne,
CERN and possibly UCLA and Rio. GridFTP will be used to move large
data files generated on the server by Clarens to the client machine.
We want to show a truly heterogeneous GAE at iGrid2002 and SC2002,
with Oracle9i storing JETMET data at CERN, SQLServer storing a
complementary set of JETMET data at Caltech, and then perhaps more
data in PostgreSQL or MySQL at UCLA and Rio. Bandwidth usage will
need to be very heavy in order to interest participants at the
conference, so we are adopting a somewhat artificial analysis scheme
which involves the use of GridFTP to move files across the WAN. Perhaps
at SC2002 we will use Clarens instead of GridFTP, if we can reproduce
the streaming performance of GridFTP on the special striped servers
being provided by the Globus team.
A particular item of work is on developing a database layer to which
the Clarens server talks in SQL, and which is responsible for returning
a vector of physics objects. Alternatively this layer would generate
a file containing the objects, which may then be transferred using
GridFTP.
We are committed to using the Globus authentication methods as part of
the demonstrations, and as part of the GAE being developed. We will
use DoE Science Grid certificates, which are being obtained for the
servers and users involved.
As far as hardware for the demonstrations is concerned, we have spent
some time analysing the market for the best motherboard to use to
maximise the Gbit NIC performance of a server. With help from
Ian Fisk (UCSD) we have ordered two powerful servers
based on the SuperMicro P4DP6 board, each equipped with dual SysKonnect
SK9843 optical NICs. Tests of the network performance in back-to-back
LAN configurations are proceeding. We also have acquired Dlink Gbit
NICs for testing. A further server has been ordered, based on the
SuperMicro P4DP8-G2 motherboard, featuring on-board dual Gbit NICs
(copper). This server will also be tested. Finally, once a suitable
configuration has been proven to allow sustained high bandwidth
transfers
in the LAN, we will purchase several for placement at CERN, Caltech and
at the STARLIGHT in Chicago.
Work on optimising the WAN performance of the GAE is an important aspect
of our activities, and we have started a very fruitful collaboration
with Steven Low's (Caltech) group on Adaptive Queue Managament for TCP,
which promises much better performance for large data transfers in the
WAN. This work is being partially funded by a successful proposal we
submitted to NSF.
|