Title: Distributed Metadata with the AMGA Metadata Catalog
1Distributed Metadata with the AMGA Metadata
Catalog
- Nuno Santos, Birger Koblitz
- 20 June 2006
- Workshop on Next-Generation Distributed Data
Management
2Abstract
- Metadata Catalogs on Data Grids The case for
replication - The AMGA Metadata Catalog
- Metadata Replication with AMGA
- Benchmark Results
- Future Work/Open Challenges
3Metadata Catalogs
- Metadata on the Grid
- File Metadata - Describe files with
application-specific information - Purpose file discovery based on their contents
- Simplified Database Service Store generic
structured data on the Grid - Not as powerful as a DB, but easier to use and
better Grid integration (security, hide DB
heterogeneity) - Metadata Services are essential for many Grid
applications - Must be accessible Grid-wide
- But Data Grids can be large
4An Example - The LCG Sites
- LCG LHC Computing Grid
- Distribute and process the data generated by the
LHC (Large Hadron Collider) at CERN - 200 sites and 5.000 users worldwide
Taken from http//goc03.grid-support.ac.uk/google
maps/lcg.html
5Challenges for Catalog Services
- Scalability
- Hundreds of grid sites
- Thousands users
- Geographical Distribution
- Network latency
- Dependability
- In a large and heterogeneous system, failures
will be common - A centralized system does not meet the
requirements - Distribution and replication required
6Off-the-shelf DB Replication?
- Most DB systems have DB replication mechanisms
- Oracle Streams, Slony for PostgreSQL, MySQL
replication
- Example 3D Project at CERN
- (Distributed Deployment of Databases)
- Uses Oracle Streams for replication
- Being deployed only at a few LCG sites (10
sites, Tier-0 and Tier-1s) - Requires Oracle () and expert on-site DBAs
() - Most sites dont have these resources
- Off-the-shelf replication is vendor-specific
- But Grids are heterogeneous by nature
- Sites have different DB systems available
Only partial solution to the problem of metadata
replication
7Replication in the Catalog
- Alternative we are exploring
- Replication in the Metadata Catalog
- Advantages
- Database independent
- Metadata-aware replication
- More efficient replicate Metadata commands
- Better functionality Partial replication,
federation - Ease of deployment and administration
- Built-in into the Metadata Catalog
- No need for dedicated DB admin
- The AMGA Metadata Catalogue is the basis for our
work on replication
8The AMGA Metadata Catalog
- Metadata Catalog of the gLite Middleware (EGEE)
- Several groups of users among the EGEE community
- High Energy Physics
- Biomed
- Main features
- Dynamic schemas
- Hierarchical organization
- Security
- Authentication user/pass, X509 Certs, GSI
- Authorization VOMS, ACLs
9AMGA Implementation
- C implementation
- Back-ends
- Oracle, MySQL, PostgreSQL, SQLite
- Front-end - TCP Streaming
- Text-based protocol like TELNET, SMTP, POP
- Examples
- Adding data
- Retrieving data
addentry /DLAudio/song.mp3 /DLAudioAuthor
John Smith /DLAudioAlbum Latest Hits
selectattr /DLAudioFILE /DLAudioAuthor
/DLAudioAlbum like(/DLAudioFILE, .mp3")
10Standalone Performance
- Single server scales well up to 100 concurrent
clients - Could not go past 100. Limited by the database
- WAN access one to two orders of magnitude slower
than LAN -
Replication can solve both bottlenecks
11- Metadata Replication with AMGA
12Requirements of EGEE Communities
- Motivation Requirements of EGEEs user
communities. - Mainly HEP and Biomed
- High Energy Physics (HEP)
- Millions of files, 5.000 users distributed
across 200 computing centres - Mainly (read-only) file metadata
- Main concerns scalability, performance and
fault-tolerance - Biomed
- Manage medical images on the Grid
- Data produced in a distributed fashion by
laboratories and hospitals - Highly sensitive data patient details
- Smaller scale than HEP
- Main concern security
13Metadata Replication
Some replication models
Partial replication
Full replication
Federation
Proxy
14Architecture
- Main design decisions
- Asynchronous replication for tolerating with
high latencies and fault-tolerance - Partial replication Replicate only what is
interesting for the remote users - Master-slave Writes only allowed on the master
- But mastership is granted to metadata
collections, not to nodes
15Status
- Initial implementation completed
- Available functionality
- Full and partial replication
- Chained replication (master ? slave1 ? slave2)
- Federation - basic support
- Data is always copied to slave
- Cross DB replication PostgreSQL ? MySQL tested
- Other combinations should work (give or take some
debugging) - Available as part of AMGA
16 17Benchmark Study
- Investigate the following
- Overhead of replication and scalability of the
master - Behaviour of the system under faults
18Scalability
- Setup
- Insertion rate at master 90 entries/s.
- Total 10,000 entries
- 0 slaves - saving replication updates, but not
shipping (slaves disconnected)
- Small increase in CPU usage as number of slaves
increases - 10 slaves, 20 increase from standalone operation
- Number of update logs sent scales almost linearly
19Fault Tolerance
- Next test illustrates fault tolerance mechanisms
- Slave fails
- Master keeps the updates for the slave
- Replication log grows
- Slave reconnects
- Master sends pending updates
- Eventually system recovers to a steady state with
the slave up-to-date - Test conditions
- Insertion rate at master 50 entries/s
- Total 20.000 entries
- Two slaves, both start connected
- Slave1 disconnects temporarily
20Fault Tolerance and Recovery
- While slave1 is disconnected, the replication log
grows in size - Limited in size. Slave unsubscribed if it does
not reconnect in time. - After slave reconnection, system recovers in
around 60 seconds.
21- Future Work/Open Challenges
22Scalability
- Support hundreds of replicas
- HEP use case. Extreme case one replica catalog
per site - Challenges
- Scalability
- Fault-tolerance tolerate failures of slaves and
of master - Current method of shipping updates (direct
streaming) might not scale - Chained replication (divide and conquer)
- Already possible with AMGA, performance needs to
be studied - Group communication
23Federation
- Federation of independent catalogs
- Biomed use case
- Challenges
- Provide a consistent view over the federated
catalogs - Shared namespace
- Security - Trust management, access control and
user management - Ideas
24Conclusion
- Replication of Metadata Catalogues necessary for
Data Grids - We are exploring replication at the Catalogue
using AMGA - Initial implementation completed
- First results are promising
- Currently working on improving scalability and on
federation - More information about our current work at
- http//project-arda-dev.web.cern.ch/project-arda-d
ev/metadata/