SC2002 Bandwidth Challenge and Data Challenge Application - PowerPoint PPT Presentation

1 / 18
About This Presentation
Title:

SC2002 Bandwidth Challenge and Data Challenge Application

Description:

calorimeter-2. energy. tracker. reconstruction. cluster ... calorimeter-1. D/A converter. cluster-2. cluster-3. track-2. track-3. track-4. track-5. jet id ... – PowerPoint PPT presentation

Number of Views:94
Avg rating:3.0/5.0
Slides: 19
Provided by: youhei
Category:

less

Transcript and Presenter's Notes

Title: SC2002 Bandwidth Challenge and Data Challenge Application


1
SC2002 Bandwidth Challenge andData Challenge
Application
  • KEK
  • Computing Research Center
  • Y. Morita

2
????
Fermion (Matter)
Gauge particles (Force)
1st gen.
2nd gen.
3rd gen.
Strong Force
Charm
Top
Up
Quark
Gluon
Electromagnetic Force
Down
Strange
Bottom
Photon
Weak Force
e neutrino
? neutrino
? neutrino
Lepton
W Boson
Z Boson
electron
muon
tau
Particle related to the Higgs field (not
discovered)
Higgs
3
Large Hadron Collider at CERN
4
ATLAS Detector
1850 physicists from 33 countries
dimensions 20x20x40 m weight 7000
ton readout ch 1.5 x 108
5
Theory vs Experiment
  • To discover a new physics in the experiments,
    signal must be separated from the background
    process (well-known physics)
  • Event simulator plays a crucial role in this
    comparison

Theory
Experiment
Event Generator
Raw Data
particles
Reconstruction Analysis
Event Simulator
Simulated Raw Data
Comparison
6
Physics Analysis Challenges
Finding a needle in a heystack
  • 109 collisions/second ? online filter ? 100
    events/second on storage ? 109 events/year
  • 1Mbytes/event ? several PetaBytes/year
  • Event reconstruction 300 SPECint95sec/event ?
    200K SPECint95 for reconstruction

High Throughput, Data-intensive Computing
7
HEP data reconstruction / analysis
1PB/year 1MB/event
1PB
300TB/year 100KB/event
10TB/year 10KB/event
8
Multi-Tier Regional Center Scheme
Multi-tier Regional Center Model for LHC
4 TIPS
24 March 2000, WW A/C Panel, P. Capiluppi
9
LCG LHC Compuing Grid Project
  • Deployment of Computing and Software model for
    the 4 LHC experiments
  • WG1 Choice of Security Middleware and Tools
  • WG2 VO management and resources
  • WG3 Registration, Authentication, Authorization
    and Security
  • WG4 Security Operational Procedures
  • LCG-1 estimatesUsers 1000User Registration
    Peak rate 25 users/day in 2003 2Q

10
ATLAS Data Challenges
  • 2002 Data Challenge 1 0.1
    test Regional Center Test High Level
    Trigger studies AprAug Phase1 Event Full
    Simulation (Fortran) OctJan Phase2 Event
    PileUp (Fortran) 3 x 106 events, 25TB
  • 20034 Data Challenge 2 "10" test Full
    chain test of C software with Grid Validation
    of the LCG computing model
  • staged validation process for the computing and
    software models with increasing magnitude

11
Atlas Software
  • Technical Design Proposal in Fortran program
  • In transition new generation C program
  • Event Generator Fortran program wrapped with C
  • Event Simulator FADS/Goofy framework with Geant4
  • Event Reconstruction/Analysis Athena/Gaudi
    framework FADS/Goofy works also as a module of
    Athena/Gaudi
  • Writing the full detector simulation,
    reconstruction, analysis modules is an ongoing
    world-wide software integration effort
  • Validation of this integration is one of the
    major goals of the Atlas Data Challenge

12
About Gfarm
  • Grid Data Farm
  • Middleware Project between AIST, KEK and Titech
  • Parallel File System taking the advantage of the
    parallel nature of event oriented data and
    statistics analysis
  • owner computes rulejob runs on the node where
    thedata resides
  • job history and the file segmentlocations are
    managed bya Metadatabase
  • File fragments are copied forbackup and load
    balancing
  • User sees the file fragmentsvia single image
    logical file URL

http//datafarm.apgrid.org/
13
CPU vs Storage in High I/O jobs
  • Simple management of system and file
  • Network and switches becomes the bottleneck in
    high I/O multi-user applications
  • Does not scale to more than a few hundrednodes
    system
  • Independent local I/O on each node
  • Scalability for more than thousands nodes
  • system and file management become complex

14
FADS/Goofy architecture for SC2002
  • FADS/GoofyFramework for ATLAS Detector
    Simulation / Geant4-based Object-oriented Folly

Atlas Detector
User Analysis Module
FADS/Goofy
Geometry Material
Particles
Tracking
Events I/O
Geant4
ROOT
MySQL
xerces-c
HepMC
ROOT
Gfarm
Gfarm
Converter
Plug-in architecture
Event Generator
Objectivity/DB
Hits files
Histogram files
15
Presto-III PC Cluster _at_ Titech
  • of Nodes 256
  • CPU AMD Athlon x 2 (Thunderbird, AXIA core)
    1.33GHz (FSB133MHz)
  • Motherboard ASUS A7V13 (VIA KT133A Chipset)
  • Memory 768MB
  • HDD 40GB
  • OS Debian/Lucie 2.14.7
  • g 2.95.4
  • Network Card 1 DEC 21140AF
  • Network Card 2 Myricom Myrinet2000
  • 47-th in TOP500 (2nd in PC cluster)

16
FADS/Goofy with Gfarm
  • Framework for Monte Carlo Detector Simulation
    using Geant4 toolkit in C
  • Parallel Event Processing with Atlas Detector
    Full Simulation
  • Parallel Object I/O capability with ROOT and
    Objectivity/DB on Gfarm file system
  • Parallel network transfer and replication over
    gigabit WAN
  • World-wide distributed data mining and
    histograming for petabyte scale data
  • Robustness TestGenerated 106 events with Titech
    PrestoIII cluster in 2 days
  • Replicated the simulated events over WAN to AIST,
    SC2002, Indiana, SDSC
  • Gfarm data replication is used as the bandwidth
    challenge-gt see Tatebe sans talk

17
Cluster and Network setting for SC2002 Bandwidth
Challenge (9/3)
18
(No Transcript)
Write a Comment
User Comments (0)
About PowerShow.com