ATLAS on UKLight Large Hadron Colider LHC will collide beams of protons at an energy of 14 TeV Using the latest super-conducting technologies, it will operate at about 2700C, just above absolute zero of temperature. With its 27 km circumference, the accelerator will be the largest superconducting installation in the world. 4 LHC Experiments ATLAS - general purpose: origin of mass, supersymmetry, micro-black holes, where did the antimatter go? -Also top quarks, standard model, heavy ions physics (quark gluon plasma) - 2,000 scientists from 34 countries As featured in FHM! LHC Data Challenges Starting from this event… Selectivity: 1 in 1013 Like looking for 1 person in a thousand world populations Or for a needle in 20 million haystacks! ..we are looking for this “signature” A particle collision = an event We need: Detectors to record Triggers to select Computing and software to process/reconstruct/simulate Computing, software & physicists to refine selection and analyse Why ATLAS needs UKLight But most data simulated/analysed not at CERN Need High Bandwidth for File Transfer ~10s PB/year (Disc-Disc SCP @ 1MB/s)/(Single site Analysis) not an option Need for Grid for particle physics LCG and UK GRIDPP needed Three Main Aims of ATLAS ESLEA collaborators Increasing capability of large bulk file transfers over WAN Real Time Analyse of Calibration/Alignment Data Monitoring Links Real Time Monitor WeatherMap Archival history / Billing Info Tier Model Tiered service varies level of support/functionality. Complex Data Flows UK has distributed Tier2s comprising multiple university sites RAL is UK Tier1 Lancaster in Tier2 NORTHGRID with: Manchester, Liverpool, Sheffield, Daresbury Lab. T0 T1 T1 Site Site T2 T2 Site T1 Site T1 Atlas Tier1 Model and data flows Tape RAW Real data storage, reprocessing and distribution ESD2 RAW Tier-0 AODm2 1.6 GB/file 0.02 Hz 1.7K f/day 32 MB/s 2.7 TB/day 0.044 Hz 3.74K f/day 44 MB/s 3.66 TB/day ESD1 AODm1 RAW AOD2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day 500 MB/file 0.04 Hz 3.4K f/day 20 MB/s 1.6 TB/day 1.6 GB/file 0.02 Hz 1.7K f/day 32 MB/s 2.7 TB/day 10 MB/file 0.2 Hz 17K f/day 2 MB/s 0.16 TB/day ESD2 AODm2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day 500 MB/file 0.036 Hz 3.1K f/day 18 MB/s 1.44 TB/day Other T1 T1 Tier-1s disk buffer CPU farm ESD2 AODm2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day 500 MB/file 0.004 Hz 0.34K f/day 2 MB/s 0.16 TB/day disk storage ESD2 AOD2 AODm2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day 10 MB/file 0.2 Hz 17K f/day 2 MB/s 0.16 TB/day 500 MB/file 0.004 Hz 0.34K f/day 2 MB/s 0.16 TB/day AODm1 AODm2 500 MB/file 0.04 Hz 3.4K f/day 20 MB/s 1.6 TB/day 500 MB/file 0.04 Hz 3.4K f/day 20 MB/s 1.6 TB/day Tier-2s T1 T1 Plus simulation & analysis data flow ESD2 AODm2 0.5 GB/file 0.02 Hz 1.7K f/day 10 MB/s 0.8 TB/day 500 MB/file 0.036 Hz 3.1K f/day 18 MB/s 1.44 TB/day Other T1 T1 Tier-1s Testing of Model Day-Day Running & Challenges Both experimental software and generic middleware stack Data challenges were used by experiments to test experiment software “Data Challenges” “Can we actually simulate data and know how to store/analyse it?” Service Challenges through (W)LCG were used to test middleware and networks “Service Challenges” “Can we provide the services that the experiments want/need?” Combination of service and data challenges towards full rates ongoing and will continue to ramp up until Full data taking in Spring ‘07 Latest phase of ATLAS involvement started 19th June Hardware Configuration Computing Element (CE) and other standard LCG EGEE Services Monitoring Node EGEE User Interface (UI) node Storage Element (SE) Storage Resource Managed (SRM) dCache Head Node 6 Pool Nodes 2 x 6TB RAID5 storage arrays Network Configuration LAN continually evolving Connected to both UKLight and Production network Production Network CE & Production Services ROW 100Mbps limit to University network for management/service RAL communication Network SE & UKLight Upgrading SE-CE connection to 1Gbps Endpoint 7609 Router UKLight Monitoring High spec 2.4 TB Raid UKLight Network •DHCP/DNS organised via University/RNO •Heavy use of static routes with backup over JANET network Software Configuration Scientific Linux 3 OS with linux_2.4.21-40 kernel Basic TCP-Tuning to increase default/max TCP_window_size, txqueuelen LCG software stack including dCache srmcp globus-url-copy EGEE File Transfer Server FTS Access to EGEE User Interface (UI) to initiate transfers ATLAS Distributed Data Management (DDM) Monitoring Ping, Traceroute, Iperf, Pathload and Pathrate for Line Testing MRTG for Cisco 7609 Monitoring/Input to RTM/Weathermap Lancaster-RAL Link T1-T2 transfer testing Bulk of T2 transfers will be to T1 1 Gbps line Iperf/Pathrate/Pathload close to design 12 to 2 hops using Traceroute Avoid production network induced bottlenecks 400 Mbps firewall at RAL File Transfer Tests Tested using : dCache version of GridFTP srmcp Also using FTS server to control transfers Achieved Peak of 948Mbps Transferred: 8TB in 24 hours - 900+ Mbps aggregate rate 36TB in 1 week - 500+ Mbps aggregate rate Parallel file transfers increase rate Better utilisation of bandwidth Staggered initialisation of transfers reduces overhead from initialisation/cessation of individual transfers. Rate increase from 150Mbps to 900Mbps 2% (18Mbps) reverse traffic flow for 900Mbps transfer File size affects rate of transfer FTS transfers not yet as successful as srmcp only transfers Greater overhead Single FTS file transfer gives 150Mbps Same as srmcp 400Mbps Max for 10 concurrent FTS transfers All single stream transfers Single stream rate varies 150 to 180 Mbps with increase from 1 to 10 GB RAL-CERN Link LINK on UKLight as initial connection To be replaced by LCG Optical Private Network Replacement/SJ5/Geant2 UKLight to SARA; then Netherlight to CERN Support/Manpower available for assistance 4 Gbps Lancaster-SARA Link Link not yet active SARA capacity underused, RAL capacity currently too small for UK simulation storage Also, SARA planned by ATLAS to be Tier1 in case of RAL downtime ( FTS catalogues etc.) Tests similar to Lancaster-RAL Tests T1-T2 testing Study of effect of International/Extended link length, alternate protocols etc.. Lancaster-Manchester Link Intra-Tier2 site Testing 1Gbps Initial IPERF 220/800 Mbps with and without “basic” TCP Tuning “Homogeneous Distributed Tier2” dCache Head node at Lancs, pool nodes at both Lancs and Mancs Test Transfers to/from RAL Test of Job submission to close CE Possible testing of xrootd Lancaster-Edinburgh Link Link not active yet UDT Protocol Testing ESLEA box 6x400GB SATA RAID disks with 3Ware SATA RAID controller Details of software configuration TBC Hand over to Barney::::::