Geneva – Krak - University of Manchester

Download Report

Transcript Geneva – Krak - University of Manchester

Geneva – Kraków network measurements for the ATLAS
Real-Time Remote Computing Farm Studies
R. Hughes-Jones (Univ. of Manchester), K. Korcyl (IFJ-PAN), C. Meirosu (CERN)
MOTIVATION
Several experiments, including ATLAS at the Large Hadron Collider (LHC) and D0 at
Fermi Lab, have expressed interest in using remote computing farms for processing
and analysing, in real time, the information from particle collision events. Different
architectures have been suggested from pseudo-real-time file transfer and
subsequent remote processing, to the real-time requesting of individual events as
described here.
Remote Computing Concepts
Remote Event Processing Farms
Copenhagen
Edmonton
Krakow
Manchester
ATLAS Detectors – Level 1 Trigger
To test the feasibility of using remote farms for real-time processing, a collaboration
was set up between members of ATLAS Trigger/DAQ community, with support from
several national research and education network operators (DARENET, Canarie,
Netera, PSNC, UKERNA and Dante) to demonstrate a Proof of Concept and measure
end-to-end network performance. The testbed was centred at CERN and used three
different types of wide area high-speed network infrastructures to link the remote
sites:
• an end-to-end lightpath (SONET circuit) to the University of Alberta in Canada
• standard Internet connectivity to the University of Manchester in the UK and the
Niels Bohr Institute in Denmark
• a Virtual Private Network (VPN) composed out of an MPLS tunnel over the GEANT
and an Ethernet VPN over the PIONIER networks to IFJ PAN Krakow in Poland.
ROB
ROB
ROB
ROB
PF
Data Collection
Network
L2PU
L2PU
L2PU
L2PU
Event
SFI Builders
SFI
PF
lightpaths
SFI
PF
PF
Back End
Network
GÉANT
Level 2 Trigger
Local Event PF
Processing Farms
Experimental Area
SFOs
CERN B513
Mass storage
Switch
PF
EQUIPMENT
We developed custom measuring equipment to measure the quality of service at Layer 2/3 .
The equipment is based on an Alteon Gigabit Ethernet (GE) network interface card (NIC)
reprogrammed to act as an IP traffic generator, a custom clock card and commercial Global
Positioning System equipment (used as a global clock time reference). We can measure oneway latency, inter-arrival time, frame loss and re-ordering on a packet-by-packet basis, as a
function of load, up to Gigabit Ethernet speed.
OWN FIBERS
GDAŃSK
Geant
10 Gb/s
KOSZALIN
OLSZTYN
SZCZECIN
GÉANT
BIAŁYSTOK
BYDGOSZCZ
TORUŃ
GÉANT
PNSC
POZNAŃ
ZIELONA
GÓRA
For the Layer 4 (TCP) measurements we developed and used the “tcpmon” program. The
tcpmon is an instrumented request-response program that emulates the communication
between the EFD and SFI components of the ATLAS Event Filter. It is builds on the
experience of UDPmon, a generic tool that can be used to automate hardware and network
performance measurements using UDP packets. UDPmon calculates the CPU load and the
number of interrupts generated by the network interface card during a given test, along with
standard network-related parameters like latency, inter-arrival time and bandwidth. TCPmon
and UDPmon run on standard PCs under the Linux operating system
10 Gb/s
LEASED CHANNELS
WARSZAWA
622 Mb/s
ŁÓDŹ
WROCŁAW
OPOLE
155 Mb/s
RADOM
CZĘSTOCHOWA
KIELCE
PUŁAWY
LUBLIN
Metropolitan
Area
Networks
KATOWICE
KRAKÓW
RZESZÓW
BIELSKO-BIAŁA
KRAKÓW
CERN
ATLAS Application Protocol
SFI and SFO
Event Filter EFD
Request event
Send event
data
Request-Response
time (Histogram)
Process event
Request Buffer
Send OK
Send processed
event
•
●●
●
Time
Event Request
– EFD requests an event from SFI
– SFI replies with the event ~2Mbytes
•
•
RESULTS – summary:
Processing of event
Return of computation
– EF asks SFO for buffer space
– SFO sends OK
– EF transfers results of the computation
• out of order frames present, even if the connection is
composed of a Layer 2 tunnel over MPLS and a pure
Ethernet VLAN
• the number of out of order frames may vary,
depending on the offered load - we have not observed
any out of order frames during our tests for loads lower
than 500 Mbit/s
• steady state request-response
latency: ~140 ms
• event rate: ~7.2 events/s
• the first event took 600 ms
(due to the start-up time on the
TCP connection)
• relevant to the 1.5 MB transfers we would have in our
application ? Yes !
• minimal: using a modern TCP stack, if frames are
“not too much out of order”, the stack will not
request re-transmits. But the CPU load, required for
the bookkeeping, is higher.
• worst case: the stack will require a re-transmit,
halving the TCP window in the process hence
reducing the maximum transfer rate for a given time
interval
Web100 parameters on the server located at CERN (data source)
Green – small requests
Blue – big responses
TCP ACK packets also counted (in each direction)
One response = 1 MB ~ 380 packets
CONCLUSIONS
The quality of service over long-distance connectivity may vary quite a lot momentarily, even though it might still meet the Service Level Agreement (based on long-term averages).
Long-distance Ethernet circuits, tunneled over routed networks, may produce out-of-order packets – which would be not the case in a LAN environment.
Application with real-time requirements should monitor the performance of the underlying network and adapt accordingly.
Out of order packets are important for our application. Studies are under way to determine the real impact.