See change log at bottom !                07.06.2013 15:57:24

Overview

The jFEX processor system performs jet, large tau, missing-ET and total-ET trigger algorithms on 0.1 × 0.1 electromagnetic and hadronic towers, as described above. The jFEX system (see fig. x) consists of a single AdvancedTCA (ATCA) shelf, equipped with several module types and is fed by an optical fibre plant. The fibres are routed via rear transition modules into jFEX modules, where physics algorithms are run. Timing distribution, ROD and control hub functionality are implemented on two hub modules located in the “fabric slots” of the shelf.

 Each jFEX module processes a phi-octant of calorimeter data, covering a strip along eta, comprising barrel, both endcap, and forward calorimeters. A total of 8 jFEX modules is required for full coverage of the calorimeter solid angle.

Jets and taus are identified on the jFEX modules as local maxima of energy deposits in a window sliding over the detector space, within an environment of 0.9 x 0.9 in eta x phi. This requires that each jFEX module is supplied with a copy of the data from neighbouring phi-octants. The data duplication is performed at the data source, the Digital Processor System. A diagram of the jFEX module is shown in fig. x.

 

Real-time data path

Real-time input data from the Digital Processor System (DPS) and the legacy processors enter the jFEX module from the back, via the rear transition module (RTM). The RTM is a passive circuit providing the mechanics for feeding fibres into the jFEX module on 72-way MTP/MPO connectors. Using standard opto backplane connectors, there is space for four of them in ATCA zone 3. That allows for a total of 288 incoming fibres per module.

On the jFEX module the four fibre bundles are each split up into six twelve-way ribbons that are terminated into “MicroPOD” electro-optical devices. The devices support data rates of up to 14Gb/s per fibre. The MicroPODs require custom-built mechanics to provide mechanical stability and to spread the heat dissipated by the devices. The input signals are routed on to the processor FPGAs as differential, CML level signals. Here they enter the “GTH” high speed receivers and are processed in the FPGA fabric after de-serialization to the LHC bunch clock frequency.

The processor FPGAs identify and report the trigger objects (TOBs, i.e. jet and tau candidates) and calculate global quantities (missing Et). The data reported by five of the processors on their fabric interface are consolidated and merged in the sixth one before being sent out optically via MicroPOD devices, on to L1Topo. So as to be able to duplicate data into more than one L1Topo module, two MicroPOD devices (i.e. 24 optical fibres) are allocated for that purpose. The fibre bundles are routed through the front panel of the jFEX on standard MTP/MPO connectors.

Signal duplication

At an environment of 0.9 x 0.9 the sliding window algorithms requires neighbour data of 4 cells to either side of each core cell to be processed (see fig.x). Due to 100% data duplication of all signals at the DPS, all environment required for processing the core cells of one calorimeter octant is routed onto each jFEX module. The algorithms are partitioned to run in six FPGAs per jFEX module. Each of the FPGAs needs access to the core cells it processes, and the environment of +/- four cells. Therefore a large amount of data needs to be duplicated between neighbouring FPGAs.

For inter-FPGA data duplication the baseline design makes use of the “PMA loopback” scheme, as described in the eFEX section: All incoming data are routed point-to-point into one FPGA receiver. After a short latency penalty of just over one bunch tick the data are routed out of the corresponding transmitter, on to the neighbouring FPGA.

The data are transmitted 8b/10b-encoded at 6.4Gb/s line rate. As outlined in sect. x the net data volume of 128 bit per bunch crossing per fibre allows for transmission of eight energy samples on each FPGA input channel, including a checksum. Data are organized in 0.4 x 0.2 in eta x phi per channel. Two samples, electromagnetic and hadronic energy, are required for each trigger tower.

A single large FPGA with 80 input channels can thus receive a maximum of 20 * 16 trigger towers. This is 12 * 8 core cells plus environment data. Due to irregularities in the calorimeter overlap region and in the FCAL area the mapping needs to be specifically implemented at PCB level. Two of the jFEX processor FPGAs will actually have to be larger, and more expensive, 96-channel devices, so as to cover the overlap region. Out of the devices currently on the market, the Xilinx devices XC7VX690T and XC7VX1140T appear well suited for the jFEX module. For reason of latency, the fastest speed grade available will have to be chosen for the real-time processors.

Timing

All real-time data transmission and processing is performed synchronously at multiples of the LHC bunch clock. The clock as well as TTC control data are received in backplane zone 2 on two signal pairs each from the hub modules. Either hub can be selected to act as a clock source. After signal conditioning (jitter reduction) and frequency multiplication the clock is routed on to the high-speed link circuitry and into the FPGA fabric.

Readout

Readout and data transmission to the 2nd level Trigger are implemented as described in the eFEX section. Processor data are read out into two hub modules on six high-speed links each (one link per processor FPGA), allocated in ATCA Zone 2. Recent ATCA backplanes support data rates of up to 10Gb/s. This data rate is supported by the processor FPGAs, and is sufficient for readout of both incoming and internal data. In normal running the input data are expected to be read out only in the unlikely event of checksum mismatches on the received real-time data.

Configuration, Control, monitoring and diagnostics

The real-time processors will be configured upon power-up through the control FPGA. The configuration bit stream will be stored locally on an SD card. The control FPGA itself is configured via an SPI flash device and the control firmware includes the algorithms to read and write the SD card and sequence the configuration data into the processor FPGAs.

For module control the IPBus firmware suite is employed. The base interface in backplane zone 2 is routed to an Ethernet phy device, controlled by the control FPGA. Here a small footprint IP stack is run directly in firmware. IPBus firmware maps Ethernet control and monitoring packets to a processor-like bus interface, which in turn reads and writes “registers” inside the FPGA. This access path is used to control and monitor the operation of the processor FPGAs. Parameters for the algorithms are set, and control of the high-speed links is performed that way, playback- and spy memories attached to the real-time path are exercised, and the environment (temperatures and voltages) is monitored.

Low level ATCA compliant control is achieved by an ATLAS standard IPMC module.

Options

The baseline design described above allows for the determination of jets and large taus in an environment of 0.9 x 0.9 in eta x phi. This is an acceptable balance of physics needs versus achievable system density and signal integrity. There is a very strong motivation though to work towards considerably higher input data rates, which allow a move towards significantly larger jet sizes. R&D projects have been started to validate the feasibility and derive limits on data rates in dense electro-optical systems with very high channel counts and in conjunction with data duplication schemes. The jFEX system has been designed to fully benefit from an increase in transmission speed to maximise the jet size.

Phase II

The jFEX system is built such that it can be operated as part of the L0 trigger after the phase-2 upgrade. The real-time data path and the processing are expected to run unchanged, though the sliding window algorithms might be upgraded. The chosen processor FPGA types are providing enough logic resources to allow for algorithmic upgrades.

In phase-2 the distribution of clock and timing signals is expected to change, as is the readout. These functionalities are, however, not handled on the jFEX modules directly. Any modifications required would affect the hub module mezzanines only.

 

 

3 drawings: system, jFEX, duplication scheme

 

Captions:

 

The jFEX processor system

 

jFEX module and major components

 

Processing area of 12*8 core cells (green) and the environment received from DPS (magenta) and duplicated on jFEX module (orange)

 

 

 

 

 

 

 

 

 

 

 

 

 

 


Change log:

07.06.14:20 major rework.

06.06. 17:32 added initial versions of drawings. Still working on that !

06.06. 17:18 changes in 1st para to better distinguish jFEX system and module.