Operation and performance of the ATLAS Semiconductor Tracker
description
Transcript of Operation and performance of the ATLAS Semiconductor Tracker
+
Operation and performance of the ATLAS Semiconductor Tracker
Nick Barlow,University of Cambridge,On behalf of the ATLAS SCT collaboration.
2+Contents
Overview of LHC/ATLAS/SCT Design Operational issues Performance Current activities
3+Introduction The Large Hadron Collider at
CERN is the world’s highest energy particle accelerator.
Beams of 4TeV protons can be made to collide head-on at 4 points around the ring, where particle detectors record the results of the collisions.
4+Introduction The Large Hadron Collider at
CERN is the world’s highest energy particle accelerator.
Beams of 4TeV protons can be made to collide head-on at 4 points around the ring, where particle detectors record the results of the collisions.
ATLAS is the largest of these detectors, designed to study the Standard Model and search for new particles.
5+Introduction The Large Hadron Collider at
CERN is the world’s highest energy particle accelerator.
Beams of 4TeV protons can be made to collide head-on at 4 points around the ring, where particle detectors record the results of the collisions.
ATLAS is the largest of these detectors, designed to study the Standard Model and search for new particles.
The Inner Detector (ID) is the innermost part of ATLAS, to measure trajectories of charged particles (“tracks”).
6+ATLAS Inner Detector The ID consists of:
Pixel detector Semiconductor Tracker (SCT) Straw tube tracker (TRT)
All within 2T solenoidal B-field.
7+ATLAS Inner Detector The ID consists of:
Pixel detector Semiconductor Tracker (SCT) Straw tube tracker (TRT)
All within 2T solenoidal B-field.
Pixel and SCT kept cold by evaporative cooling, using C3F8
8+ATLAS Inner Detector The ID consists of:
Pixel detector Semiconductor Tracker (SCT) Straw tube tracker (TRT)
All within 2T solenoidal B-field.
Pixel and SCT kept cold by evaporative cooling, using C3F8
The SCT is made up of 4 cylindrical barrel layers and 9 endcap disks on each side.
9+ATLAS Inner Detector The ID consists of:
Pixel detector Semiconductor Tracker (SCT) Straw tube tracker (TRT)
All within 2T solenoidal B-field. Pixel and SCT kept cold by
evaporative cooling, using C3F8
The SCT is made up of 4 cylindrical barrel layers and 9 endcap disks on each side.
Consists of 4088 double-sided silicon modules. ‘p-in-n’ silicon strip sensors.
10+ATLAS Inner Detector and SCT
Each module side has: 768 aluminium strips (pitch=80
microns) (>6 Million channels in total!)
6 ABCD3TA ASIC readout chips.
The 2 sides of a module have stereo angle of 40mrad between strip directions, in order to give 2D position of “hits”.
The ID consists of: Pixel detector Semiconductor Tracker (SCT) Straw tube tracker (TRT)
All within 2T solenoidal B-field. Pixel and SCT kept cold by
evaporative cooling, using C3F8
The SCT is made up of 4 cylindrical barrel layers and 9 endcap disks on each side.
Consists of 4088 double-sided silicon modules. ‘p-in-n’ silicon strip sensors.
11+SCT readout / Data Acquisition (DAQ)
Data from the SCT are read out by off-detector electronics in eight DAQ crates.
Trigger signals received via optical link to Trigger and Timing Crate (TTC).
12+SCT readout / Data Acquisition (DAQ)
Data from the SCT are read out by off-detector electronics in eight DAQ crates.
Trigger signals received via optical link to Trigger and Timing Crate (TTC).
Trigger and clock signals sent to modules along optical “TX” link (one fibre per module).
13+SCT readout / Data Acquisition (DAQ)
Data from the SCT are read out by off-detector electronics in eight DAQ crates.
Trigger signals received via optical link to Trigger and Timing Crate (TTC).
Trigger and clock signals sent to modules along optical “TX” link (one fibre per module).
Modules then return hit data along “RX” link (one fibre per side)
14+SCT readout / Data Acquisition (DAQ)
Data from the SCT are read out by off-detector electronics in eight DAQ crates.
Trigger signals received via optical link to Trigger and Timing Crate (TTC).
Trigger and clock signals sent to modules along optical “TX” link (one fibre per module).
Modules then return hit data along “RX” link (one fibre per side)
Binary readout – either “1” or “0” in each 25ns time bin, depending on whether or not charge exceeded a configurable threshold.
15+SCT readout / Data Acquisition (DAQ)
Data from the SCT are read out by off-detector electronics in eight DAQ crates.
Trigger signals received via optical link to Trigger and Timing Crate (TTC).
Trigger and clock signals sent to modules along optical “TX” link (one fibre per module).
Modules then return hit data along “RX” link (one fibre per side)
ROD assembles and formats data from 48 modules, then sends along “S-link” to ‘ROS’ (central ATLAS DAQ).
16+Redundancy in optical communications
Two types of redundancy implemented, in case of failure of optical transmitter or damage to optical fibre. TX redundancy: module can receive clock and command
signals electronically from neighbouring module. Cannot “daisy-chain” – if two adjacent modules lose
optical input on TX line, we will lose data from at least one.
RX redundancy: both sides of a module can be read out through one RX link. In most barrel modules, this involves bypassing one chip
– lose data from 128 strips.
17+Timing Upon receipt of a trigger (via TX
line), the ABCD will send back (along RX)data from the last three 25ns time bins in its “pipeline” If we are correctly “timed in”, middle
one should correspond to same bunch crossing as event that fired the trigger.
Need to adjust delay to trigger signal module-by-module to account for fibre lengths and time-of flight.
Timing scan performed once or twice per year to find optimum delays.
18+LHC/ATLAS operations LHC delivered:
40nb-1 in 2010 5fb-1 in 2011 22fb-1 in 2012
50ns bunch spacing.
High instantaneous lumi leads to up to 40 pp interactions per bunch crossing (μ). High detector occupancy.
Non-zero rate of Single Event Upsets (SEUs)
ATLAS Trigger system selects interesting events. “Level 1” hardware
trigger, rate ~70kHz Detector subsystems
must read out their data at this rate!
Software-based High Level Trigger further reduces rate to 400Hz for data recording.
19+SCT operations
SCT bias voltage maintained at safe 50V level until “Stable Beams” declared, at which point HV is ramped to 150V. Automated action
in 2012, though with shifter oversight.
Year 2010 2011 2012Lumi-weighted SCT good data fraction
99.9% 99.6%
99.1%
99% of readout channels operational.
Total Out-of-readout (Barrel/Endcap)
Fraction (Total)
Modules 4088 11/19 0.73%Chips 49056 38/11 0.10%Strips 627916
84111/8020 0.21% Automatic recovery (reconfiguration) of
modules giving persistent readout errors. Reconfiguration of all modules every 30
minutes during running (recover from SEUs).
20+Operational issues – ROD busy
If, for any reason, a ROD is not able to send data to ROS fast enough to keep up with L1 trigger rate, it can assert “BUSY”, which will stop triggers. Will then be automatically “stoplessly removed”, and can then
be recovered by a shifter action. If >1 SCT ROD is out of the readout at the same time, data is
considered “bad” for physics.
Running at high trigger rates and large occupancies uncovered a flaw in our ROD firmware, such that a ROD could go BUSY if a small number of modules returned no data, or too much data, or nonsensical data. Several issues in ROD firmware were identified and fixed, but
problem persisted until the end of the Run. High priority for detailed investigation during current shutdown!
21+Operational issues: CiS modules Approximately 25% of endcap modules manufactured by
CiS (remainder manufactured by Hamamatsu). Slightly different design:
In May 2012, we started observing strange behaviour from some of these modules. About 2hrs into high-lumi runs,
leakage current would increase dramatically, and one side of module would become noisy.
Eventually ROD would go BUSY, and/or module HV would trip.
Could be mitigated in the short term by reducing HV from 150V.
22+CiS modules
Mainly affected side 0 of “middle” modules.
Middle
Inner
Outer
Side 0Side 0
Side 0
Problem still not fully understood, but was mitigated for 2012 running by reducing “Standby” voltage (HV during inter-fill periods) from 50V to 5V for all CiS modules.
Current would still increase during run, but would plateau before reaching problematic levels.
23+Operational issues – TX failures TX channels (each corresponding to one module) on TX plugins
began failing in 2009. Analysis of failed units indicated ESD damage during manufacturing
process. Replacements ordered, with improved ESD precautions during
manufacturing, and installed.
After ~3 months operation, replacements also began to fail. Some evidence that humidity was damaging the units. Replacements ordered from a different vendor, with improved resistance
to humidity.
TX failures continued during 2012 run (though at a lower rate). Believed to be due to thermal mismatch between epoxy and VCSEL
array. Third full set of replacements, using a commercial VCSEL package
(LightABLE engine), will be installed during the current shutdown period. Minimal quantity of data was lost due to this problem, due to provision of TX
redundancy (modules can receive clock and command signals electronically from neighbouring modules), and vigilance of shifters and on-call experts.
24+Efficiency
Define intrinsic hit efficiency as “hits-per-possible-hit”, i.e. ignore non-operational modules from both numerator and denominator.
To measure efficiency of each module side, perform track fits ignoring that side, and then see if we see a hit.
Efficiency well above 99% for all layers+sides!
25+Noise Too many fake hits from noise could impair the pattern
recognition in tracking software. SCT was designed to have noise occupancy lower than
5x10-4. Occupancy can be measured
either in standalone calibration runs, or as part of normal ATLAS data-taking (look in empty bunch-crossings).
Noise is well within design limits. (At high luminosity, there are
many more hits from low pT tracks (inc “loopers”) than from intrinsic noise.)
26+Tracking performance Up to 4,000 tracks per event
in high-pile-up conditions seen in 2012!
Excellent agreement between data and Monte Carlo simulation.
27+Alignment Track-based alignment is iterative
process, where residuals (i.e. difference in position between hit-on-track and the track intersection point) are minimized. First align large structures, e.g. SCT
wrt TRT and Pixel, then eventually go down to individual modules.
Alignment in Barrel region, particularly in horizontal (“x”) direction, was already good in 2009 as a result of cosmics running..
SCT is now close to perfectly aligned.
28+Frequency Scanning Interferometer Laser-based alignment system, can precisely track
distances between nodes.
Large movements observed by FSI can alert track-based alignment team that a new set of constants may be needed.
29+Lorentz Angle In absence of B-field, charge
carriers produced by ionization of the silicon would be expected to travel in direction of E-field, i.e. perpendicular to surface of sensor.
Solenoidal B-field will deflect charge carries by some angle – Lorentz Angle.
Can be measured by looking at distribution of <cluster-size>-vs-incident angle
Charge carriers
Lorentz Angle
Incidence Angle
30+Radiation damage
FLUKA simulation indicates that dose received to date is still some way short of that required for type inversion.
Ionising radiation can have several effects on silicon sensors, including: Increased leakage current. Charge trapping / defects Transition from n-type to p-type.
31+Radiation damage
Radiation damage will increase the leakage current across the silicon sensors. This can be measured by the
Power Supplies, and compared to model predictions, as a function of dose received (using FLUKA and Hamburg/Dortmund model)
Excellent agreement for barrel modules over several orders of magnitude! Agreement in endcap is less
spectacular, but still within 20%. Radiation damage is not yet having a significant impact on the
operating characteristics of SCT modules. Dose received does not necessitate keeping SCT cold during
current shutdown period.
+For LHC Run 2, will need to handle μ~80, and trigger rate of 100 kHz!
SCT DAQ Bottlenecks
ABCDs:8-deep event buffer, read out at 40MbpsSufficient for 100kHz L1 with m~87
ROD/BOC pair x90
ROD
inpu
t dec
oder
and
FIF
O: 5
12
deep
S-link:32bit wide transfer at 40MHz = 1.28GbpsSufficient for 100kHz readout with m~30-40
ROS x8
+Expanded ROD System after shutdown
ABCDs:8-deep event buffer, read out at 40MbpsSufficient for 100kHz L1 with m~87
ROD/BOC pair x128
ROD
inpu
t dec
oder
and
FIF
O: 5
12
deep
S-link:32bit wide transfer at 40MHz = 1.28GbpsSufficient for 100kHz readout with m~87
1. 90128 RODs2. Improved data compression on
RODBandwidth matches that of front-end
ROS x12
34+Ongoing updates/improvements (cooling and TX transmitters) Evaporative cooling upgraded to use new thermosyphon
system. Use 90m drop
from surface to cavern to provide pressure.
No moving parts – is expected to be more reliable then current compressor-based system.
All TX plugins will be replaced again. Commercial VCSEL package, expected to be much more reliable.
35+Conclusions
The SCT has performed extremely well during LHC Run 1. 99% readout channels operational. >99% data “good” for physics analysis.
Efficiency and noise match or exceed design specifications, and effects of radiation damage are in good agreement with model predictions.
Updates to DAQ system, cooling, and optical transmitters under way during the current shutdown. Expect to have even more robust and reliable system for
LHC Run 2, able to deal with even higher occupancies and trigger rates.
36+
Backup
37+Calibration ROD can also generate triggers
for calibration. Several types of calibration, e.g.:
Opto scans – make sure communication between BOC and modules is working well.
Response Curve – ensure that hit threshold is correctly set to desired charge-on-strip.
Noise occupancy – send lots of triggers, count the hits, to ensure SCT modules are operating within design parameters for noise.
Typically have 1-2 hour periods available for calibration between LHC fills.
38+Readout
39+Location of CiS modules
40+Location of problematic CiS modules