Trackml Report
Trackml Report
Project Report
Sharad Chitlangia
ID No.: 2017A8PS0472G
Acknowledgements
I would like to thank Dr. Kinjal Banerjee for giving me the opportunity to work on this
wonderful project, and for providing constant guidance throughout the project duration. The
project involved some concepts from particle physics which I didn’t understand until I was given
direction by him. He also encouraged me to make submissions for the competition which led to a
Contents
1. Introduction
3.1. Data
3.4. Benchmarks
4. Outlook
1
The competition page is at https://competitions.codalab.org/competitions/20112. My final rank was 6th on the
official challenge and 1st in the private competition. This can be found under the ‘Development’ tab in Results tab.
2
The code for all the algorithms can be found at https://github.com/Sharad24/Particle-Track-Reconstruction
4
Chapter 1
Introduction
Particle Physics is a vast field in Physics where elementary particles, their interactions, and
the theories are studied in great detail. Experimental Particle Physics attempts to verify the
results that theoretical Particle Physics predicts, through very advanced and expensive
technology. The most famous of these is the Large Hadron Collider at CERN, Geneva, where the
major discoveries of particle physics occurred.
Machine Learning is a very important and upcoming field of Computer Science which uses
complex algorithms to teach computers how to learn patterns in data and predict outcomes for
events based on some experience. This area of computer science is extremely powerful and is
nowadays used everywhere, like in recommender systems, handwriting recognition, weather
prediction, gadget customization and so on. There are several subfields of machine learning like
deep learning, reinforcement learning etc., each introduced to solve different kinds of problems.
This report is an effort to understand the use of Machine Learning techniques in Particle
Physics. Experimental Particle Physics annually analyses several terabytes of data, and machine
learning plays a big role in this analysis. To understand this better, this report looks at how
machine learning helps improve predictions and saves a lot of effort from the physics community
annually to devise smarter and smarter ways to get the results they want.
5
Chapter 2
Colliders like the Large Hadron Collider at CERN, Tevatron at Fermilab, DESY at
Hamburg all aim to solve unanswered questions in Physics by extracting useful information from
collisions of beams of particles at great energies. These collisions give rise to more particles and
radiation that we may already have guessed, or ones we do not expect to find. The goal of
Particle physics in colliders today, is to investigate processes that will help find new particles or
new interactions between particles, to solve some mysteries plaguing Physics.
Theorists work on hypothesizing and predicting what interactions particles must have with
each other, and what might be the outcome of collisions between particles. This must then be
verified in accelerators where particles are actually collided with each other, using very long
accelerator tubes, so that at very high energies, colliding beams may bring the particles close
enough to experience interactions that are not often seen in daily life, like weak and strong
interactions. For example, ATLAS is one of the accelerators at CERN that collides protons with
each other. For the longest time, it was not confirmed that protons had substructure, but by
colliding very high speed electrons with protons, and later protons with protons, many
elementary particles like quarks, Higgs Boson were discovered. These particles are extremely
unstable and are not directly detected, analysis of data from colliders about the final products
helps to back track and tell which particles might have decayed to them.
The process of colliding an electron or muon with a proton to probe its constituents like
quarks, gluons and other particles is called deep inelastic scattering. Quarks and gluons interact
with each other to produce new quarks, and gluons interact with each other to produce particles,
6
some of which may be unknown and can help probe new physics. The final products have hit the
detectors and their momenta, velocity, charge and angles are measured. This helps to identify
which particles have been produced, and given their momenta and angular distribution, which
particles might have decayed to them.
The 3 dimensional momentum i s measured by the first layer of the detector, called an inner
tracker. It has very strong magnets that create a magnetic field strong enough to curve the paths
of particles based on their momenta. The direction of the curve (left or rightward bending) is
decided by the charge o f the particle. Since these particles are moving relativistically, high
momentum particles curve less and low momentum particles curve more.
Particle identification happens by a combination of velocity and momentum measurements.
While momentum is measured in a tracking chamber, velocity i s measured in various ways. In a
Time of Flight detector, the time required to travel from the interaction point to the detector, or
the time required to travel between two time of flight detectors is measured, and velocity is
calculated taking into account other resistive forces. Time of flight decreases as particles move
closer to speed of light.
Reactions in which the intermediate products are those of interest to us, we call signal. The
other reactions that produce the same final particles are called background. The physics
community devises several features to identify the difference between these two reactions,
simply by observations made in detectors. Some features are directly measured- called low-level
features, and some features are linear combinations, or derivable from the low level features,
called high-level features.
7
In the LHC, proton bunches (beams) circulates and collide at high energy. Each collision
produces a firework of new particles. To identify the types and measure the kinematic properties
of these particles, a complex apparatus, the detector bends them in an externally applied
magnetic field and records the small energy deposited by the particles when they impact
well-defined locations in the detector.
The tracking problem refers to reconstructing the trajectories from the information recorded
by the detector. Given the coordinates of the impact of particles on the detector (3D points), the
problem is to “connect the dots” or rather the points, i.e. return all sets of points belonging to
alleged particles trajectories.
8
Chapter 3
As the detectors at LHC, Cern become more stronger, the amount of data generated becomes
higher. Due to some natural errors like Noise, Lack of constant magnetic fields and secondary
trajectories, combinatorial methods cannot be applied. Machine Learning models are robust to
small noise and are great at detecting patterns in a large amount of data.
3.1 Data
Event hits
● hit id: a numerical hit identifier, unique inside the event.
● x, y, z : the reconstructed position (x, y, z)
● volume id,layer id, module id: the module where the hit was read out by using the
identification scheme
Event Particles
● hit id: numerical identifier of the hit as defined in the hits file.
9
● particle id: numerical identifier of the particle generating this hit as defined in
the particles file.
● tx, ty, tz: true hit position (in mm) at the intersection with the module. Remember that the
true hit position is the intersection of the trajectory with the module mid-plane.
● tpx, tpy, tpz true particle momentum (in GeV/c) in the global coordinate system at the
intersection point. The corresponding unit vector is tangent to the particle trajectory.
● weight: per-hit weight used by the scoring function; note that the total sum of weights
within one event equals to one.
● hit id: numerical identifier of the hit as defined in the hits file.
● ch0, ch1: cell identifier/coordinates unique with one module, encoded as in figure 6.
Depending on the detector type only one of the channel identifiers may be valid.
● value: Signal value information, e.g. how much charge a particle has deposited.
Approach
I developed a general framework to be followed for training algorithms. This framework was
largely motivated by some of the algorithms exposed in the first phase (the Accuracy phase) of
the competition on Kaggle. The framework is described below:
Preprocessing
The input features considered for pair classification were the hit locations, cell and signal values.
The hit location is in cartesian geometry. The cartesian coordinates were converted into
cylindrical as it simplifies the task for classification for the classifier. This is primarily because a
helix in cylindrical coordinates is like a straight line in cartesian coordinates.
The task is develop a method for identification of Promising pairs and triplets. This is plausible
as two hits cannot be in two very different sections of the detector if they are part of a trajectory.
A multitude of algorithms were tested for pair classification. They are described below:
● Logistic Regression: In its basic form uses a logistic function to model a binary
dependent variable. The logistic function is described as:
1
f (x) = 1+e
−x
where,
f 1(.) is the logistic function as described above.
The test accuracy achieved was 75.3%.
● Artificial Neural Network: An ANN is based on a collection of connected units or nodes
called artificial neurons, which loosely model the neurons in a biological brain. Each
connection, like the synapses in a biological brain, can transmit a signal from one
artificial neuron to another. An artificial neuron that receives a signal can process it and
then signal additional artificial neurons connected to it. Each node represents what is
called a perceptron. The perceptron equation is described as:
11
f (x) = w · x + b
where,
w represents the weight for that node
b represents the bias for the node
Essentially a neural network is a combination of such perceptrons, which makes it such a
powerful function approximator. The ANN is trained using the Backpropagation
algorithm. For the purpose of this challenge a network comprising of five layers was
used. The number of neurons in each layer, in sequence, was 4000, 2000, 2000, 1000, 1
(Figure 1).
Figure 1
The test accuracy achieved with this Neural Network is 84.3 % for pair classification.
Overall test track reconstruction accuracy was 80.4 %.
● Random Forest: They are an ensemble learning method for classification, regression and
other tasks that operates by constructing a multitude of decision trees at training time and
outputting the class. The prediction is made by a majority vote among the trees. A
random forest consisting of 500 trees with a depth of 10 was trained on particular event
data. The resultant trained Random Forest gave an accuracy of 94.6 % ontest data.
3.4 Benchmarks
2
The total inference time includes the time for pair prediction and the time for trajectory reconstruction
1
ε signifies negligible time. Less than a second
14
Chapter 4
Outlook
This report discusses the applications of Machine Learning and Deep Learning to one of the
tasks at CERN’s experiments. A faster algorithm can be built in C++ for faster inference using
CERN’s ROOT library. An algorithm which also organises the event data in a directed Acyclic
Graph should work better. Another easy extension would be to train an algorithm to predict
triples from pairs since as of now I have just extended the trajectory using 3d geometry.
15
Appendix A
The reports for the top algorithms can be found in the drive folder:
https://drive.google.com/open?id=1cnp8oHTaS8Nx33cKPwCzsPea4wAloKd7
16
Bibliography
[1] The TrackML challenge https://sites.google.com/site/trackmlparticle/