ترغب بنشر مسار تعليمي؟ اضغط هنا

Speeding up Particle Track Reconstruction using a Parallel Kalman Filter Algorithm

74   0   0.0 ( 0 )
 نشر من قبل Allison Hall
 تاريخ النشر 2020
  مجال البحث فيزياء
والبحث باللغة English
 تأليف Steven Lantz




اسأل ChatGPT حول البحث

One of the most computationally challenging problems expected for the High-Luminosity Large Hadron Collider (HL-LHC) is determining the trajectory of charged particles during event reconstruction. Algorithms used at the LHC today rely on Kalman filtering, which builds physical trajectories incrementally while incorporating material effects and error estimation. Recognizing the need for faster computational throughput, we have adapted Kalman-filter-based methods for highly parallel, many-core SIMD architectures that are now prevalent in high-performance hardware. In this paper, we discuss the design and performance of the improved tracking algorithm, referred to as mkFit. A key piece of the algorithm is the Matriplex library, containing dedicated code to optimally vectorize operations on small matrices. The physics performance of the mkFit algorithm is comparable to the nominal CMS tracking algorithm when reconstructing tracks from simulated proton-proton collisions within the CMS detector. We study the scaling of the algorithm as a function of the parallel resources utilized and find large speedups both from vectorization and multi-threading. mkFit achieves a speedup of a factor of 6 compared to the nominal algorithm when run in a single-threaded application within the CMS software framework.



قيم البحث

اقرأ أيضاً

62 - Giuseppe Cerati 2020
One of the most computationally challenging problems expected for the High-Luminosity Large Hadron Collider (HL-LHC) is finding and fitting particle tracks during event reconstruction. Algorithms used at the LHC today rely on Kalman filtering, which builds physical trajectories incrementally while incorporating material effects and error estimation. Recognizing the need for faster computational throughput, we have adapted Kalman-filter-based methods for highly parallel, many-core SIMD and SIMT architectures that are now prevalent in high-performance hardware. Previously we observed significant parallel speedups, with physics performance comparable to CMS standard tracking, on Intel Xeon, Intel Xeon Phi, and (to a limited extent) NVIDIA GPUs. While early tests were based on artificial events occurring inside an idealized barrel detector, we showed subsequently that our mkFit software builds tracks successfully from complex simulated events (including detector pileup) occurring inside a geometrically accurate representation of the CMS-2017 tracker. Here, we report on advances in both the computational and physics performance of mkFit, as well as progress toward integration with CMS production software. Recently we have improved the overall efficiency of the algorithm by preserving short track candidates at a relatively early stage rather than attempting to extend them over many layers. Moreover, mkFit formerly produced an excess of duplicate tracks; these are now explicitly removed in an additional processing step. We demonstrate that with these enhancements, mkFit becomes a suitable choice for the first iteration of CMS tracking, and eventually for later iterations as well. We plan to test this capability in the CMS High Level Trigger during Run 3 of the LHC, with an ultimate goal of using it in both the CMS HLT and offline reconstruction for the HL-LHC CMS tracker.
91 - Pierre Billoir 2021
We present an alternative implementation of the Kalman filter employed for track fitting within the LHCb experiment. It uses simple parametrizations for the extrapolation of particle trajectories in the field of the LHCb dipole magnet and for the eff ects of multiple scattering in the detector material. A speedup of more than a factor of four is achieved while maintaining the quality of the estimated track quantities. This Kalman filter implementation could be used in the purely software-based trigger of the LHCb upgrade.
We describe the Hybrid seeding, a standalone pattern recognition algorithm aiming at finding charged particle trajectories for the LHCb upgrade. A significant improvement to the charged particle reconstruction efficiency is accomplished by exploiting the knowledge of the LHCb magnetic field and the position of energy deposits in the scintillating fibre tracker detector. Moreover, we achieve a low fake rate and a small contribution to the overall timing budget of the LHCb real-time data processing.
Real-time data processing is one of the central processes of particle physics experiments which require large computing resources. The LHCb (Large Hadron Collider beauty) experiment will be upgraded to cope with a particle bunch collision rate of 30 million times per second, producing $10^9$ particles/s. 40 Tbits/s need to be processed in real-time to make filtering decisions to store data. This poses a computing challenge that requires exploration of modern hardware and software solutions. We present Compass, a particle tracking algorithm and a parallel raw input decoding optimised for GPUs. It is designed for highly parallel architectures, data-oriented and optimised for fast and localised data access. Our algorithm is configurable, and we explore the trade-off in computing and physics performance of various configurations. A CPU implementation that delivers the same physics performance as our GPU implementation is presented. We discuss the achieved physics performance and validate it with Monte Carlo simulated data. We show a computing performance analysis comparing consumer and server grade GPUs, and a CPU. We show the feasibility of using a full GPU decoding and particle tracking algorithm for high-throughput particle trajectories reconstruction, where our algorithm improves the throughput up to 7.4$times$ compared to the LHCb baseline.
Muons are the most abundant charged particles arriving at sea level originating from the decay of secondary charged pions and kaons. These secondary particles are created when high-energy cosmic rays hit the atmosphere interacting with air nuclei ini tiating cascades of secondary particles which led to the formation of extensive air showers (EAS). They carry essential information about the extra-terrestrial events and are characterized by large flux and varying angular distribution. To account for open questions and the origin of cosmic rays, one needs to study various components of cosmic rays with energy and arriving direction. Because of the close relation between muon and neutrino production, it is the most important particle to keep track of. We propose a novel tracking algorithm based on the Geometric Deep Learning approach using graphical structure to incorporate domain knowledge to track cosmic ray muons in our 3-D scintillator detector. The detector is modeled using the GEANT4 simulation package and EAS is simulated using CORSIKA (COsmic Ray SImulations for KAscade) with a focus on muons originating from EAS. We shed some light on the performance, robustness towards noise and double hits, limitations, and application of the proposed algorithm in tracking applications with the possibility to generalize to other detectors for astrophysical and collider experiments.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا