ترغب بنشر مسار تعليمي؟ اضغط هنا

Accuracy of the geometric-mean method for determining spatial resolutions of tracking detectors in the presence of multiple Coulomb scattering

118   0   0.0 ( 0 )
 نشر من قبل Aiwu Zhang
 تاريخ النشر 2016
  مجال البحث فيزياء
والبحث باللغة English




اسأل ChatGPT حول البحث

The geometric-mean method is often used to estimate the spatial resolution of a position-sensitive detector probed by tracks. It calculates the resolution solely from measured track data without using a detailed tracking simulation and without considering multiple Coulomb scattering effects. Two separate linear track fits are performed on the same data, one excluding and the other including the hit from the probed detector. The geometric mean of the widths of the corresponding exclusive and inclusive residual distributions for the probed detector is then taken as a measure of the intrinsic spatial resolution of the probed detector: $sigma=sqrt{sigma_{ex}cdotsigma_{in}}$. The validity of this method is examined for a range of resolutions with a stand-alone Geant4 Monte Carlo simulation that specifically takes multiple Coulomb scattering in the tracking detector materials into account. Using simulated as well as actual tracking data from a representative beam test scenario, we find that the geometric-mean method gives systematically inaccurate spatial resolution results. Good resolutions are estimated as poor and vice versa. The more the resolutions of reference detectors and probed detector differ, the larger the systematic bias. An attempt to correct this inaccuracy by statistically subtracting multiple-scattering effects from geometric-mean results leads to resolutions that are typically too optimistic by 10-50%. This supports an earlier critique of this method based on simulation studies that did not take multiple scattering into account.

قيم البحث

اقرأ أيضاً

The spatial resolution of a detector, using a reference detector telecscope, can be measured applying the geometric mean method, with tracks reconstructed from hits of all the detectors, including ($sigma_mathrm{in}$) and excluding ($sigma_mathrm{ex} $) the hit from the detector under study. The geometric mean of the two measured resolution values ($sigma=sqrt{sigma_mathrm{ex}sigma_mathrm{in}}$), is proposed to provide a more accurate estimate of the intrinsic detector resolution. This method has been tested using a Monte Carlo algorithm and is proven to give accurate results, independently of the distance between the detectors used for the track fitting. The method does not give meaningful results if all the detectors do not carry the same characteristics.
The GERDA and Majorana experiments will search for neutrinoless double-beta decay of germanium-76 using isotopically enriched high-purity germanium detectors. Although the experiments differ in conceptual design, they share many aspects in common, an d in particular will employ similar data analysis techniques. The collaborations are jointly developing a C++ software library, MGDO, which contains a set of data objects and interfaces to encapsulate, store and manage physical quantities of interest, such as waveforms and high-purity germanium detector geometries. These data objects define a common format for persistent data, whether it is generated by Monte Carlo simulations or an experimental apparatus, to reduce code duplication and to ease the exchange of information between detector systems. MGDO also includes general-purpose analysis tools that can be used for the processing of measured or simulated digital signals. The MGDO design is based on the Object-Oriented programming paradigm and is very flexible, allowing for easy extension and customization of the components. The tools provided by the MGDO libraries are used by both GERDA and Majorana.
96 - N. J. Ayres , G. Ban , G. Bison 2019
Psychological bias towards, or away from, a prior measurement or a theory prediction is an intrinsic threat to any data analysis. While various methods can be used to avoid the bias, e.g. actively not looking at the result, only data blinding is a tr aceable and thus trustworthy method to circumvent the bias and to convince a public audience that there is not even an accidental psychological bias. Data blinding is nowadays a standard practice in particle physics, but it is particularly difficult for experiments searching for the neutron electric dipole moment, as several cross measurements, in particular of the magnetic field, create a self-consistent network into which it is hard to inject a fake signal. We present an algorithm that modifies the data without influencing the experiment. Results of an automated analysis of the data are used to change the recorded spin state of a few neutrons of each measurement cycle. The flexible algorithm is applied twice to the data, to provide different data to various analysis teams. This gives us the option to sequentially apply various blinding offsets for separate analysis steps with independent teams. The subtle modification of the data allows us to modify the algorithm and to produce a re-blinded data set without revealing the blinding secret. The method was designed for the 2015/2016 measurement campaign of the nEDM experiment at the Paul Scherrer Institute. However, it can be re-used with minor modification for the follow-up experiment n2EDM, and may be suitable for comparable efforts.
The measurement of muon energy is critical for many analyses in large Cherenkov detectors, particularly those that involve separating extraterrestrial neutrinos from the atmospheric neutrino background. Muon energy has traditionally been determined b y measuring the specific energy loss (dE/dx) along the muons path and relating the dE/dx to the muon energy. Because high-energy muons (E_mu > 1 TeV) lose energy randomly, the spread in dE/dx values is quite large, leading to a typical energy resolution of 0.29 in log10(E_mu) for a muon observed over a 1 km path length in the IceCube detector. In this paper, we present an improved method that uses a truncated mean and other techniques to determine the muon energy. The muon track is divided into separate segments with individual dE/dx values. The elimination of segments with the highest dE/dx results in an overall dE/dx that is more closely correlated to the muon energy. This method results in an energy resolution of 0.22 in log10(E_mu), which gives a 26% improvement. This technique is applicable to any large water or ice detector and potentially to large scintillator or liquid argon detectors.
GELATIO is a new software framework for advanced data analysis and digital signal processing developed for the GERDA neutrinoless double beta decay experiment. The framework is tailored to handle the full analysis flow of signals recorded by high pur ity Ge detectors and photo-multipliers from the veto counters. It is designed to support a multi-channel modular and flexible analysis, widely customizable by the user either via human-readable initialization files or via a graphical interface. The framework organizes the data into a multi-level structure, from the raw data up to the condensed analysis parameters, and includes tools and utilities to handle the data stream between the different levels. GELATIO is implemented in C++. It relies upon ROOT and its extension TAM, which provides compatibility with PROOF, enabling the software to run in parallel on clusters of computers or many-core machines. It was tested on different platforms and benchmarked in several GERDA-related applications. A stable version is presently available for the GERDA Collaboration and it is used to provide the reference analysis of the experiment data.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا