ترغب بنشر مسار تعليمي؟ اضغط هنا

Strategic Plan for a Scientific Software Innovation Institute (S2I2) for High Energy Physics

213   0   0.0 ( 0 )
 نشر من قبل Peter Elmer
 تاريخ النشر 2017
  مجال البحث فيزياء
والبحث باللغة English




اسأل ChatGPT حول البحث

The quest to understand the fundamental building blocks of nature and their interactions is one of the oldest and most ambitious of human scientific endeavors. Facilities such as CERNs Large Hadron Collider (LHC) represent a huge step forward in this quest. The discovery of the Higgs boson, the observation of exceedingly rare decays of B mesons, and stringent constraints on many viable theories of physics beyond the Standard Model (SM) demonstrate the great scientific value of the LHC physics program. The next phase of this global scientific project will be the High-Luminosity LHC (HL-LHC) which will collect data starting circa 2026 and continue into the 2030s. The primary science goal is to search for physics beyond the SM and, should it be discovered, to study its details and implications. During the HL-LHC era, the ATLAS and CMS experiments will record circa 10 times as much data from 100 times as many collisions as in LHC Run 1. The NSF and the DOE are planning large investments in detector upgrades so the HL-LHC can operate in this high-rate environment. A commensurate investment in R&D for the software for acquiring, managing, processing and analyzing HL-LHC data will be critical to maximize the return-on-investment in the upgraded accelerator and detectors. The strategic plan presented in this report is the result of a conceptualization process carried out to explore how a potential Scientific Software Innovation Institute (S2I2) for High Energy Physics (HEP) can play a key role in meeting HL-LHC challenges.

قيم البحث

اقرأ أيضاً

60 - Shangen Li 2021
We analyze a game of technology development where players allocate resources between exploration, which continuously expands the public domain of available technologies, and exploitation, which yields a flow payoff by adopting the explored technologi es. The qualities of the technologies are correlated and initially unknown, and this uncertainty is fully resolved once the technologies are explored. We consider Markov perfect equilibria with the quality difference between the best available technology and the latest technology under development as the state variable. In all such equilibria, while the players do not fully internalize the benefit of failure owing to free-riding incentives, they are more tolerant of failure than in the single-agent optimum thanks to an encouragement effect. In the unique symmetric equilibrium, the cost of exploration determines whether free-riding prevails as team size grows. Pareto improvements over the symmetric equilibrium can be achieved by asymmetric equilibria where players take turns performing exploration.
The scale of scientific High Performance Computing (HPC) and High Throughput Computing (HTC) has increased significantly in recent years, and is becoming sensitive to total energy use and cost. Energy-efficiency has thus become an important concern i n scientific fields such as High Energy Physics (HEP). There has been a growing interest in utilizing alternate architectures, such as low power ARM processors, to replace traditional Intel x86 architectures. Nevertheless, even though such solutions have been successfully used in mobile applications with low I/O and memory demands, it is unclear if they are suitable and more energy-efficient in the scientific computing environment. Furthermore, there is a lack of tools and experience to derive and compare power consumption between the architectures for various workloads, and eventually to support software optimizations for energy efficiency. To that end, we have performed several physical and software-based measurements of workloads from HEP applications running on ARM and Intel architectures, and compare their power consumption and performance. We leverage several profiling tools (both in hardware and software) to extract different characteristics of the power use. We report the results of these measurements and the experience gained in developing a set of measurement techniques and profiling tools to accurately assess the power consumption for scientific workloads.
Computing plays an essential role in all aspects of high energy physics. As computational technology evolves rapidly in new directions, and data throughput and volume continue to follow a steep trend-line, it is important for the HEP community to dev elop an effective response to a series of expected challenges. In order to help shape the desired response, the HEP Forum for Computational Excellence (HEP-FCE) initiated a roadmap planning activity with two key overlapping drivers -- 1) software effectiveness, and 2) infrastructure and expertise advancement. The HEP-FCE formed three working groups, 1) Applications Software, 2) Software Libraries and Tools, and 3) Systems (including systems software), to provide an overview of the current status of HEP computing and to present findings and opportunities for the desired HEP computational roadmap. The fin
We describe a laser-plasma platform for photon-photon collision experiments to measure fundamental quantum electrodynamic processes such as the linear Breit-Wheeler process with real photons. The platform has been developed using the Gemini laser fac ility at the Rutherford Appleton Laboratory. A laser wakefield accelerator and a bremsstrahlung convertor are used to generate a collimated beam of photons with energies of hundreds of MeV, that collide with keV x-ray photons generated by a laser heated plasma target. To detect the pairs generated by the photon-photon collisions, a magnetic transport system has been developed which directs the pairs onto scintillation-based and hybrid silicon pixel single particle detectors. We present commissioning results from an experimental campaign using this laser-plasma platform for photon-photon physics, demonstrating successful generation of both photon sources, characterisation of the magnetic transport system and calibration of the single particle detectors, and discuss the feasibility of this platform for the observation of the Breit-Wheeler process. The design of the platform will also serve as the basis for the investigation of strong-field quantum electrodynamic processes such as the nonlinear Breit-Wheeler and the Trident process, or eventually, photon-photon scattering.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا