No Arabic abstract
We describe a novel application of the end-to-end deep learning technique to the task of discriminating top quark-initiated jets from those originating from the hadronization of a light quark or a gluon. The end-to-end deep learning technique combines deep learning algorithms and low-level detector representation of the high-energy collision event. In this study, we use low-level detector information from the simulated CMS Open Data samples to construct the top jet classifiers. To optimize classifier performance we progressively add low-level information from the CMS tracking detector, including pixel detector reconstructed hits and impact parameters, and demonstrate the value of additional tracking information even when no new spatial structures are added. Relying only on calorimeter energy deposits and reconstructed pixel detector hits, the end-to-end classifier achieves an AUC score of 0.975$pm$0.002 for the task of classifying boosted top quark jets. After adding derived track quantities, the classifier AUC score increases to 0.9824$pm$0.0013, serving as the first performance benchmark for these CMS Open Data samples. We additionally provide a timing performance comparison of different processor unit architectures for training the network.
Characterization of the electronic band structure of solid state materials is routinely performed using photoemission spectroscopy. Recent advancements in short-wavelength light sources and electron detectors give rise to multidimensional photoemission spectroscopy, allowing parallel measurements of the electron spectral function simultaneously in energy, two momentum components and additional physical parameters with single-event detection capability. Efficient processing of the photoelectron event streams at a rate of up to tens of megabytes per second will enable rapid band mapping for materials characterization. We describe an open-source workflow that allows user interaction with billion-count single-electron events in photoemission band mapping experiments, compatible with beamlines at $3^{text{rd}}$ and $4^{text{th}}$ generation light sources and table-top laser-based setups. The workflow offers an end-to-end recipe from distributed operations on single-event data to structured formats for downstream scientific tasks and storage to materials science database integration. Both the workflow and processed data can be archived for reuse, providing the infrastructure for documenting the provenance and lineage of photoemission data for future high-throughput experiments.
We study jet substructures of a boosted polarized top quark, which undergoes the semileptonic decay $tto bell u$, in the perturbative QCD framework. The jet mass distribution (energy profile) is factorized into the convolution of a hard top-quark decay kernel with the bottom-quark jet function (jet energy function). Computing the hard kernel to leading order in QCD and inputting the latter functions from the resummation formalism, we observe that the jet mass distribution is not sensitive to the helicity of the top quark, but the energy profile is: energy is accumulated faster within a left-handed top jet than within a right-handed one, a feature related to the $V-A$ structure of weak interaction. It is pointed out that the energy profile is a simple and useful jet observable for helicity discrimination of a boosted top quark, which helps identification of physics beyond the Standard Model at the Large Hadron Collider. The extension of our analysis to other jet substructures, including those associated with a hadronically decaying polarized top quark, is proposed.
We analyse the semileptonic decay of a polarised top-quark with a large velocity based on the perturbative QCD factorisation framework. Thanks to the factorisation and the spin decomposition, the production part and the decay part can be factorised and the spin dependence is introduced in the decay part. The decay part is converted to the top-jet function which describes the distribution of jet observables and the spin is translated to the helicity of the boosted top. Using this top-jet function, the energy profile of b-jet is investigated and it is turned out that the sub-jet energy for the helicity-minus top is accumulated faster than that for the helicity-plus top. This behaviour for the boosted top can be understood with the negative spin-analysing-power of b-quark in the polarised-top decay.
Audio classification can distinguish different kinds of sounds, which is helpful for intelligent applications in daily life. However, it remains a challenging task since the sound events in an audio clip is probably multiple, even overlapping. This paper introduces an end-to-end audio classification system based on raw waveforms and mix-training strategy. Compared to human-designed features which have been widely used in existing research, raw waveforms contain more complete information and are more appropriate for multi-label classification. Taking raw waveforms as input, our network consists of two variants of ResNet structure which can learn a discriminative representation. To explore the information in intermediate layers, a multi-level prediction with attention structure is applied in our model. Furthermore, we design a mix-training strategy to break the performance limitation caused by the amount of training data. Experiments show that the mean average precision of the proposed audio classification system on Audio Set dataset is 37.2%. Without using extra training data, our system exceeds the state-of-the-art multi-level attention model.
One of the core components of conventional (i.e., non-learned) video codecs consists of predicting a frame from a previously-decoded frame, by leveraging temporal correlations. In this paper, we propose an end-to-end learned system for compressing video frames. Instead of relying on pixel-space motion (as with optical flow), our system learns deep embeddings of frames and encodes their difference in latent space. At decoder-side, an attention mechanism is designed to attend to the latent space of frames to decide how different parts of the previous and current frame are combined to form the final predicted current frame. Spatially-varying channel allocation is achieved by using importance masks acting on the feature-channels. The model is trained to reduce the bitrate by minimizing a loss on importance maps and a loss on the probability output by a context model for arithmetic coding. In our experiments, we show that the proposed system achieves high compression rates and high objective visual quality as measured by MS-SSIM and PSNR. Furthermore, we provide ablation studies where we highlight the contribution of different components.