No Arabic abstract
Concussion and repeated exposure to mild traumatic brain injury are risks for athletes in many sports. While direct head impacts are analyzed to improve the detection and awareness of head acceleration events so that an athletes brain health can be appropriately monitored and treated. However, head accelerations can also be induced by impacts with little or no head involvement. In this work we evaluated if impacts that do not involve direct head contact, such as being pushed in the torso, can be sufficient in collegiate American football to induce head accelerations comparable to direct head impacts. Datasets of impacts with and without direct head contact were collected and compared. These datasets were gathered using a state-of-the-art impact detection algorithm embedded in an instrumented mouthguard to record head kinematics. Video analysis was used to differentiate between impact types. In total, 15 impacts of each type were used in comparison, with clear video screenshots available to distinguish each impact type. Analysis of the kinematics showed that the impacts without direct head contact achieved similar levels of linear and angular accelerations during impact compared to those from direct head impacts. Finite element analyses using the median and peak kinematic signals were used to calculate maximum principal strain of the brain. Statistical analysis revealed that no significant difference was found between the two datasets based on a Bonferroni-adjusted p-value threshold of 0.017 , with the exception of peak linear acceleration. Impacts without direct head contact showed higher mean values of peak linear acceleration values of 17.6 g compared to the direct-head impact mean value of 6.1g. These results indicated that impacts other than direct head impacts could still produce meaningful kinematic loads in the head and as such should be included in athlete health monitoring.
Head-related impulse responses (HRIRs) are subject-dependent and direction-dependent filters used in spatial audio synthesis. They describe the scattering response of the head, torso, and pinnae of the subject. We propose a structural factorization of the HRIRs into a product of non-negative and Toeplitz matrices; the factorization is based on a novel extension of a non-negative matrix factorization algorithm. As a result, the HRIR becomes expressible as a convolution between a direction-independent emph{resonance} filter and a direction-dependent emph{reflection} filter. Further, the reflection filter can be made emph{sparse} with minimal HRIR distortion. The described factorization is shown to be applicable to the arbitrary source signal case and allows one to employ time-domain convolution at a computational cost lower than using convolution in the frequency domain.
Multi-head attention plays a crucial role in the recent success of Transformer models, which leads to consistent performance improvements over conventional attention in various applications. The popular belief is that this effectiveness stems from the ability of jointly attending multiple positions. In this paper, we first demonstrate that jointly attending multiple positions is not a unique feature of multi-head attention, as multi-layer single-head attention also attends multiple positions and is more effective. Then, we suggest the main advantage of the multi-head attention is the training stability, since it has less number of layers than the single-head attention, when attending the same number of positions. For example, 24-layer 16-head Transformer (BERT-large) and 384-layer single-head Transformer has the same total attention head number and roughly the same model size, while the multi-head one is significantly shallower. Meanwhile, we show that, with recent advances in deep learning, we can successfully stabilize the training of the 384-layer Transformer. As the training difficulty is no longer a bottleneck, substantially deeper single-head Transformer achieves consistent performance improvements without tuning hyper-parameters.
When people deliver a speech, they naturally move heads, and this rhythmic head motion conveys prosodic information. However, generating a lip-synced video while moving head naturally is challenging. While remarkably successful, existing works either generate still talkingface videos or rely on landmark/video frames as sparse/dense mapping guidance to generate head movements, which leads to unrealistic or uncontrollable video synthesis. To overcome the limitations, we propose a 3D-aware generative network along with a hybrid embedding module and a non-linear composition module. Through modeling the head motion and facial expressions1 explicitly, manipulating 3D animation carefully, and embedding reference images dynamically, our approach achieves controllable, photo-realistic, and temporally coherent talking-head videos with natural head movements. Thoughtful experiments on several standard benchmarks demonstrate that our method achieves significantly better results than the state-of-the-art methods in both quantitative and qualitative comparisons. The code is available on https://github.com/ lelechen63/Talking-head-Generation-with-Rhythmic-Head-Motion.
In this work we present a new physics-informed machine learning model that can be used to analyze kinematic data from an instrumented mouthguard and detect impacts to the head. Monitoring player impacts is vitally important to understanding and protecting from injuries like concussion. Typically, to analyze this data, a combination of video analysis and sensor data is used to ascertain the recorded events are true impacts and not false positives. In fact, due to the nature of using wearable devices in sports, false positives vastly outnumber the true positives. Yet, manual video analysis is time-consuming. This imbalance leads traditional machine learning approaches to exhibit poor performance in both detecting true positives and preventing false negatives. Here, we show that by simulating head impacts numerically using a standard Finite Element head-neck model, a large dataset of synthetic impacts can be created to augment the gathered, verified, impact data from mouthguards. This combined physics-informed machine learning impact detector reported improved performance on test datasets compared to traditional impact detectors with negative predictive value and positive predictive values of 88% and 87% respectively. Consequently, this model reported the best results to date for an impact detection algorithm for American Football, achieving an F1 score of 0.95. In addition, this physics-informed machine learning impact detector was able to accurately detect true and false impacts from a test dataset at a rate of 90% and 100% relative to a purely manual video analysis workflow. Saving over 12 hours of manual video analysis for a modest dataset, at an overall accuracy of 92%, these results indicate that this model could be used in place of, or alongside, traditional video analysis to allow for larger scale and more efficient impact detection in sports such as American Football.
Because of the relatively rigid coupling between the upper dentition and the skull, instrumented mouthguards have been shown to be a viable way of measuring head impact kinematics for assisting in understanding the underlying biomechanics of concussions. This has led various companies and institutions to further develop instrumented mouthguards. However, their use as a research tool for understanding concussive impacts makes quantification of their accuracy critical, especially given the conflicting results from various recent studies. Here we present a study that uses a pneumatic impactor to deliver impacts characteristic to football to a Hybrid III headform, in order to validate and compare five of the most commonly used instrumented mouthguards. We found that all tested mouthguards gave accurate measurements for the peak angular acceleration (mean relative error, MRE < 13%), the peak angular velocity (MRE < 8%), brain injury criteria values (MRE < 13%) and brain deformation (described as maximum principal strain and fiber strain, calculated by a convolutional neural network based brain model, MRE < 9%). Finally, we found that the accuracy of the measurement varies with the impact locations yet is not sensitive to the impact velocity for the most part.