We present a practical approach to capturing ear-to-ear face models comprising both 3D meshes and intrinsic textures (i.e. diffuse and specular albedo). Our approach is a hybrid of geometric and photometric methods and requires no geometric calibration. Photometric measurements made in a lightstage are used to estimate view dependent high resolution normal maps. We overcome the problem of having a single photometric viewpoint by capturing in multiple poses. We use uncalibrated multiview stereo to estimate a coarse base mesh to which the photometric views are registered. We propose a novel approach to robustly stitching surface normal and intrinsic texture data into a seamless, complete and highly detailed face model. The resulting relightable models provide photorealistic renderings in any view.
Since the insertion area in the middle ear or in the sinus cavity is very narrow, the mobility of the endoscope is reduced to a rotation around a virtual point and a translation for the insertion of the camera. This article first presents the anatomy of these regions obtained from 3D scanning and then a mechanism based on the architecture of the agile eye coupled to a double parallelogram to create an RCM. This mechanism coupled with a positioning mechanism is used to handle an endoscope. This tool is used in parallel to the surgeon to allow him to have better rendering of the medium ear than the use of Binocular scope. The mechanism offers a wide working space without singularity whose borders are fixed by joint limits. This feature allows ergonomic positioning of the patients head on the bed as well as for the surgeon and allows other applications such as sinus surgery.
Classification of cognitive workload promises immense benefit in diverse areas ranging from driver safety to augmenting human capability through closed loop brain computer interface. The brain is the most metabolically active organ in the body and increases its metabolic activity and thus oxygen consumption with increasing cognitive demand. In this study, we explore the feasibility of in-ear SpO2 cognitive workload tracking. To this end, we preform cognitive workload assessment in 8 subjects, based on an N-back task, whereby the subjects are asked to count and remember the number of odd numbers displayed on a screen in 5 second windows. The 2 and 3-back tasks lead to either the lowest median absolute SpO2 or largest median decrease in SpO2 in all of the subjects, indicating a robust and measurable decrease in blood oxygen in response to increased cognitive workload. Using features derived from in-ear pulse oximetry, including SpO2, pulse rate and respiration rate, we were able to classify the 4 N-back task categories, over 5 second epochs, with a mean accuracy of 94.2%. Moreover, out of 21 total features, the 9 most important features for classification accuracy were all SpO2 related features. The findings suggest that in-ear SpO2 measurements provide valuable information for classification of cognitive workload over short time windows, which together with the small form factor promises a new avenue for real time cognitive workload tracking.
The precise detection of blood vessels in retinal images is crucial to the early diagnosis of the retinal vascular diseases, e.g., diabetic, hypertensive and solar retinopathies. Existing works often fail in predicting the abnormal areas, e.g, sudden brighter and darker areas and are inclined to predict a pixel to background due to the significant class imbalance, leading to high accuracy and specificity while low sensitivity. To that end, we propose a novel error attention refining network (ERA-Net) that is capable of learning and predicting the potential false predictions in a two-stage manner for effective retinal vessel segmentation. The proposed ERA-Net in the refine stage drives the model to focus on and refine the segmentation errors produced in the initial training stage. To achieve this, unlike most previous attention approaches that run in an unsupervised manner, we introduce a novel error attention mechanism which considers the differences between the ground truth and the initial segmentation masks as the ground truth to supervise the attention map learning. Experimental results demonstrate that our method achieves state-of-the-art performance on two common retinal blood vessel datasets.
The detection of sound begins when energy derived from acoustic stimuli deflects the hair bundles atop hair cells. As hair bundles move, the viscous friction between stereocilia and the surrounding liquid poses a fundamental challenge to the ears high sensitivity and sharp frequency selectivity. Part of the solution to this problem lies in the active process that uses energy for frequency-selective sound amplification. Here we demonstrate that a complementary part involves the fluid-structure interaction between the liquid within the hair bundle and the stereocilia. Using force measurement on a dynamically scaled model, finite-element analysis, analytical estimation of hydrodynamic forces, stochastic simulation and high-resolution interferometric measurement of hair bundles, we characterize the origin and magnitude of the forces between individual stereocilia during small hair-bundle deflections. We find that the close apposition of stereocilia effectively immobilizes the liquid between them, which reduces the drag and suppresses the relative squeezing but not the sliding mode of stereociliary motion. The obliquely oriented tip links couple the mechanotransduction channels to this least dissipative coherent mode, whereas the elastic horizontal top connectors stabilize the structure, further reducing the drag. As measured from the distortion products associated with channel gating at physiological stimulation amplitudes of tens of nanometres, the balance of forces in a hair bundle permits a relative mode of motion between adjacent stereocilia that encompasses only a fraction of a nanometre. A combination of high-resolution experiments and detailed numerical modelling of fluid-structure interactions reveals the physical principles behind the basic structural features of hair bundles and shows quantitatively how these organelles are adapted to the needs of sensitive mechanotransduction.
Myocardial Velocity Mapping Cardiac MR (MVM-CMR) can be used to measure global and regional myocardial velocities with proved reproducibility. Accurate left ventricle delineation is a prerequisite for robust and reproducible myocardial velocity estimation. Conventional manual segmentation on this dataset can be time-consuming and subjective, and an effective fully automated delineation method is highly in demand. By leveraging recently proposed deep learning-based semantic segmentation approaches, in this study, we propose a novel fully automated framework incorporating a 3D-UNet backbone architecture with Embedded multichannel Attention mechanism and LSTM based Recurrent neural networks (RNN) for the MVM-CMR datasets (dubbed 3D-EAR segmentor). The proposed method also utilises the amalgamation of magnitude and phase images as input to realise an information fusion of this multichannel dataset and exploring the correlations of temporal frames via the embedded RNN. By comparing the baseline model of 3D-UNet and ablation studies with and without embedded attentive LSTM modules and various loss functions, we can demonstrate that the proposed model has outperformed the state-of-the-art baseline models with significant improvement.