ترغب بنشر مسار تعليمي؟ اضغط هنا

Application of the Cox Regression Model for Analysis of Railway Safety Performance

47   0   0.0 ( 0 )
 نشر من قبل Jens Braband
 تاريخ النشر 2020
  مجال البحث الاحصاء الرياضي
والبحث باللغة English




اسأل ChatGPT حول البحث

The assessment of in-service safety performance is an important task, not only in railways. For example it is important to identify deviations early, in particular possible deterioration of safety performance, so that corrective actions can be applied early. On the other hand the assessment should be fair and objective and rely on sound and proven statistical methods. A popular means for this task is trend analysis. This paper defines a model for trend analysis and compares different approaches, e. g. classical and Bayes approaches, on real data. The examples show that in particular for small sample sizes, e. g. when railway operators shall be assessed, the Bayesian prior may influence the results significantly.



قيم البحث

اقرأ أيضاً

Recently the European Union Agency for Railways (ERA) has received a mandate for the development of common safety methods for assessing the safety level and the safety performance of railway operators at national and Union level. Currently, several m ethods are under development. It is of interest how a possible candidate would behave and what would be the advantages and disadvantages of a particular method. In this paper, we study a version of the procedure. On the one hand side we analyze it based on the theory of mathematical statistics. As a result, we present a statistically efficient method the rate-ratio test based on a quantity that has smaller variance than the quantity handled by the ERA. Then, we support the theoretical results with the help of a simple simulation study in order to estimate failure probabilities of the first and second kinds. We construct such alternative distributions which the decision procedure cannot distinguish. We will show that the use of procedures that are optimal in the sense of mathematical statistics combined with the use of a characteristics that has small spread, here the number of accidents, is advantageous.
231 - Wenjie Jiang , Chang Lu , Jing Qu 2020
SKCM is the most dangerous one of skin cancer, its high degree of malignant, is the leading cause of skin cancer. And the level of radiation treatment and chemical treatment is minimal, so the mortality is high. Because of its complex molecular and c ellular heterogeneity, the existing prediction model of skin cancer risk is not ideal. In this study, we developed an immune-related lncRNAs model to predict the prognosis of patients with SKCM. Screening for SKCM-related differential expression of lncRNA from TCGA. Identified immune-related lncRNAs and lncRNA-related mRNA based on the co-expression method. Through univariate and multivariate analysis, an immune-related lncRNA model is established to analyze the prognosis of SKCM patients. A 4-lncRNA skin cancer prediction model was constructed, including MIR155HG, AL137003.2, AC011374.2, and AC009495.2. According to the model, SKCM samples were divided into a high-risk group and low-risk group, and predict the survival of the two groups in 30 years. The area under the ROC curve is 0.749, which shows that the model has excellent performance. We constructed a 4-lncRNA model to predict the prognosis of patients with SKCM, indicating that these lncRNAs may play a unique role in the carcinogenesis of SKCM.
Survival analysis is a challenging variation of regression modeling because of the presence of censoring, where the outcome measurement is only partially known, due to, for example, loss to follow up. Such problems come up frequently in medical appli cations, making survival analysis a key endeavor in biostatistics and machine learning for healthcare, with Cox regression models being amongst the most commonly employed models. We describe a new approach for survival analysis regression models, based on learning mixtures of Cox regressions to model individual survival distributions. We propose an approximation to the Expectation Maximization algorithm for this model that does hard assignments to mixture groups to make optimization efficient. In each group assignment, we fit the hazard ratios within each group using deep neural networks, and the baseline hazard for each mixture component non-parametrically. We perform experiments on multiple real world datasets, and look at the mortality rates of patients across ethnicity and gender. We emphasize the importance of calibration in healthcare settings and demonstrate that our approach outperforms classical and modern survival analysis baselines, both in terms of discriminative performance and calibration, with large gains in performance on the minority demographics.
Point processes in time have a wide range of applications that include the claims arrival process in insurance or the analysis of queues in operations research. Due to advances in technology, such samples of point processes are increasingly encounter ed. A key object of interest is the local intensity function. It has a straightforward interpretation that allows to understand and explore point process data. We consider functional approaches for point processes, where one has a sample of repeated realizations of the point process. This situation is inherently connected with Cox processes, where the intensity functions of the replications are modeled as random functions. Here we study a situation where one records covariates for each replication of the process, such as the daily temperature for bike rentals. For modeling point processes as responses with vector covariates as predictors we propose a novel regression approach for the intensity function that is intrinsically nonparametric. While the intensity function of a point process that is only observed once on a fixed domain cannot be identified, we show how covariates and repeated observations of the process can be utilized to make consistent estimation possible, and we also derive asymptotic rates of convergence without invoking parametric assumptions.
A rapid growth in spatial open datasets has led to a huge demand for regression approaches accommodating spatial and non-spatial effects in big data. Regression model selection is particularly important to stably estimate flexible regression models. However, conventional methods can be slow for large samples. Hence, we develop a fast and practical model-selection approach for spatial regression models, focusing on the selection of coefficient types that include constant, spatially varying, and non-spatially varying coefficients. A pre-processing approach, which replaces data matrices with small inner products through dimension reduction dramatically accelerates the computation speed of model selection. Numerical experiments show that our approach selects the model accurately and computationally efficiently, highlighting the importance of model selection in the spatial regression context. Then, the present approach is applied to open data to investigate local factors affecting crime in Japan. The results suggest that our approach is useful not only for selecting factors influencing crime risk but also for predicting crime events. This scalable model selection will be key to appropriately specifying flexible and large-scale spatial regression models in the era of big data. The developed model selection approach was implemented in the R package spmoran.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا