We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings.
To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge.org
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
Operational Risk is one of the most difficult risks to model. It is a large and diverse category covering anything from cyber losses to mis-selling fines; and from processing errors to HR issues. Data is usually lacking, particularly for low frequency, high impact losses, and consequently there can be a heavy reliance on expert judgement. This paper seeks to help actuaries and other risk professionals tasked with the challenge of validating models of operational risks. It covers the loss distribution and scenario-based approaches most commonly used to model operational risks, as well as Bayesian Networks. It aims to give a comprehensive yet practical guide to how one may validate each of these and provide assurance that the model is appropriate for a firm’s operational risk profile.
When very young children are brought to a doctor or hospital with signs and symptoms consistent with head injury, it is important to determine the cause. For almost 50 years, the triad of subdural haematoma (SDH), retinal haemorrhage (RH) and encephalopathy has been regarded as an accurate predictor of deliberate shaking and widely used to diagnose shaken baby syndrome (SBS). Statistical analyses by Cardiff University researchers and others claim to show that certain combinations of findings are highly predictive of abuse and as a result of this conclusion, protocols such as mandatory reporting to police are invoked in the name of protecting the child. However, concerns have been raised about the circularity of approach used in the statistical analyses which requires each case to be classified explicitly as either abuse or non-abuse. By producing a causal model of the problem, we show that these findings are actually a poor predictor of SBS, even where there is some evidence of risk factors indicative of abuse.
Major depressive disorder (MDD) is one of the growing human mental health challenges facing the global health care system. In this study, the structural connectivity between symptoms of MDD is explored using two different network modeling approaches.
Methods
Data are from ‘the Virginia Adult Twin Study of Psychiatric and Substance Use Disorders (VATSPSUD)’. A cohort of N = 2163 American Caucasian female-female twins was assessed as part of the VATSPSUD study. MDD symptoms were assessed using personal structured clinical interviews. Two network analyses were conducted. First, an undirected network model was estimated to explore the connectivity between the MDD symptoms. Then, using a Bayesian network, we computed a directed acyclic graph (DAG) to investigate possible directional relationships between symptoms.
Results
Based on the results of the undirected network, the depressed mood symptom had the highest centrality value, indicating its importance in the overall network of MDD symptoms. Bayesian network analysis indicated that depressed mood emerged as a plausible driving symptom for activating other symptoms. These results are consistent with DSM-5 guidelines for MDD. Also, somatic weight and appetite symptoms appeared as the strongest connections in both networks.
Conclusions
We discuss how the findings of our study might help future research to detect clinically relevant symptoms and possible directional relationships between MDD symptoms defining major depression episodes, which would help identify potential tailored interventions. This is the first study to investigate the network structure of VATSPSUD data using both undirected and directed network models.
Reconstruction of gene interaction networks from experimental data provides a deep understanding of the underlying biological mechanisms. The noisy nature of the data and the large size of the network make this a very challenging task. Complex approaches handle the stochastic nature of the data but can only do this for small networks; simpler, linear models generate large networks but with less reliability.
Methods:
We propose a divide-and-conquer approach using probabilistic graph representations and external knowledge. We cluster the experimental data and learn an interaction network for each cluster, which are merged using the interaction network for the representative genes selected for each cluster.
Results:
We generated an interaction atlas for 337 human pathways yielding a network of 11,454 genes with 17,777 edges. Simulated gene expression data from this atlas formed the basis for reconstruction. Based on the area under the curve of the precision-recall curve, the proposed approach outperformed the baseline (random classifier) by ∼15-fold and conventional methods by ∼5–17-fold. The performance of the proposed workflow is significantly linked to the accuracy of the clustering step that tries to identify the modularity of the underlying biological mechanisms.
Conclusions:
We provide an interaction atlas generation workflow optimizing the algorithm/parameter selection. The proposed approach integrates external knowledge in the reconstruction of the interactome using probabilistic graphs. Network characterization and understanding long-range effects in interaction atlases provide means for comparative analysis with implications in biomarker discovery and therapeutic approaches. The proposed workflow is freely available at http://otulab.unl.edu/atlas.
Given a combinatorial search problem, it may be highly useful to enumerate its (all) solutions besides just finding one solution, or showing that none exists. The same can be stated about optimal solutions if an objective function is provided. This work goes beyond the bare enumeration of optimal solutions and addresses the computational task of solution enumeration by optimality (SEO). This task is studied in the context of answer set programming (ASP) where (optimal) solutions of a problem are captured with the answer sets of a logic program encoding the problem. Existing answer set solvers already support the enumeration of all (optimal) answer sets. However, in this work, we generalize the enumeration of optimal answer sets beyond strictly optimal ones, giving rise to the idea of answer set enumeration in the order of optimality (ASEO). This approach is applicable up to the best k answer sets or in an unlimited setting, which amounts to a process of sorting answer sets based on the objective function. As the main contribution of this work, we present the first general algorithms for the aforementioned tasks of answer set enumeration. Moreover, we illustrate the potential use cases of ASEO. First, we study how efficiently access to the next-best solutions can be achieved in a number of optimization problems that have been formalized and solved in ASP. Second, we show that ASEO provides us with an effective sampling technique for Bayesian networks.
This chapter is concerned with analysing the expected runtime of probabilistic programs by exploiting program verification techniques. We introduce a weakest pre-conditioning framework á la Dijkstra that enables to determine the expected runtime in a compositional manner. Like weakest pre-conditions, it is a reasoning framework at the syntax level of programs. Applications of the weakest pre-conditioning framework include determining the expected runtime of randomised algorithms, as well as determining whether a program is positive almost-surely terminating, i.e., whether the expected number of computation steps until termination is finite for every possible input. For Bayesian networks, a restricted class of probabilistic programs, we show that the expected runtime analysis can be fully automated. In this way, the simulation time under rejection sampling can be determined. This is particularly useful for ill-conditioned inference queries.
Certain hypotheses cannot be directly confirmed for theoretical, practical, or moral reasons. For some of these hypotheses, however, there might be a workaround: confirmation based on analogical reasoning. In this paper we take up Dardashti, Hartmann, Thébault, and Winsberg’s (2019) idea of analyzing confirmation based on analogical inference Bayesian style. We identify three types of confirmation by analogy and show that Dardashti et al.’s approach can cover two of them. We then highlight possible problems with their model as a general approach to analogical inference and argue that these problems can be avoided by supplementing Bayesian update with Jeffrey conditionalization.
The relation between the understanding and belief of the site-specific dangers of climate change and the behaviour that individuals take to mitigate their impacts was assessed to investigate the psychological antecedent to pro-environmental behaviour; a necessity to mitigate anthropogenic climate change at the individual level. A quantitative cross-sectional design was employed to measure beliefs and behaviour of university students. Correlation was measured between the belief in one’s ability to affect change and pro-environmental behaviour. The hypothesis that nations facing greater climate threat would behave accordingly was tested on the two largest national representatives of the sample, China and Australia. In addition, a naïve Bayesian network, coupled with a self-organising map, was developed to explore correlations between self-efficacy and participants’ socio-demographic features. Results showed that Chinese students are more likely to have higher self-efficacy, while such trend was not noticed for Australians. Similarly, participants with higher educational qualifications, older, and with higher paid jobs also have a higher chance of presenting pro-environmental behaviour. Despite the study limitations, there seems to be evidence suggesting that educational and climate change policies have affected students’ self-efficacy and individual commitment to mitigation.
Bayesian networks are convenient graphical expressions for high-dimensional probability distributions which represent complex relationships between a large number of random variables. They have been employed extensively in areas such as bioinformatics, artificial intelligence, diagnosis, and risk management. The recovery of the structure of a network from data is of prime importance for the purposes of modeling, analysis, and prediction. There has been a great deal of interest in recent years in the NP-hard problem of learning the structure of a Bayesian network from observed data. Typically, one assigns a score to various structures and the search becomes an optimization problem that can be approached with either deterministic or stochastic methods. In this paper, we introduce a new search strategy where one walks through the space of graphs by modeling the appearance and disappearance of edges as a birth and death process. We compare our novel approach with the popular Metropolis–Hastings search strategy and give empirical evidence that the birth and death process has superior mixing properties.
Risk aggregation is a popular method used to estimate the sum of a collection of financial assets or events, where each asset or event is modelled as a random variable. Applications include insurance, operational risk, stress testing and sensitivity analysis. In practice, the sum of a set of random variables involves the use of two well-known mathematical operations: n-fold convolution (for a fixed number n) and N-fold convolution, defined as the compound sum of a frequency distribution N and a severity distribution, where the number of constant n-fold convolutions is determined by N, where the severity and frequency variables are independent, and continuous, currently numerical solutions such as, Panjer’s recursion, fast Fourier transforms and Monte Carlo simulation produce acceptable results. However, they have not been designed to cope with new modelling challenges that require hybrid models containing discrete explanatory (regime switching) variables or where discrete and continuous variables are inter-dependent and may influence the severity and frequency in complex, non-linear, ways. This paper describes a Bayesian Factorisation and Elimination (BFE) algorithm that performs convolution on the hybrid models required to aggregate risk in the presence of causal dependencies. This algorithm exploits a number of advances from the field of Bayesian Networks, covering methods to approximate statistical and conditionally deterministic functions to factorise multivariate distributions for efficient computation. Experiments show that BFE is as accurate on conventional problems as competing methods. For more difficult hybrid problems BFE can provide a more general solution that the others cannot offer. In addition, the BFE approach can be easily extended to perform deconvolution for the purposes of stress testing and sensitivity analysis in a way that competing methods do not.
This chapter focuses on machine learning as a general way of thinking about the world, and provides a high-level characterization of the major goals of machine learning. Structural inference is the basis of many, and arguably most, machine learning frameworks and methods, including many well-known ones such as various forms of regression, neural-network learning algorithms such as back propagation, and causal learning algorithms using Bayesian networks. Machine learning algorithms must balance three factors: complexity of the learned model, which provides increased accuracy in representing the input dataset; generalizability of the learned model to new data, which enables the use of the model in novel contexts; and computational tractability of learning and using the model, which is a necessary precondition for the algorithms to have practical value. The practice of machine learning inevitably involves some human element to specify and control the algorithm, test various assumptions, and interpret the algorithm output.
This paper argues that most of the problems that actuaries have to deal with in the context of non-life insurance can be usefully cast in the framework of computational intelligence (a.k.a. artificial intelligence), the discipline that studies the design of agents which exhibit intelligent behaviour. Finding an adequate framework for actuarial problems has more than a simply theoretical interest: it also allows a technological transfer from the computational intelligence discipline to general insurance, wherever techniques have been developed for problems which are common to both contexts. This has already happened in the past (neural networks, clustering, data mining have all found applications to general insurance) but not in a systematic way. One of the objectives of this paper will therefore be to introduce some useful techniques such as sparsity-based regularisation and dynamic decision networks that are not yet known to the wider actuarial community.
Whilst in the first part of this paper we dealt mainly with data-driven loss modelling under the assumption that all the data were accurate and fully relevant to the exercise, in this second part of the paper we explore how to deal with uncertain knowledge, whether this uncertainty comes from the fact that the data are not fully reliable (e.g. they are estimates) or from the fact that the knowledge is “soft” (e.g. expert beliefs) or not fully relevant (e.g. market information on a given risk). Most importantly, we will deal with the problem of making pricing, reserving and capital decisions under uncertainty. It will be concluded that a Bayesian framework is the most adequate for dealing with uncertainty, and we will present a number of computational intelligence techniques to do this in practice.
The relative contributions of fat-free mass (FFM) and fat mass (FM) to body weight are key indicators for several major public health issues. Predictive models could offer new insights into body composition analysis. A non-parametric equation derived from a probabilistic Bayesian network (BN) was established by including sex, age, body weight and height. We hypothesised that it would be possible to assess the body composition of any subject from easily accessible covariables by selecting an adjusted FFM value within a reference dual-energy X-ray absorptiometry (DXA) measurement database (1999–2004 National Health and Nutrition Examination Survey (NHANES), n 10 402). FM was directly calculated as body weight minus FFM. A French DXA database (n 1140) was used (1) to adjust the model parameters (n 380) and (2) to cross-validate the model responses (n 760). French subjects were significantly different from American NHANES subjects with respect to age, weight and FM. Despite this different population context, BN prediction was highly reliable. Correlations between BN predictions and DXA measurements were significant for FFM (R2 0·94, P < 0·001, standard error of prediction (SEP) 2·82 kg) and the percentage of FM (FM%) (R2 0·81, P < 0·001, SEP 3·73 %). Two previously published linear models were applied to the subjects of the French database and compared with BN predictions. BN predictions were more accurate for both FFM and FM than those obtained from linear models. In addition, BN prediction generated stochastic variability in the FM% expressed in terms of BMI. The use of such predictions in large populations could be of interest for many public health issues.
Recommend this
Email your librarian or administrator to recommend adding this to your organisation's collection.