Events
Thu 16 Jan, '14- |
CRiSM Seminar - Chenlei Leng (Warwick), John Fox (Oxford & UCL/Royal Free Hospital)A1.01John Fox (Oxford & UCL/Royal Free Hospital) Despite the practical success of argumentation methods in risk management and other kinds of decision making the main theories ignore quantitative measurement of uncertainty, or they combine qualitative reasoning with quantitative uncertainty in ad hoc ways. After a brief introduction to argumentation theory I will demonstrate some medical applications and invite suggestions for ways of incorporating uncertainty probabilistically that are mathematically satisfactory. Chenlei Leng (Warwick)
|
|
Thu 30 Jan, '14- |
CRiSM Seminar - Judith Rousseau (Paris Dauphine), Jean-Michel Marin (Université Montpellier)A1.01Jean-Michel Marin Consistency of the Adaptive Multiple Importance Sampling (joint work with Pierre Pudlo and Mohammed Sedki Among Monte Carlo techniques, the importance sampling requires fine tuning of a proposal distribution, which is now fluently resolved through iterative schemes. The Adaptive Multiple Importance Sampling (AMIS) of Cornuet et al. (2012) provides a significant improvement in stability and Effective Sample Size due to the introduction of a recycling procedure. However, the consistency of the AMIS estimator remains largely open. In this work, we prove the convergence of the AMIS, at a cost of a slight modification in the learning process. Numerical experiments exhibit that this modification might even improve the original scheme. Judith Rousseau Asymptotic properties of Empirical Bayes procedures – in parametric and non parametric models
In this work we investigate frequentist properties of Empirical Bayes procedures. Empirical Bayes procedures are very much used in practice in more or less formalized ways as it is common practice to replace some hyperparameter in the prior by some data dependent quantity. There are typically two ways of constructing these data dependent quantities : using some king of moment estimator or some quantity whose behaviour is well understood or using a maximum marginal likelihood estimator. In this work we first give some general results on how to determine posterior concentration rates under the former setting, which we apply in particular to two types of Dirichlet process mixtures. We then shall discuss more parametric models in the context of maximum marginal likelihood estimation. We will in particular explain why some pathological behaviour can be expected in this case. |
|
Thu 13 Feb, '14- |
CRiSM Seminar - Amanda Turner (Lancaster)A1.01Amanda Turner (Lancaster) Small particle limits in a regularized Laplacian random growth model
|
|
Thu 13 Feb, '14- |
CRiSM Seminar - Vasileios Maroulas (Bath/Tennessee))A1.01Vasileios Maroulas (Bath/Tennessee)
|
|
Thu 13 Mar, '14- |
CRiSM Seminar - Darren Wilkinson (Newcastle), Richard Everitt (Reading)A1.01Darren Wilkinson (Newcastle) Saccharomyces cerevisiae (often known as budding yeast, or brewers yeast) is a single-celled micro-organism that is easy to grow and genetically manipulate. As it has a cellular organisation that has much in common with the cells of humans, it is often used as a model organism for studying genetics. High-throughput robotic genetic technologies can be used to study the fitness of many thousands of genetic mutant strains of yeast, and the resulting data can be used to identify novel genetic interactions relevant to a target area of biology. The processed data consists of tens of thousands of growth curves with a complex hierarchical structure requiring sophisticated statistical modelling of genetic independence, genetic interaction (epistasis), and variation at multiple levels of the hierarchy. Starting from simple stochastic differential equation (SDE) modelling of individual growth curves, a Bayesian hierarchical model can be built with variable selection indicators for inferring genetic interaction. The methods will be applied to data from experiments designed to highlight genetic interactions relevant to telomere biology. Richard Everitt (Reading) Inexact approximations for doubly and triply intractable problems Markov random field models are used widely in computer science, statistical physics and spatial statistics and network analysis. However, Bayesian analysis of these models using standard Monte Carlo methods is not possible due to an intractable likelihood function. Several methods have been developed that permit exact, or close to exact, simulation from the posterior distribution. However, estimating the marginal likelihood and Bayes' factors for these models remains challenging in general. This talk will describe new methods for estimating Bayes' factors that use simulation to circumvent the evaluation of the intractable likelihood, and compare them to approximate Bayesian computation. We will also discuss more generally the idea of "inexact approximations". |
|
Thu 27 Mar, '14- |
CRiSM Seminar - Professor Adrian Raftery (Washington)A1.01Professor Adrian Raftery (Washington) Bayesian Reconstruction of Past Populations for Developing and Developed Countries |
|
Thu 1 May, '14- |
Oxford-Warwick Seminar: David Dunson (Duke) and Eric Moulines (Télécom ParisTech)MS.03David Dunson (Duke University) Robust and scalable Bayes via the median posterior Bayesian methods have great promise in big data sets, but this promise has not been fully realized due to the lack of scalable computational methods. Usual MCMC and SMC algorithms bog down as the size of the data and number of parameters increase. For massive data sets, it has become routine to rely on penalized optimization approaches implemented on distributed computing systems. The most popular scalable approximation algorithms rely on variational Bayes, which lacks theoretical guarantees and badly under-estimates posterior covariance. Another problem with Bayesian inference is the lack of robustness; data contamination and corruption is particularly common in large data applications and cannot easily be dealt with using traditional methods. We propose to solve both the robustness and the scalability problem using a new alternative to exact Bayesian inference we refer to as the median posterior. Data are divided into subsets and stored on different computers prior to analysis. For each subset, we obtain a stochastic approximation to the full data posterior, and run MCMC to generate samples from this approximation. The median posterior is defined as the geometric median of the subset-specific approximations, and can be rapidly approximated. We show several strong theoretical results for the median posterior, including general theorems on concentration rates and robustness. The methods are illustrated through simple examples, including Gaussian process regression with outliers. Eric Moulines (Télécom ParisTech) Proximal Metropolis adjusted Langevin algorithm for sampling sparse distribution over high-dimensional spaces This talk introduces a new Markov Chain Monte Carlo method to sampling sparse distributions or to perform Bayesian model choices in high dimensional settings. The algorithm is a Hastings-Metropolis sampler with a proposal mechanism which combines (i) a Metropolis adjusted Langevin step to propose local moves associated with the differentiable part of the target density with (ii) a proximal step based on the non-differentiable part of the target density which provides sparse solutions such that small components are shrunk toward zero. Several implementations of the proximal step will be investigated adapted to different sparsity priors or allowing to perform variable selections, in high-dimensional settings. The performance of these new procedures are illustrated on both simulated and real data sets. Preliminary convergence results will also be presented. |
|
Thu 15 May, '14- |
CRiSM Seminar - Mark Fiecas (Warwick)A1.01Mark Fiecas (Warwick) In recent years, research into analyzing brain signals has dramatically increased, and these rich data sets require more advanced statistical tools in order to perform proper statistical analyses. Consider an experiment where a stimulus is presented many times, and after each stimulus presentation (trial), time series data is collected. The time series data per trial exhibit nonstationary characteristics. Moreover, across trials the time series are non-identical because their spectral properties change over the course of the experiment. In this talk, we will look at a novel approach for analyzing nonidentical nonstationary time series data. We consider two sources of nonstationarity: 1) within each trial of the experiment and 2) across the trials, so that the spectral properties of the time series data are evolving over time within a trial, and are also evolving over the course of the experiment. We extend the locally stationary time series model to account for nonidentical data. We analyze a local field potential data set to study how the spectral properties of the local field potentials obtained from the nucleus accumbens and the hippocampus of a monkey evolve over the course of a learning association experiment. |
|
Thu 15 May, '14- |
CRiSM Seminar - David Leslie (Bristol)A1.01David Leslie (Bristol) Stochastic approximation was introduced as a tool to find the zeroes of a function under only noisy observations of the function value. A classical statistical example is to find the zeroes of the score function when observations can only be processed sequentially. The method has since been developed and used mainly in the control theory, machine learning and economics literature to analyse iterative learning algorithms, but I contend that it is time for statistics to re-discover the power of stochastic approximation. I will introduce the main ideas of the method, and describe an extension; the parameter of interest is an element of a function space, and we wish to analyse its stochastic evolution through time. This extension allows the analysis of online nonparametric algorithms - we present an analysis of Newton's algorithm to estimate nonparametric mixing distributions. It also allows the investigation of learning in games with a continuous strategy set, where a mixed strategy is an arbitrary distribution on an interval. (Joint work with Steven Perkins) |
|
Thu 29 May, '14- |
CRiSM Seminar - Rajen Shah (Cambridge)A1.01Rajen Shah (Cambridge) Random Intersection Trees for finding interactions in large, sparse datasets Many large-scale datasets are characterised by a large number (possibly tens of thousands or millions) of sparse variables. Examples range from medical insurance data to text analysis. While estimating main effects in regression problems involving such data is now a reasonably well-studied problem, finding interactions between variables remains a serious computational challenge. As brute force searches through all possible interactions are infeasible, most approaches build up interaction sets incrementally, adding variables in a greedy fashion. The drawback is that potentially informative high-order interactions may be overlooked. Here, we propose an alternative approach for classification problems with binary predictor variables, called Random Intersection Trees. It works by starting with a maximal interaction that includes all variables, and then gradually removing variables if they fail to appear in randomly chosen observations of a class of interest. We show that with this method, under some weak assumptions, interactions can be found with high probability, and that the computational complexity of our procedure is much smaller than for a brute force search. |
|
Thu 29 May, '14- |
CRiSM Seminar - Randal DoucA1.01Randal Douc (TELECOM SudParis) Identifiability conditions for partially-observed Markov chains By R. Douc, F. Roueff and T. Sim This paper deals with a parametrized family of partially-observed bivariate Markov chains. We establish that the limit of the normalized log-likelihood is maximized when the parameter belongs to the equivalence class of the true parameter, which is a key feature for obtaining consistency the Maximum Likelihood Estimators (MLE) in well-specified models. This result is obtained in a general framework including both fully dominated or partially dominated models, and thus applies to both Hidden Markov models or Observation-Driven times series. In contrast with previous approaches, the identifiability is addressed by relying on the unicity of the invariant distribution of the Markov chain associated to the complete data, regardless its rate of convergence to the equilibrium. We use this approach to obtain a set of easy-to-check conditions which imply the consistency of the MLE of a general observation driven time series. |
|
Thu 12 Jun, '14- |
CRiSM Seminar - Ben Graham (Warwick)A1.01Ben Graham (University of Warwick) Handwriting, signatures, and convolutions The 'signature', from the theory of differential equations driven by rough paths, provides a very efficient way of characterizing curves. From a machine learning perspective, the elements of the signature can be used as a set of features for consumption by a classification algorithm. Using datasets of letters, digits, Indian characters and Chinese characters, we see that this improves the accuracy of online character recognition---that is the task of reading characters represented as a collection of pen strokes. |
|
Thu 12 Jun, '14- |
CRiSM Seminar - Emmanuele Giorgi (Lancaster)Emmanuele Giorgi (Lancaster) Combining data from multiple spatially referenced prevalence surveys using generalized linear geostatistical models Geostatistical methods are becoming more widely used in epidemiology to analyze spatial variation in disease prevalence. These methods are especially useful in resource-poor settings where disease registries are either non-existent or geographically incomplete, and data on prevalence must be obtained by survey sampling of the population of interest. In order to obtain good geographical coverage of the population, it is often necessary also to combine information from multiple prevalence surveys in order to estimate model parameters and for prevalence mapping. However, simply fitting a single model to the combined data from multiple surveys is inadvisable without testing the implicit assumption that both the underlying process and its realization are common to all of the surveys. We have developed a multivariate generalized linear geostatistical model to combine data from multiple spatially referenced prevalence surveys so as to address each of two common sources of variation across surveys: variation in prevalence over time; variation in data-quality. In the case of surveys that differ in quality, we assume that at least one of the surveys delivers unbiased gold-standard estimates of prevalence, whilst the others are potentially biased. For example, some surveys might use a random sampling design, the others opportunistic convenience samples. For parameter estimation and spatial predictions, we used Monte Carlo Maximum Likelihood methods. We describe an application to malaria prevalence data from Chikhwawa District, Malawi. The data consist of two Malaria Indicator Surveys (MISs) and an Easy Access Group (EAG) study, conducted over the period 2010-2012. In the two MISs, the data were collected by random selection of households in an area of 50 villages within 400 square kilometers, whilst the EAG study enrolled a random selection of children attending the vaccination clinic in Chikhwawa District Hospital. The second sampling strategy is more economical, but the sampling bias inherent to such "convenience" samples needs to be taken into account. |
|
Wed 16 Jul, '14- |
CRiSM Seminar - Adelchi AzzaliniA1.01Adelchi Azzalini (University of Padova) Clustering based on non-parametric density estimation: A proposal Cluster analysis based on non-parametric density estimation represents an approach to the clustering problem whose roots date back several decades, but it is only in recent times that this approach could actually be developed. The talk presents one proposal within this approach which is among the few ones which have been brought up to the operational stage. |
|
Wed 8 Oct, '14- |
CRiSM SeminarMS.03Christophe Ley - Universite Livre de Bruxelles Stein's method, Information theory, and Bayesian statistics In this talk, I will first describe a new general approach to the celebrated Stein method for asymptotic approximations and apply it to diverse approximation problems. Then I will show how Stein’s method can be successfully used in two a priori unrelated domains, namely information theory and Bayesian statistics. In the latter case, I will evaluate the influence of the choice of the prior on the posterior distribution at given sample size n. Based on joint work with Gesine Reinert (Oxford) and Yvik Swan (Liege). |
|
Thu 16 Oct, '14- |
CRiSM SeminarA1.01Karthik Bharath People use models in all fields of science, technology, management, etc. These can range from highly complex mathematical models based on systems of differential equations to relatively simple empirical, statistical, models. This talk is about the uncertainty in the predictions made by models. One aspect of this has come to be called Uncertainty Quantification (UQ), and is concerned with deriving the uncertainty in model outputs induced by uncertainty in the inputs. But there is another component of uncertainty that is much more important: all models are wrong. This talk is about just how badly misled we can be if we forget this fact. |
|
Thu 30 Oct, '14- |
CRiSM Seminar - Pierre Jacob & Leonardo BottoloA1.01Pierre Jacob |
|
Thu 13 Nov, '14- |
CRiSM Seminar - Michael Eichler (Maastricht) & Richard Huggins (Melbourne)A1.01Michael Eichler (Maastricht) In time series analysis, inference about cause-effect relationships among multiple time series is commonly based on the concept of Granger causality, which exploits temporal structure to achieve causal ordering of dependent variables. One major and well known problem in the application of Granger causality for the identification of causal relationships is the possible presence of latent variables that affect the measured components and thus lead to so-called spurious causalities. We present a new graphical approach for describing and analysing Granger-causal relationships in multivariate time series that are possibly affected by latent variables. It is based on mixed graphs in which directed edges represent direct influences among the variables while dashed edges---directed or undirected---indicate associations that are induced by latent variables. We show how such representations can be used for inductive causal learning from time series and discuss the underlying assumptions and their implications for causal learning. Finally we will discuss tetrad constraints in the time series context and how the can be exploited for causal inference. Richard Huggins (Melbourne) |
|
Thu 27 Nov, '14- |
CRiSM Seminar - Daniel Williamson (Exeter) & David van Dyk (Imperial)A1.01David van Dyk (Imperial) |
|
Tue 2 Dec, '14- |
CRiSM Seminar - David Draper (UC-Santa Cruz), Luis Nieto Barajas (ITAM - Instituto Tecnologico Autonomo de Mexico)A1.01Luis Nieto Barajas (ITAM - Instituto Tecnologico Autonomo de Mexico) |
|
Fri 23 Jan, '15- |
CRiSM Seminar - Rebecca Killick (Lancaster), Peter Green (Bristol)B1.01 (Maths)Rebecca Killick (Lancaster) |
|
Fri 6 Feb, '15- |
CRiSM Seminar - Gareth Peters (UCL), Leonhard Held (University of Zurich)B1.01 (Maths)Gareth Peters (UCL) Leonard Held (University of Zurich) |
|
Fri 20 Feb, '15- |
CRiSM Seminar - Marina Knight (York)B1.01 (Maths)Marina Knight (York) Hurst exponent estimation for long-memory processes using wavelet lifting |
|
Fri 1 May, '15- |
CRiSM Seminar - Marcelo Pereyra (Bristol), Magnus Rattray (Manchester)D1.07 (Complexity)Marcelo Pereyra (Bristol)
Proximal Markov chain Monte Carlo: stochastic simulation meets convex optimisation
Convex optimisation and stochastic simulation are two powerful computational methodologies for performing statistical inference in high-dimensional inverse problems. It is widely acknowledged that these methodologies can complement each other very well, yet they are generally studied and used separately. This talk presents a new Langevin Markov chain Monte Carlo method that uses elements of convex analysis and proximal optimisation to simulate efficiently from high-dimensional densities that are log-concave, a class of probability distributions that is widely used in modern high-dimensional statistics and data analysis. The method is based on a new first-order approximation for Langevin diffusions that uses Moreau-Yoshida approximations and proximity mappings to capture the log-concavity of the target density and construct Markov chains with favourable convergence properties. This approximation is closely related to Moreau-Yoshida regularisations for convex functions and uses proximity mappings instead of gradient mappings to approximate the continuous-time process. The proposed method complements existing Langevin algorithms in two ways. First, the method is shown to have very robust stability properties and to converge geometrically for many target densities for which other algorithms are not geometric, or only if the time step is sufficiently small. Second, the method can be applied to high-dimensional target densities that are not continuously differentiable, a class of distributions that is increasingly used in image processing and machine learning and that is beyond the scope of existing Langevin and Hamiltonian Monte Carlo algorithms. The proposed methodology is demonstrated on two challenging models related to image resolution enhancement and low-rank matrix estimation, which are not well addressed by existing MCMC methodology.
|
|
Fri 15 May, '15- |
CRiSM Seminar - Carlos Carvalho (UT Austin), Andrea Riebler (Norwegian University of Science & Technology)D1.07 (Complexity)Carlos Carvalho, (The University of Texas) Decoupling Shrinkage and Selection in Bayesian Linear Models: A Posterior Summary Perspective Andrea Riebler, (Norwegian University of Science and Technology) |
|
Fri 29 May, '15- |
CRiSM Seminar - Clifford Lam (LSE), Zoltan Szabo (UCL)D1.07 (Complexity)Zoltán Szabó, (UCL) Regression on Probability Measures: A Simple and Consistent Algorithm We address the distribution regression problem: we regress from probability measures to Hilbert-space valued outputs, where only samples are available from the input distributions. Many important statistical and machine learning problems can be phrased within this framework including point estimation tasks without analytical solution, or multi-instance learning. However, due to the two-stage sampled nature of the problem, the theoretical analysis becomes quite challenging: to the best of our knowledge the only existing method with performance guarantees requires density estimation (which often performs poorly in practise) and the distributions to be defined on a compact Euclidean domain. We present a simple, analytically tractable alternative to solve the distribution regression problem: we embed the distributions to a reproducing kernel Hilbert space and perform ridge regression from the embedded distributions to the outputs. We prove that this scheme is consistent under mild conditions (for distributions on separable topological domains endowed with kernels), and construct explicit finite sample bounds on the excess risk as a function of the sample numbers and the problem difficulty, which hold with high probability. Specifically, we establish the consistency of set kernels in regression, which was a 15-year-old-open question, and also present new kernels on embedded distributions. The practical efficiency of the studied technique is illustrated in supervised entropy learning and aerosol prediction using multispectral satellite images. [Joint work with Bharath Sriperumbudur, Barnabas Poczos and Arthur Gretton.]
Clifford Lam, (LSE) Nonparametric Eigenvalue-Regularized Precision or COvariance Matrix Estimator for Low and High Frequency Data Analysis We introduce nonparametric regularization of the eigenvalues of a sample covariance matrix through splitting of the data (NERCOME), and prove that NERCOME enjoys asymptotic optimal nonlinear shrinkage of eigenvalues with respect to the Frobenius norm. One advantage of NERCOME is its computational speed when the dimension is not too large. We prove that NERCOME is positive definite almost surely, as long as the true covariance matrix is so, even when the dimension is larger than the sample size. With respect to the inverse Stein’s loss function, the inverse of our estimator is asymptotically the optimal precision matrix estimator. Asymptotic efficiency loss is defined through comparison with an ideal estimator, which assumed the knowledge of the true covariance matrix. We show that the asymptotic efficiency loss of NERCOME is almost surely 0 with a suitable split location of the data. We also show that all the aforementioned optimality holds for data with a factor structure. Our method avoids the need to first estimate any unknowns from a factor model, and directly gives the covariance or precision matrix estimator. Extension to estimating the integrated volatility matrix for high frequency data is presented as well. Real data analysis and simulation experiments on portfolio allocation are presented for both low and high frequency data. |
|
Fri 12 Jun, '15- |
CRiSM Seminar - Sara van der Geer (Zurich), Daniel Simpson (Warwick)D1.07 (Complexity)Daniel Simpson (University of Warwick) Penalising model component complexity: A principled practical approach to constructing priors Setting prior distributions on model parameters is the act of characterising the nature of our uncertainty and has proven a critical issue in applied Bayesian statistics. Although the prior distribution should ideally encode the users’ uncertainty about the parameters, this level of knowledge transfer seems to be unattainable in practice and applied statisticians are forced to search for a “default” prior. Despite the development of objective priors, which are only available explicitly for a small number of highly restricted model classes, the applied statistician has few practical guidelines to follow when choosing the priors. An easy way out of this dilemma is to re-use prior choices of others, with an appropriate reference. In this talk, I will introduce a new concept for constructing prior distributions. We exploit the natural nested structure inherent to many model components, which defines the model component to be a flexible extension of a base model. Proper priors are defined to penalise the complexity induced by deviating from the simpler base model and are formulated after the input of a user- defined scaling parameter for that model component, both in the univariate and the multivariate case. These priors are invariant to reparameterisations, have a natural connection to Jeffreys’ priors, are designed to support Occam’s razor and seem to have excellent robustness properties, all which are highly desirable and allow us to use this approach to define default prior distributions. Through examples and theoretical results, we demonstrate the appropriateness of this approach and how it can be applied in various situations, like random effect models, spline smoothing, disease mapping, Cox proportional hazard models with time-varying frailty, spatial Gaussian fields and multivariate probit models. Further, we show how to control the overall variance arising from many model components in hierarchical models. This joint work with Håvard Rue, Thiago G. Martins, Andrea Riebler, Geir-Arne Fuglstad (NTNU) and Sigrunn H. Sørbye (Univ. of Tromsø). Sara van de Geer (ETH Zurich) Norm-regularized Empirical Risk Minimization |
|
Fri 26 Jun, '15- |
CRiSM Seminar - Thomas Hamelryck (University of Copenhagan), Anjali Mazumder (Warwick)D1.07 (Complexity)Thomas Hamelryck (Bioinformatics Center, University of Copenhagen) Inference of protein structure and ensembles using Bayesian statistics and probability kinematics The so-called protein folding problem is the loose designation for an amalgam of closely related, unsolved problems that include protein structure prediction, protein design and the simulation of the protein folding process. We adopt a unique Bayesian approach to modelling bio-molecular structure, based on graphical models, directional statistics and probability kinematics. Notably, we developed a generative probabilistic model of protein structure in full atomic detail. I will give an overview of how rigorous probabilistic models of something as complicated as a protein's atomic structure can be formulated, focusing on the use of graphical models and directional statistics to model angular degrees of freedom. I will also discuss the reference ratio method, which is needed to "glue" several probabilistic models of protein structure together in a consistent way. The reference ratio method is based on "probability kinematics", a little known method to perform Bayesian inference proposed by the philosopher Richard C. Jeffrey at the end of the fifties. Probability kinematics might find widespread application in statistics and machine learning as a way to formulate complex, high dimensional probabilistic models for multi-scale problems by combining several simpler models. Anjali Mazumder (University of Warwick)
Probabilistic Graphical Models for planning and reasoning of scientific evidence in the courts
The use of probabilistic graphical models (PGMs) has gained prominence in the forensic science and legal literature when evaluating evidence under uncertainty. The graph-theoretic and modular nature of the PGMs provide a flexible and graphical representation of the inference problem, and propagation algorithms facilitate the calculation of laborious marginal and conditional probabilities of interest. In giving expert testimony regarding, for example, the source of a DNA sample, forensic scientists under much scrutiny, are often asked to justify their decision-making-process. Using information-theoretic concepts and a decision-theoretic framework, we define a value of evidence criterion as a general measure of informativeness for a forensic query and collection of evidence to determine which and how much evidence contributes to the reduction of uncertainty. In this talk, we demonstrate how this approach can be used for a variety of planning problems and the utility of PGMs for scientific and legal reasoning.
|
|
Mon 12 Oct, '15- |
CRiSM Seminar - Dan Roy (University of Toronto)A1.01Dan Roy (University of Toronto) For finite parameter spaces under finite loss, there is a close link between optimal frequentist decision procedures and Bayesian procedures: every Bayesian procedure derived from a prior with full support is admissible, and every admissible procedure is Bayes. This relationship breaks down as we move beyond finite parameter spaces. There is a long line of work relating admissible procedures to Bayesian ones in more general settings. Under some regularity conditions, admissible procedures can be shown to be the limit of Bayesian procedures. Under additional regularity, they are generalized Bayesian, i.e., they minimize the average loss with respect to an improper prior. In both these cases, one must venture beyond the strict confines of Bayesian analysis. Using methods from mathematical logic and nonstandard analysis, we introduce the notion of a hyperfinite statistical decision problem defined on a hyperfinite probability space and study the class of nonstandard Bayesian decision procedures---namely, those whose average risk with respect to some prior is within an infinitesimal of the optimal Bayes risk. We show that if there is a suitable hyperfinite approximation to a standard statistical decision problem, then every admissible decision procedure is nonstandard Bayes, and so the nonstandard Bayesian procedures form a complete class. We give sufficient regularity conditions on standard statistical decision problems admitting hyperfinite approximations. Joint work with Haosui (Kevin) Duanmu. |
|
Mon 26 Oct, '15- |
CRiSM Seminar - Hernando Ombao (UC Irvine, Dept of Statistics))A1.01Hernando Ombao (UC Irvine, Dept of Statistics) exhibits abnormal firing behavior which then spreads to other subpopulations of neurons. This abnormal firing behavior is captured by increases in signal amplitudes (which can be easily spotted by visual inspection) and changes in the decomposition of the waveforms and in the strength of dependence between different regions (which are more subtle). The proposed frequency-specific change-point detection method (FreSpeD) uses a cumulative sum test statistic within a binary segmentation algorithm. Theoretical optimal properties of the FreSpeD method will be developed. We demonstrate that, when applied to an epileptic seizure EEG data, FreSpeD identifies the correct brain region as the focal point of seizure, the time of seizure onset and the very subtle changes in cross-coherence immediately preceding seizure onset. The goal of the second project to track changes in spatial boundaries (or more generally spatial sets or clusters) as the seizure process unfolds. A pair of channels (or a pair of sets of channels) are merged into one cluster if they exhibit synchronicity as measured by, for example, similarities in their spectra or by the strength of their coherence. We will highlight some open problems including developing a model for the evolutionary clustering of non-stationary time series. The first project is in collaboration with Anna Louise Schröder (London School of Economics); the second is with Carolina Euan (CIMAT, Mexico), Joaquin Ortega (CIMAT, Mexico) and Ying Sun (KAUST, Saudi Arabia). |