Skip to main content Skip to navigation

Events

Select tags to filter on
  More events Jump to any date

Search calendar

Enter a search term into the box below to search for all events matching those terms.

Start typing a search term to generate results.

How do I use this calendar?

You can click on an event to display further information about it.

The toolbar above the calendar has buttons to view different events. Use the left and right arrow icons to view events in the past and future. The button inbetween returns you to today's view. The button to the right of this shows a mini-calendar to let you quickly jump to any date.

The dropdown box on the right allows you to see a different view of the calendar, such as an agenda or a termly view.

If this calendar has tags, you can use the labelled checkboxes at the top of the page to select just the tags you wish to view, and then click "Show selected". The calendar will be redisplayed with just the events related to these tags, making it easier to find what you're looking for.

 
Thu 17 Jan, '08
-
CRiSM Seminar

Dr Elena Kulinskaya, Statistical Advisory Service, Imperial College
Meta analysis on the right scale

This talk is about an approach to meta analysis and to statistical evidence developed jointly with Stephan Morgenthaler and Robert Staudte, and now written up in our book 'Meta Analysis: a guide to calibrating and combining statistical evidence'  to be published by Wiley very soon. The traditional ways of measuring evidence, in particular with p-values, are neither intuitive nor useful when it comes to making comparisons between experimental results, or when combining them. We measure evidence for an alternative hypothesis, not evidence against a null. To do this, we have in a sense adopted standardized scores for the calibration scale. Evidence for us is simply a transformation of a test statistic S to another one (called evidence T=T(S)) whose distribution is close to normal with variance 1, and whose mean grows from 0 with the parameter as it moves away from the null. Variance stabilization is used to arrive on this scale. For meta analysis the results from different studies are transformed to a common calibration scale, where it is simpler to combine and interpret them. 

Thu 24 Jan, '08
-
CRiSM Seminar
A1.01

Dr Richard Samworth, Statistical Laboratory, Cambridge
Computing the maximum likelihood estimator of a multidimensional log-concave density

We show that if $X_1,...,X_n$ are a random sample from a log-concave density $f$ in $\mathbb{R}^d$, then with probability one there exists a unique maximum likelihood estimator $\hat{f}_n$ of $f$.  The use of this estimator is attractive because, unlike kernel density estimation, the estimator is fully automatic, with no smoothing parameters to choose. The existence proof is non-constructive, however, and in practice we require an iterative algorithm that converges to the estimator.  By reformulating the problem as one of non-differentiable convex optimisation, we are able to exhibit such an algorithm.  We will also show how the method can be combined with the EM algorithm to fit finite mixtures of log-concave densities.  The talk will be illustrated with pictures from the R package LogConcDEAD.

This is joint work with Madeleine Cule (Cambridge), Bobby Gramacy (Cambridge) and Michael Stewart (University of Sydney).

Thu 31 Jan, '08
-
CRiSM Seminar
A1.01

Dr Robert Gramacy, Statistical Laboratory Cambridge
Importance Tempering
Simulated tempering (ST) is an established Markov Chain Monte Carlo (MCMC) methodology for sampling from a multimodal density pi(theta).  The technique involves introducing an auxiliary variable k taking values in a finite subset of [0,1] and indexing a set of tempered distributions, say pi_k(theta) = pi(theta)^k.  Small values of k encourage better mixing, but samples from pi are only obtained when the joint chain for (theta,k) reaches k=1. However, the entire chain can be used to estimate expectations under pi of functions of interest, provided that importance sampling (IS) weights are calculated.  Unfortunately this method, which we call importance tempering (IT),  has tended not work well in practice.  This is partly because the most immediately obvious implementation is naive and can lead to high variance estimators.  We derive a new optimal method for combining multiple IS estimators and prove that this optimal combination has a highly desirable property related to the notion of effective sample size.  The methodology is applied in two modelling scenarios requiring reversible-jump MCMC, where the naive approach to IT fails: model averaging in treed models, and model selection for mark--recapture data. 

Thu 14 Feb, '08
-
CRiSM Seminar
A1.01

Professor Simon Wood, University of Bath
Generalized Additive Smooth Modelling
Generalized Additive Models are GLMs in which the linear predictor is made up, partly, of a sum of smooth functions of predictor variables. I will talk about the penalized regression spline approach to GAM, as implemented, for example, in R package mgcv. In particular I will focus on two interesting aspects: low rank representation of smooth functions of several covariates and stable computation of both model coefficients and smoothing parameters. More than half the slides will have pictures.
 

Mon 18 Feb, '08
-
CRiSM Seminar
A1.01

Terry Speed, University of California, Berkeley
Alternative Splicing in Tumors: Detection and Interpretation
In this talk I will discuss using the Affymetrix GeneChip Human Exon and Human Gene 1.0 ST arrays for the detection of genes spliced differently in some tumors in comparison with others. I plan to begin by introducing the arrays and the expression data they produce. Next I will outline the way in which we use such data in our attempts to identify exon-tumor combinations exhibiting splicing patterns different from the majority. This will be illustrated by examples from publicly available tissue and mixture data. Then I will briefly discuss some of the additional issues which arise when we seek to enumerate such alternative splicing patterns on a genome-wide scale. Finally, I will exhibit some of the results we have found applying these methods to glioblastoma tissue samples collected as part of The Cancer Genome Atlas (TCGA) project. (This is joint work with ELizabeth Purdom, Mark Robinson, Ken Simpson, and members of the Berkeley Cancer Genome Center.)

Thu 28 Feb, '08
-
CRiSM Seminar
A1.01

Alexey Koloydenko & Juri Lember (Joint Talk), University of Nottingham
Adjusted Viterbi Training for Hidden Markov Models
The Expectation Maximisation (EM) procedure is a principal tool for parameter estimation in hidden Markov models (HMMs). However, in applications EM is sometimes replaced by Viterbi training, or extraction, (VT). VT is computationally less intensive and more stable, and has more of an intuitive appeal, but VT estimation is biased and does not satisfy the following fixed point property: Hypothetically, given an infinitely large sample and initialized to the true parameters, VT will generally move away from the initial values. We propose adjusted Viterbi training (VA), a new method to restore the fixed point property and thus alleviate the overall imprecision of the VT estimators, while preserving the computational advantages of the baseline VT algorithm. Simulations show that VA indeed improves estimation precision appreciably in both the special case of mixture models and more general HMMs.

We will discuss the main idea of the adjusted Viterbi training. This will also touch on tools developed specifically to analyze asymptotic behaviour of maximum a posteriori (MAP) hidden paths, also known as Viterbi alignments. Our VA correction is analytic and relies on infinite Viterbi alignments and associated limiting probability distributions. While explicit in the special case of mixture models, these limiting measures are not obvious to exist for more general HMMs. We will conclude by presenting a result that under certain mild conditions, general (discrete time) HMMs do possess the limiting distributions required for the construction of VA. 

Thu 6 Mar, '08
-
CRiSM Seminar
A1.01
Dr Cliona Golden, UCD, Dublin
On the validity of ICA for fMRI data
Functional Magnetic Resonance Imaging (fMRI) is a brain-imaging technique which, over time, records changes in blood oxygenation level that can be associated with underlying neural activity. However, fMRI images are very noisy and extracting useful information from them calls for a variety of methods of analysis.

I will discuss the validity of the use of two popular Independent Component Analysis (ICA) algorithms, InfoMax and FastICA, which are commonly used for fMRI data analysis.

Tests of the two algorithms on simulated, as well as real, fMRI data, suggest that their successes are related to their ability to detect "sparsity" rather than the independence which ICA is designed to seek.
Thu 13 Mar, '08
-
CRiSM Seminar
A1.01

Prof Antony Pettitt, Lancaster University
Statistical inference for assessing infection control measures for the transmission of pathogens in hospitals
Patients can acquire infections from pathogen sources within hospitals and certain pathogens appear to be found mainly in hospitals.  Methicillin-resistant Staphylococcus Aureus (MRSA) is an example of a hospital acquired pathogen that continues to be of particular concern to patients and hospital management.  Patients infected with MRSA can develop severe infections which lead to increased patient morbidity and costs for the hospital.  Pathogen transmission to a patient can occur via health-care workers that do not regularly perform hand hygiene.  Infection control measures that can be considered include isolation for colonised patients and improved hand hygiene for health-care workers.

The talk develops statistical methods and models in order to assess the effectiveness of the two control measures (i) isolation and (ii) improved hand hygiene.  For isolation, data from a prospective study carried out in a London hospital is considered and statistical models based on detailed patient data are used to determine the effectiveness of isolation.  The approach is Bayesian.
For hand hygiene it is not possible, for ethical and practical reasons, to carry out a prospective study to investigate various levels of hand hygiene.  Instead hand hygiene effects are investigated by simulation using parameter values estimated from data on health-care worker hand hygiene and weekly colonisation incidence collected from a hospital ward in Brisbane.  The approach uses profile likelihoods.Both approaches involve transmission parameters where there is little information available and contrasting compromises have to be made. Conclusions about the effectiveness of the two infection control measures will be discussed. The talk involves collaborative work with Marie Forrester, Emma McBryde, Chris Drovandi, Ben Cooper, Gavin Gibson, Sean McElwain.

 

Thu 3 Apr, '08
-
CRiSM Seminar
A1.01
Professor Jay Kadane, Carnegie Mellon University
Driving While Black: Statisticians Measure Discriminatory Law Enforcement (joint work with John Lamberth)
The US Constitution guarantees "equal protection under the law" regardless of race, but sometimes law enforcement practices have failed to adhere to this standard.In the 1990's, a suit was brought alleging that the New Jersey State Police were stopping Blacks at disproportionately high rates in the southern end of the New Jersey Turnpike. In this talk I
* review the evidence in that case, the decision, and its immediate aftermath
* discuss criticisms of that decision
* examine new evidence that rebuts those criticisms
* comment on the extent to which the Constitutional standard is now being met.
Thu 1 May, '08
-
CRiSM Seminar
A1.01

Alastair Young, Imperial College London
Objective Bayes and Conditional Inference
In Bayesian parametric inference, in the absence of subjective prior information about the parameter of interest, it is natural to consider use of an objective prior which leads to posterior probability quantiles which have, at least to some higher order approximation in terms of the sample size, the correct frequentist interpretation. Such priors are termed probability matching priors. In many circumstances, however, the appropriate frequentist inference is a conditional one. The key contexts involve inference in multi-parameter exponential families, where conditioning eliminates the nuisance parameter, and models which admit ancillary statistics, where conditioning on the ancillary is indicated by the conditionality principle of inference. In this talk, we consider conditions on the prior under which posterior quantiles have, to high order, the correct conditional frequentist interpretation. The key motivation for the work is that the conceptually simple objective Bayes route may provide accurate approximation to more complicated frequentist procedures. We focus on the exponential family context, where it turns out that the condition for higher order conditional frequentist accuracy reduces to a condition on the model, not the prior: when the condition is satisfied, as it is in many key situations, any first order probability matching prior (in the unconditional sense) automatically yields higher order conditional probability matching. We provide numerical illustrations, discuss the relationship between the objective Bayes inference and the parametric bootstrap, as well as giving a brief account appropriate to the ancillary statistic context, where conditional frequentist probability matching is more difficult. [This is joint work with Tom DiCiccio, Cornell].

Thu 8 May, '08
-
CRiSM Seminar
A1.01
Cees Diks, University of Amsterdam
Linear and Nonlinear Causal Relations in Exchange Rates and Oil Spot and Futures Prices
Various tests have been proposed recently in the literature for detecting causal relationships between time series. I will briefly review the traditional linear methods and some more recent contributions on testing for nonlinear Granger causality. The relative benefits and limitations of these methods are then compared in two different case studies with real data. In the first case study causal relations between six main currency exchange rates are considered. After correcting for linear causal dependence using VAR models there is still evidence presence for nonlinear causal relations between these currencies. ARCH and GARCH effects are insufficient to fully account for the nonlinear causality found. The second case study focuses on nonlinear causal linkages between daily spot and futures prices at different maturities of West Texas Intermediate crude oil. The results indicate that after correcting for possible cointegration, linear dependence and multivariate GARCH effects, some causal relations are still statistically significant. In both case studies the conclusion is that non-standard models need to be developed to fully capture the higher-order nonlinear dependence in the data.
Thu 22 May, '08
-
CRiSM Seminar
A1.01

Thomas Richardson, University of Washington
Estimation of the Relative Risk and Risk Difference

I will first review well-known differences between odds ratios, relative risks and risk differences. These results motivate the development of methods, analogous to logistic regression, for estimating the latter two quantities. I will then describe simple parametrizations that facilitate maximum-likelihood estimation of the relative risk and risk-difference. Further, these parametrizations allow for doubly-robust g-estimation of the relative risk and risk difference. (Joint work with James Robins, Harvard School of Public Health).

Thu 29 May, '08
-
CRiSM Seminar
A1.01

Geoff McLachlan, University of Queensland
On Simple Mixture Models in High-Dimensional Testing in the Detection of Differentially Expressed Genes

An important problem in microarray experiments is the detection of genes that are differentially expresse in a given number of classes. As there are usually thousands of genes to be considered simultaneously, one encounters high-dimensional testing problems. We provide a straightforward and easily implemented method for estimating the posterior probability that an individual gene is null (not differentially expressed).  The problem can be expressed in a two-component mixture framework.  Current methods of implementing this approach either have some limitations due to the minimal assumptions made or with the computationally intensive nature of more specific assumptions. By converting to a z-score the value of the test statistic used to test the significance of each gene, we propose a simple two-component normal mixture that models adequately the distribution of this score. The approach provides an estimate of the local false discovery rate (FDR) for each gene, which is taken to be the posterior probability that the gene is null.  Genes with the local FDR less than a specified threshold C are taken to be differentially expressed.  For a given C, this approach also provides estimates of the implied overall errors such as the (global) FDR and the false negative/positive rates.

Thu 5 Jun, '08
-
CRiSM Seminar
A1.01

Thomas Nichols, GlaxoSmithKline Clinical Imaging Centre
Cluster mass inference - a new random field theory method for neuroimaging
Functional and structural brain imaging analyses typically fit univariate models at each volume element, or voxel, producing images of T-statistics. These statistics images must be threshold or assessed insome way to identify significant regions while controlling some specified false positive rate. A widely used method for assessing signficance is the cluster size test, where an arbitary threshold is applied to the statistic image and contiguous suprathreshold voxels are grouped into clusters. The size of a cluster forms a test statistic on the null hypothesis of no activation anywhere within the cluster, and P-values are found with Random Field Theory (RFT). Various authors have reported on improved sensitivity with a modified version of the cluster size test, the cluster mass test. Cluster mass is the integral of suprathreshold intensities within the cluster, and captures information about both the extent and magnitude of the effect. However, all previous work has relied on permutation inference as no distributional results have been available. I will show recent work on deriving a null distribution for cluster mass using RFT. Using the approximate parabolic shape about local maxima and distributional results for the cluster extent and peak curvature, we produce a joint distribution for mass and peak height which is marginalized to produce a P-value for the mass statistic. We show results on simulated and real data which demonstrate the tests validity and power. 

Thu 12 Jun, '08
-
CRiSM Seminar
A1.01
Jonathan Dark, University of Melbourne
Dynamic hedging with futures that are subject to price limits
The standard approaches to estimating minimum variance hedge ratios (MVHRs) are mis-specified when futures prices are subject to price limits. This paper proposes a bivariate tobit-FIGARCH model with maturity effects to estimate dynamic MVHRs using single and multiple period approaches. Simulations and an application to a commodity futures hedge support the proposed approach and highlight the importance of allowing for price limits when hedging.   
Thu 19 Jun, '08
-
CRiSM Seminar
A1.01

Ian Dryden, University of Nottingham
Non-Euclidean statistics for covariance matrices, with applications to diffusion tensor imaging
The statistical analysis of covariance matrices occurs in many important applications, e.g. in diffusion tensor imaging or longitudinal data analysis. Methodology is discussed for estimating covariance matrices which takes into account the non-Euclidean nature of the space of positive semi-definite symmetric matrices. We make connections with the use of Procrustes methods in shape analysis, and comparisons are made with other estimation techniques, including using the matrix logarithm, Riemannian metric, matrix square root and Cholesky decomposition. Our main application will be diffusion tensor imaging which is used to study the fibre structure in the white matter of the brain. Diffusion weighted MR images involve applying a set of gradients in a design of directions, and the recorded data are related to the Fourier transforms of the displacement of water molecules at each voxel. Under a multivariate Gaussian diffusion model a single tensor (3 x 3 covariance matrix) is fitted at each voxel. We discuss the statistical analysis of diffusion tensors, including construction of means, spatial interpolation, geoedsics, and principal components analysis. This is joint work with Alexey Koloydenko and Diwei Zhou.

Thu 26 Jun, '08
-
CRiSM Seminar
A1.01
Gersende Fort, ENST (Ecole Nationale Superieure Des Telecommunications, France
Stability of Markov Chains based on fluid limit techniques. Applications to MCMC
We propose a transformation of some Markov chains which will allow us to define its fluid limit: by renormalization in time, space, and initial value of the chain, we exhibit a time-continuous process which governs the dynamic of the initial chain. The goal is to identify the quantities that govern the ergodic behavior of the Markov chain, by showing their impact on the dynamics of the associated fluid  process which, by definition, gives information on  the transient steps of the chain. We will consider  applications of these techniques to the choice of the design parameters of some MCMC samplers.
Mon 14 Jul, '08
-
CRiSM Seminar
A1.01
Prof Donald Martin, North Carolina State University
Markov chain pattern distributions
We give a method for predicting statistics of hidden state sequences, where the conditional distribution of states given observations is modeled by a factor graph with factors that depend on past states but not future ones.  Model structure is exploited to develop a deterministic finite automaton and an associated Markov chain that facilitates efficient computation of the distributions.  Examples of applications of the methodology are the computation of distributions of patterns and statistics in a discrete hidden state sequence perturbed by noise and/or missing values, and patterns in a state sequence that serves to classify the observations.  Two detailed examples are given to illustrate the computational procedure. 
Mon 25 Aug, '08
-
Economics/Stats Seminar
S2.79
Donald Rubin (Harvard)
For Objective Causal Inference, Design Trumps Analysis
For obtaining causal inference that are objective, and therefore have the best chance of revealing scientific truths, carefully designed and executed randomized experiments are generally considered to be the gold standard.  Observational studies, in contrast, are generally fraught with problems that compromise any claim for objectivity of the resulting causal inferences.  The thesis here is that observational studies have to be carefully designed to approximate randomized experiments, in particular, without examining any final outcome data.  Often a candidate data set will have to be rejected as inadequate because of lack of data on key covariates, or because of lack of overlap in the distributions of key covariates between treatment and control groups, often revealed by careful propensity score analyses.  Sometimes the template for the approximating randomized experiment will have to be altered, and the use of principal stratification can be helpful in doing this.  These issues are discussed and illustrated using the framework of potential outcomes to define causal effects, which greatly clarifies critical issues.
Tue 9 Sep, '08
-
CRiSM Seminar
A1.01
Professor P Cheng, Academia Sinica, Taipei, Republic of China
Linear Information Models and Applications
Log-likelihood information identities and Venn diagrams for categorical data exhibit fundamental differences from those of continuous variables. This presentation will start with three-way contingency tables and the associated likelihood ratio tests. It will introduce linear information models that deviate from hierarchical log-linear models, beginning with three-way tables. A connection to latent class analysis with two-way tables and the geometry of the one-degree-freedom chi-square test and exact test for two-way independence is also investigated.

Key Names: Pearson; Fisher; Neyman and Pearson; Kullback and Leibler; Cochran, Mantel, and Heanszel; Goodman.

Co-authors: John A. D. Aston, Jiun W. Liou, and Michelle Liou.

Mon 22 Sep, '08
-
CRiSM Seminar
A1.01
Dr Jonathan Evans, Institute of Linguistics, Academia Sinica
Statistical Modelling in Linguistics: Approaches and Challenges in pitch analysis

This talk introduces the use of Linear Mixed Effects (LME) analysis to model f0 (pitch) production in a language with two  tones, and demonstrates the advantages of using such a method of analysis.  LME can be used to weigh the impact of a large number of effects, it can demonstrate the interaction among those effects, and can also show how both fixed and random effects contribute to the model.  Unlike previous analytical methods for modeling f0 in tone languages, LME analysis allows researchers to have more freedom in designing experiments, and to have sufficient variety in the dataset without having to rely on nonsense words and phrases to fill out a data matrix. LME makes it is possible to put a multitude of effects and interactions into a single comprehensive model of f0.  The ensuing model is easy to interpret and straightforward to compare crosslinguistically.  LME analysis makes possible a quantitative typology that shows clearly how linguistic and nonlinguistic factors combine in the production of f0 for each language thus analyzed. The talk will also veer into discussion of how to model f0 based on the pitch curve of each syllable.  Although each curve contains an infinite number of points, there is striking similarity between the curve-based model and the point-based model.

 
Thu 2 Oct, '08
-
CRiSM Seminar: Martin Baxter
A1.01
Dr Martin Baxter, Nomura International
Levy Modelling of Credit

This talk will start with some simple models of credit dynamics, and embed them in a general Levy process framework. A particular instance, the Gamma process, will then be studied with reference to both its theoretical and practical properties. A brief analysis of the ongoing credit crisis in terms of Levy modelling. Time permitting, we will also look at some other applications.

Thu 23 Oct, '08
-
Oxford-Warwick Joint Seminar
Oxford

 

3rd JOINT WARWICK-OXFORD STATISTICS SEMINAR

2:30 – 5:00 pm  at The Mary Ogilvie Lecture Theatre,

St. Anne’s College, University of Oxford

2:30 p.m.

Speaker 1:  Julian Besag (University of Bath, University of Washington, Seattle)

Title:  Continuum limits of Gaussian Markov random fields: resolving the conflict with geostatistics

Abstract:  For more than 30 years, Markov random fields (equivalently, graphical models with undirected edges) have been used with some success to account for spatial variation in data. Applications include agricultural crop trials, geographical epidemiology, medical imaging, remote sensing, astronomy, and microarrays. Almost all of the examples involve (hidden) Gaussian MRF formulations.

MRFs refer to fixed regular or irregular discrete lattices or arrays and questions arise regarding inconsistencies between MRFs specified at differing scales, especially for regional data. Ideally, one would often prefer an underlying continuum formulation, as in geostatistics, which can then be integrated to the regions of interest. However, limiting continuum versions of MRFs, as lattice spacing decreases, proved elusive until recently.

This talk briefly presents some motivating examples and shows that limiting processes indeed exist but are defined on arbitrary regions of the plane rather than pointwise. Especially common is the conformally invariant de Wijs process, which coincidentally was used originally by mining engineers but which became unfashionable as geostatistics developed. Convergence is generally very fast. The de Wijs process is also shown to be a natural extension of Brownian motion to the plane. Other processes, including the thin-plate spline, can be derived as limits of MRFs. The talk closes by briefly discussing data analysis.

 

ÍÎ

 

3:30 to 4.00 - Tea, Coffee and biscuits in foyer outside lecture theatre

 

4:00 p.m.

Speaker 2:   Susan Lewis (University of Southampton, UK)

Title:  Screening experiments

Abstract:  Discovery and development in science and industry often involves investigation of many features or factors that could potentially affect the performance of a product or process. In factor screening, designed experiments are used to identify efficiently the few features that influence key properties of the system under study. A brief overview of this broad area will be presented. This will be followed by discussion of a variety of methods with particular emphasis on industrial screening. Ideas will be motivated and illustrated through examples, including a case study from the automotive industry.

5:00 reception in the foyer
Thu 30 Oct, '08
-
CRiSM Seminar: Christopher Sherlock
A1.01
Dr Christopher Sherlock
Optimal scaling of the random walk Metropolis
University of Lancaster
Abstract: The random walk Metropolis (RWM) is one of the most commonly used Metropolis-Hasting algorithms, and choosing the appropriate scaling for the proposal is an important practical problem. Previous theoretical approaches have focussed on high-dimensional algorithms and have revolved around a diffusion approximation of the trajectory. For certain specific classes of targets it has been possible to show that the algorithm is optimal when the acceptance rate is approximately
0.234.

We develop a novel approach which avoids the need for diffusion limits. Focussing on spherically symmetric targets, it is possible to derive simple exact formulae for efficiency and acceptance rate for a "real" RWM algorithm, as opposed to a limit process. The limiting behaviour of these formulae can then be explored. This in some sense "simpler" approach allows important general intuitions as to when and why the 0.234 rule holds, when the rule fails, and what may happen when it does fail. By extending the theory to include elliptically symmetric targets we obtain further intuitions about the role of the proposal's shape.
Thu 6 Nov, '08
-
CRiSM Seminar: Ming-Yen Cheng
A1.01

Prof Ming-Yen Cheng UCL
Generalized Multiparameter Likelihood Models
Multiparameter likelihood models (MLM) with multiple covariates have a wide range of applications. However, they encounter the ``curse of dimensionality" problem when the dimension of the covariates is large. We develop a generalized multiparameter likelihood model that copes with multiple covariates and adapts to dynamic structural changes well. And it includes some popular models, such as the partially linear and varying coefficient models, as special cases. When the model is fixed, a simple and effective two-step method is developed to estimate both the parametric and the nonparametric components.  The proposed estimator of the parametric component has the root-n convergence rate, and the estimator of the nonparametric component enjoys an adaptivity property. A data-driven procedure is suggested to select the bandwidths involved. Also proposed is a new initial estimator in profile likelihood estimation of the parametric part to ensure stability of the approach in general settings. We further develop an automatic procedure to identify constant parameters in the underlying model. A simulation study and an application to the  infant mortality data of China are given to demonstrate performance of the proposed methods.

Wed 12 Nov, '08
-
CRiSM Seminar - John Cussens
A1.01, Zeeman Building

James Cussens, University of York

 Model Selection using weighted MAX-SAT Solvers

This talk concerns encoding problems of statistical model selection in such a way that "weighted MAX-SAT solvers" can be used to search for the   'best' model. In this approach each model is (implicitly) encoded as a joint instantiation of n binary variables. Each of these binary variables encodes the truth/falsity of a logical proposition and weighted logical formulae are used to represent the model selection problem. Once encoded in this way we can tap into years of research and use any of the state-of-the-art solvers to conduct the search. In the talk I will show how to use this approach when the model class is that of Bayesian networks, and also for clustering. I will briefly touch on related methods which permit the calculation of marginal probabilities in discrete distributions.

Wed 12 Nov, '08
-
Joint Stats/Econometrics Seminar: Sylvia Fruehwirth-Schnatter
A1.01, Zeeman Building

Prof Sylvia Fruehwirth-Schnatter, Johannes Kepler University, Austria


Bayesian Variable Selection Problems for State Space and other Latent Variable Models

Latent Variable models are widely used in applied statistics and econometrics to deal with data where the underlying processes change either over time or between units.  Whereas estimation of these models is well understood, model selection problems are rarely studies, because such an issue usually leads to a non-regular testing problem.

Bayesian statistics offers in principle a framework for model selection even for non-regular problems, as is shortly discussed in the first part of the talk.  The practical application of the Bayesian approach, however, proves to be challenging and numerical technique like marginal likelihoods, RJMCMC or the variable selection approach have to be used.

The main contribution of this talk is to demonstrate that the Bayesian variable selection approach is useful far beyond the common problem of selecting covariates in a classical regression model and may be extended to deal model selection problems in various latent variable models.  First, it is extended to testing for the presence of unobserved heterogeneity in random effects models.  Second, dynamic regression models are considered, where one has to choose between fixed and random coefficients.  Finally, the variable selection approach is extended to state space models, where testing problems like discriminating between models with a stochastic trend, a deterministic trend and a model without trend arise.

Case studies from marketing, economics and finance will be considered for illustration.

Thu 27 Nov, '08
-
CRiSM Seminar - Anthony Reveillac
A1.01
Anthony Reveillac
Humboldt University - Berlin

 

Stein estimators and SURE shrinkage estimation for Gaussian processes using the Malliavin calculus

In this talk we present a construction of Stein type estimators for the drift of Gaussian processes using the Malliavin integration by parts formula. Then we construct an estimation and de-noising procedure for an input signal perturbed by a continuous-time Gaussian noise using the local and occupation times of Gaussian processes. This talk is based on joint works with Nicolas Privault.

Fri 12 Dec, '08
-
Joint CRiSM/Applied Maths/Stats Seminar
A1.01
Professor Jeff Rosenthal, University of Toronto
Adaptive MCMC
Thu 22 Jan, '09
-
CRiSM Seminar
Room A1.01, Zeeman Building

Dr Laura Sangalli, Politecnico di Milano

Title : Efficient estimation of curves in more than one dimension by free-knot regression splines, with applications to the analysis of 3D cerebral vascular geometries.

Abstract : We deal with the problem of efficiently estimating a 3D curve and its derivatives, starting from a discrete and noisy observation of the curve.  We develop a regression technique based on free-knot splines, ie. regression splines where the number and position of knots are not fixed in advance but chosen in a way to minimize a penalized sum of squared errors criterion.  We thoroughly compare this technique to a classical regression method, local polynomial smoothing, via simulation studies and application to the analysis of inner carotid artery centerlines (AneuRisk Project dataset).  We show that 3D free-knot regression splines yield more accurate and efficient estimates.

 Joint work with Piercesare Secchi and Simone Vantini.

Placeholder