Research Seminar

Would you like to be notified about these presentations via e-mail? Please subscribe here.

×

Modal title

Modal content

Spring Semester 2011

Date / Time Speaker Title Location
11 March 2011
15:15-16:30
Elvezio Ronchetti
University of Geneva
Event Details

Research Seminar in Statistics

Title Accurate and Robust Indirect Inference
Speaker, Affiliation Elvezio Ronchetti, University of Geneva
Date, Time 11 March 2011, 15:15-16:30
Location HG G 19.1
Abstract Indirect inference provides a broad class of estimators and testing procedures that can be used to carry out inference in complex models, where e.g. the likelihood function is not available in closed form. These techniques have now been successfully applied in a variety of fields, including engineering, biostatistics, and finance. Given a model and the data, an estimate of the parameter of an auxiliary (simpler) model is first obtained. Then, pseudo-data are simulated from the original (complex) model and the auxiliary estimate is computed on the pseudo-data. Finally, the estimate of the parameter of the original model is obtained by minimizing a distance between the auxiliary estimates computed on the data and on the pseudo-data. In this talk we address two important issues. First it is known that classical (especially over-identification) tests based on the asymptotic theory have a poor finite sample accuracy. Therefore, we introduce new accurate parameter and over-identification tests for indirect inference which exhibit an excellent finite sample behavior. Secondly, we address the robustness issue of these procedures by providing estimators and tests for indirect inference which are not unduly influenced by small deviations from the assumed model. By combining these two properties, we obtain accurate and reliable procedures for indirect inference. The theoretical results are illustrated in various models, including nonlinear regression, Poisson regression with over-dispersion, and diffusion models. Joint work with Veronika Czellar, HEC Paris.
Accurate and Robust Indirect Inferenceread_more
HG G 19.1
1 April 2011
15:15-16:30
Nicolas Vayatis
CMLA de Cachan
Event Details

Research Seminar in Statistics

Title Nonparametric scoring and ranking trees
Speaker, Affiliation Nicolas Vayatis, CMLA de Cachan
Date, Time 1 April 2011, 15:15-16:30
Location HG G 19.1
Abstract In a world of overwhelming information, ranking has become one of the most critical tasks for high level data processing. In the talk, we focus on the problem of learning to rank high dimensional vectors based on past data with binary feedback, say positive and negative. From a mathematical point of view, the problem is to find a real-valued scoring function which leads to the highest possible density of positive observations among the highest values of the scoring function. We discuss the nature of optimal scoring functions, performance measures, statistical aspects and practical algorithms inspired from decision trees and random forests.
Nonparametric scoring and ranking treesread_more
HG G 19.1
* 5 April 2011
16:00-17:15
Volkan Cevher
EPFL
Event Details

Research Seminar in Statistics

Title Compressible priors for high-dimensional statistics
Speaker, Affiliation Volkan Cevher, EPFL
Date, Time 5 April 2011, 16:00-17:15
Location HG E 22
Abstract Special Talk, invited and organized by Prof. Andreas Krause, Learning and Adaptive Systems, Department of Computer Science, ETH Zurich We develop a principled way of identifying probability distributions whose independent and identically distributed (iid) realizations are compressible, i.e., can be approximated as sparse. We focus on the context of Gaussian random underdetermined linear regression (GULR) problems, where compressibility is known to ensure the success of estimators exploiting sparse regularization. We prove that many of the conventional priors revolving around probabilistic interpretations of the p-norm (p<=1) regularization algorithms are in fact incompressible in the limit of large problem sizes. To show this, we identify nontrivial undersampling regions in GULR where the simple least squares solution almost surely outperforms an oracle sparse solution, when the data is generated from a prior such as the Laplace distribution. We provide rules of thumb to characterize large families of compressible and incompressible priors based on their second and fourth moments. Generalized Gaussians and generalized Pareto distributions serve as running examples for concreteness. We then conclude with a study of the statistics of wavelet coefficients of natural images in the context of compressible priors. Bio: Prof. Volkan Cevher received his BSc degree (valedictorian) in Electrical Engineering from Bilkent University in 1999, and his PhD degree in Electrical and Computer Engineering from Georgia Institute of Technology in 2005. He held Research Scientist positions at University of Maryland, College Park during 2006-2007 and at Rice University during 2008-2009. Currently, he is an Assistant Professor at Ecole Polytechnique Federale de Lausanne with joint appointment at the Idiap Research Institute and a Faculty Fellow at Rice University. His research interests include signal processing theory, machine learning, graphical models, and information theory. Time and location: April 5 2011, 4pm in HG E 22 More information: Andreas Krause, http://las.ethz.ch Volkan Cevher, http://lions.epfl.ch
Compressible priors for high-dimensional statisticsread_more
HG E 22
8 April 2011
15:15-16:30
Stanislav Anatolyev
New Economic School, Moscow
Event Details

Research Seminar in Statistics

Title Sequential testing with uniformly distributed size
Speaker, Affiliation Stanislav Anatolyev, New Economic School, Moscow
Date, Time 8 April 2011, 15:15-16:30
Location HG G 19.1
Abstract Sequential procedures of testing for structural stability do not provide enough guidance on the shape of boundaries that are used to decide on acceptance or rejection, requiring only that the overall size of the test is asymptotically controlled. We introduce and motivate a reasonable criterion for a shape of boundaries which requires that the test size be uniformly distributed over the testing period. Under this criterion, we numerically construct boundaries for most popular sequential tests that are characterized by a test statistic behaving asymptotically either as a Wiener process or Brownian bridge. We handle this problem both in a context of retrospecting a historical sample and in a context of monitoring newly arriving data. We tabulate the boundaries by Ötting them to certain áexible but parsimonious functional forms. Interesting patterns emerge in an illustrative application of sequential tests to the Phillips curve model. Key words: Structural stability; sequential tests; CUSUM; retrospection; monitoring; boundaries; asymptotic size. _
Sequential testing with uniformly distributed sizeread_more
HG G 19.1
* 13 May 2011
15:15-16:30
Michael Chichignoud
Université de Provence, Marseille
Event Details

Research Seminar in Statistics

Title Statistical performances of a bayesian estimator
Speaker, Affiliation Michael Chichignoud, Université de Provence, Marseille
Date, Time 13 May 2011, 15:15-16:30
Location HG G 19.2
Abstract We study the nonparametric regression model ((X1, Y1),...(Xn ,Yn)), where (Xi)i is the deterministic design and (Yi)i is a sequence of independent variables. Assume that the density Yi is known and can be written as g(., f(Xi)), which depends on a regression function f at the point Xi. This function is assumed smooth, i.e. belonging at a Hölder ball. The aim is to estimate the regression function at a given point y from the observations (pointwise estimation) and to find the optimal estimator (in the sense of rates of convergence) for each density g. We use the locally parametric approach to construct a new local bayesian estimator. Under some conditions on g, we propose an adaptive procedure based on the so-called Lepski’s method (adaptive selection of the bandwidth) which allows us to construct an optimal adaptive bayesian estimator.
Statistical performances of a bayesian estimatorread_more
HG G 19.2
* 26 May 2011
16:15-17:30
Robert G. Staudte
La Trobe University, Melbourne
Event Details

Research Seminar in Statistics

Title The relationship between variance stabilized statistics and the Kullback-Leibler symmetrized divergence.
Speaker, Affiliation Robert G. Staudte, La Trobe University, Melbourne
Date, Time 26 May 2011, 16:15-17:30
Location HG G 19.1
Abstract In the beginning was R.A. Fisher, who created variance stabilization and saw that it was good. Many statistical descendants have found it to be a powerful tool in applications. A recent book Kulinskaya, Morgenthaler and Staudte (Wiley, 2008) formalized this powerful tool to define statistical evidence as the transformed statistic, because it allowed for simple calibration and interpretation, as well as combination of evidence from different studies in a metaanalysis. It turns out that the Key function of their book that arises in variance stabilization has a remarkably strong link to the Kullback-Leibler symmetrized divergence. This result leads to some surprising global approximations and new variance stabilized statistics.
The relationship between variance stabilized statistics and the Kullback-Leibler symmetrized divergence.read_more
HG G 19.1
* 16 June 2011
15:15-16:30
Jonas Peters
Max-Planck-Campus, Tübingen
Event Details

Research Seminar in Statistics

Title Causal Inference using Identifiable Functional Model Classes.
Speaker, Affiliation Jonas Peters, Max-Planck-Campus, Tübingen
Date, Time 16 June 2011, 15:15-16:30
Location HG G 19.1
Abstract This work addresses the following question: Under what assumptions on the data generating process can one infer the causal graph from the joint distribution? Constraint-based methods like the PC algorithm assume the Markov condition and faithfulness. These two conditions relate conditional independences and the graph structure, which allows to infer properties of the graph from conditional independences that can be found in the joint distribution. These methods, however, encounter the following difficulties: (1) One can discover causal structures only up to Markov equivalence classes, in particular one cannot distinguish between X -> Y and Y -> X. (2) Conditional independence testing is very difficult in practice. (3) When the process is not faithful, the results may be wrong, but the user does not realize it. We propose an alternative by defining dentifiable Functional Model classes (IFMOCs) and provide the example of additive noise models with additional constraints (e.g. X3=f(X1,X2)+N, but N should not be Gaussian when f is linear). Based on these classes we develop a causal inference method that overcomes some of the difficulties from before: (1) One can identify causal relationships even within an equivalence class. (2)Intuitively, fitting the model is in a sense easier than conditional independence testing. (3) We do not require faithfulness, but rather impose a model class on the data. When the model assumptions are violated, however, (e.g. the data do not follow the considered IFMOC or some of the variables are unobserved), the method would output "I do not know" rather than giving wrong answers.We regard our work as being theoretical. Although results on simulated data and on some real world data sets look promising, extensive experiments on real systems are necessary to verify the proposed principles.
Causal Inference using Identifiable Functional Model Classes.read_more
HG G 19.1

Notes: events marked with an asterisk (*) indicate that the time and/or location are different from the usual time and/or location and if you want you can subscribe to the iCal/ics Calender.

Organizers: Peter Bühlmann, Reinhard Furrer, Leonhard Held, Hans-Rudolf Künsch, Marloes Maathuis, Sara van de Geer, Michael Wolf

JavaScript has been disabled in your browser