Statistical Inference Via Convex Optimization

Statistical Inference Via Convex Optimization
Author: Anatoli Juditsky
Publisher: Princeton University Press
Total Pages: 655
Release: 2020-04-07
Genre: Mathematics
ISBN: 0691197296

Download Statistical Inference Via Convex Optimization Book in PDF, Epub and Kindle

This authoritative book draws on the latest research to explore the interplay of high-dimensional statistics with optimization. Through an accessible analysis of fundamental problems of hypothesis testing and signal recovery, Anatoli Juditsky and Arkadi Nemirovski show how convex optimization theory can be used to devise and analyze near-optimal statistical inferences. Statistical Inference via Convex Optimization is an essential resource for optimization specialists who are new to statistics and its applications, and for data scientists who want to improve their optimization methods. Juditsky and Nemirovski provide the first systematic treatment of the statistical techniques that have arisen from advances in the theory of optimization. They focus on four well-known statistical problems—sparse recovery, hypothesis testing, and recovery from indirect observations of both signals and functions of signals—demonstrating how they can be solved more efficiently as convex optimization problems. The emphasis throughout is on achieving the best possible statistical performance. The construction of inference routines and the quantification of their statistical performance are given by efficient computation rather than by analytical derivation typical of more conventional statistical approaches. In addition to being computation-friendly, the methods described in this book enable practitioners to handle numerous situations too difficult for closed analytical form analysis, such as composite hypothesis testing and signal recovery in inverse problems. Statistical Inference via Convex Optimization features exercises with solutions along with extensive appendixes, making it ideal for use as a graduate text.

High-dimensional Statistical Inference from Coarse and Nonlinear Data

High-dimensional Statistical Inference from Coarse and Nonlinear Data
Author: Haoyu Fu
Publisher:
Total Pages: 142
Release: 2019
Genre: Machine learning
ISBN:

Download High-dimensional Statistical Inference from Coarse and Nonlinear Data Book in PDF, Epub and Kindle

Moving to the context of machine learning, we study several one-hidden-layer neural network models for nonlinear regression using both cross-entropy and least-squares loss functions. The neural-network-based models have attracted a significant amount of research interest due to the success of deep learning in practical domains such as computer vision and natural language processing. Learning such neural-network-based models often requires solving a non-convex optimization problem. We propose different strategies to characterize the optimization landscape of the non-convex loss functions and provide guarantees on the statistical and computational efficiency of optimizing these loss functions via gradient descent.

Learning Theory

Learning Theory
Author: Hans Ulrich Simon
Publisher: Springer
Total Pages: 667
Release: 2006-09-29
Genre: Computers
ISBN: 3540352961

Download Learning Theory Book in PDF, Epub and Kindle

This book constitutes the refereed proceedings of the 19th Annual Conference on Learning Theory, COLT 2006, held in Pittsburgh, Pennsylvania, USA, June 2006. The book presents 43 revised full papers together with 2 articles on open problems and 3 invited lectures. The papers cover a wide range of topics including clustering, un- and semi-supervised learning, statistical learning theory, regularized learning and kernel methods, query learning and teaching, inductive inference, and more.

Fast Randomized Algorithms for Convex Optimization and Statistical Estimation

Fast Randomized Algorithms for Convex Optimization and Statistical Estimation
Author: Mert Pilanci
Publisher:
Total Pages: 234
Release: 2016
Genre:
ISBN:

Download Fast Randomized Algorithms for Convex Optimization and Statistical Estimation Book in PDF, Epub and Kindle

With the advent of massive datasets, statistical learning and information processing techniques are expected to enable exceptional possibilities for engineering, data intensive sciences and better decision making. Unfortunately, existing algorithms for mathematical optimization, which is the core component in these techniques, often prove ineffective for scaling to the extent of all available data. In recent years, randomized dimension reduction has proven to be a very powerful tool for approximate computations over large datasets. In this thesis, we consider random projection methods in the context of general convex optimization problems on massive datasets. We explore many applications in machine learning, statistics and decision making and analyze various forms of randomization in detail. The central contributions of this thesis are as follows: (i) We develop random projection methods for convex optimization problems and establish fundamental trade-offs between the size of the projection and accuracy of solution in convex optimization. (ii) We characterize information-theoretic limitations of methods that are based on random projection, which surprisingly shows that the most widely used form of random projection is, in fact, statistically sub-optimal. (iii) We present novel methods, which iteratively refine the solutions to achieve statistical optimality and enable solving large scale optimization and statistical inference problems orders-of-magnitude faster than existing methods. (iv) We develop new randomized methodologies for relaxing cardinality constraints in order to obtain checkable and more accurate approximations than the state of the art approaches.

Robust Optimization

Robust Optimization
Author: Aharon Ben-Tal
Publisher: Princeton University Press
Total Pages: 565
Release: 2009-08-10
Genre: Mathematics
ISBN: 1400831059

Download Robust Optimization Book in PDF, Epub and Kindle

Robust optimization is still a relatively new approach to optimization problems affected by uncertainty, but it has already proved so useful in real applications that it is difficult to tackle such problems today without considering this powerful methodology. Written by the principal developers of robust optimization, and describing the main achievements of a decade of research, this is the first book to provide a comprehensive and up-to-date account of the subject. Robust optimization is designed to meet some major challenges associated with uncertainty-affected optimization problems: to operate under lack of full information on the nature of uncertainty; to model the problem in a form that can be solved efficiently; and to provide guarantees about the performance of the solution. The book starts with a relatively simple treatment of uncertain linear programming, proceeding with a deep analysis of the interconnections between the construction of appropriate uncertainty sets and the classical chance constraints (probabilistic) approach. It then develops the robust optimization theory for uncertain conic quadratic and semidefinite optimization problems and dynamic (multistage) problems. The theory is supported by numerous examples and computational illustrations. An essential book for anyone working on optimization and decision making under uncertainty, Robust Optimization also makes an ideal graduate textbook on the subject.

Information Theory, Inference and Learning Algorithms

Information Theory, Inference and Learning Algorithms
Author: David J. C. MacKay
Publisher: Cambridge University Press
Total Pages: 694
Release: 2003-09-25
Genre: Computers
ISBN: 9780521642989

Download Information Theory, Inference and Learning Algorithms Book in PDF, Epub and Kindle

Information theory and inference, taught together in this exciting textbook, lie at the heart of many important areas of modern technology - communication, signal processing, data mining, machine learning, pattern recognition, computational neuroscience, bioinformatics and cryptography. The book introduces theory in tandem with applications. Information theory is taught alongside practical communication systems such as arithmetic coding for data compression and sparse-graph codes for error-correction. Inference techniques, including message-passing algorithms, Monte Carlo methods and variational approximations, are developed alongside applications to clustering, convolutional codes, independent component analysis, and neural networks. Uniquely, the book covers state-of-the-art error-correcting codes, including low-density-parity-check codes, turbo codes, and digital fountain codes - the twenty-first-century standards for satellite communications, disk drives, and data broadcast. Richly illustrated, filled with worked examples and over 400 exercises, some with detailed solutions, the book is ideal for self-learning, and for undergraduate or graduate courses. It also provides an unparalleled entry point for professionals in areas as diverse as computational biology, financial engineering and machine learning.

Statistical Inference and Optimization for Low-rank Matrix and Tensor Learning

Statistical Inference and Optimization for Low-rank Matrix and Tensor Learning
Author: Yuetian Luo (Ph.D.)
Publisher:
Total Pages: 0
Release: 2022
Genre:
ISBN:

Download Statistical Inference and Optimization for Low-rank Matrix and Tensor Learning Book in PDF, Epub and Kindle

High dimensional statistical problems with matrix or tensor type data are ubiquitous in modern data analysis. In many applications, the dimension of the matrix or tensor is high and much bigger than the sample size and some structural assumptions are often imposed to ensure the problem is well-posed. One of the most popular structures in matrix and tensor data analysis is the low-rankness. In this thesis, we make contributions to the statistical inference and optimization in low-rank matrix and tensor data analysis from the following three aspects. First, first-order algorithms have been the workhorse in modern data analysis, including matrix and tensor problems, for their simplicity and efficiency. Second-order algorithms suffer from high computational costs and instability. The first part of the thesis explores the following question: can we develop provable efficient second-order algorithms for high-dimensional matrix and tensor problems with low-rank structures? We provide a positive answer to this question, where the key idea is to explore smooth Riemannian structures of the sets of low-rank matrices and tensors and the connection to the second-order Riemannian optimization methods. In particular, we demonstrate that for a large class of tensor-on-tensor regression problems, the Riemannian Gauss-Newton algorithm is computationally fast and achieves provable second-order convergence. We also discuss the case when the intrinsic rank of the parameter matrix/tensor is unknown and a natural rank overspecification is implemented. In the second part of the thesis, we explore an interesting question: is there any connection between different non-convex optimization approaches for solving the general low-rank matrix optimization? We find from a geometric point of view, the common non-convex factorization formulation has a close connection with the Riemannian formulation and there exists an equivalence between them. Moreover, we discover that two notable Riemannian formulations, i.e., formulations under Riemannian embedded and quotient geometries, are also closely related from a geometric point of view. In the final part of the thesis, we are dedicated to studying one intriguing phenomenon in high dimensional statistical problems, statistical and computational trade-offs, which refers to the commonly appearing gaps between the different signal-to-noise ratio thresholds that make the problem information-theoretically solvable or polynomial-time solvable. Here we focus on the statistical-computational trade-offs induced by tensor structures. We would provide rigorous evidence for the computational barriers to two important classes of problems: tensor clustering and tensor regression. We show these computational limits by the average-case reduction and restricted class of low-degree polynomials arguments.

Computer Age Statistical Inference, Student Edition

Computer Age Statistical Inference, Student Edition
Author: Bradley Efron
Publisher: Cambridge University Press
Total Pages: 514
Release: 2021-06-17
Genre: Mathematics
ISBN: 1108915876

Download Computer Age Statistical Inference, Student Edition Book in PDF, Epub and Kindle

The twenty-first century has seen a breathtaking expansion of statistical methodology, both in scope and influence. 'Data science' and 'machine learning' have become familiar terms in the news, as statistical methods are brought to bear upon the enormous data sets of modern science and commerce. How did we get here? And where are we going? How does it all fit together? Now in paperback and fortified with exercises, this book delivers a concentrated course in modern statistical thinking. Beginning with classical inferential theories - Bayesian, frequentist, Fisherian - individual chapters take up a series of influential topics: survival analysis, logistic regression, empirical Bayes, the jackknife and bootstrap, random forests, neural networks, Markov Chain Monte Carlo, inference after model selection, and dozens more. The distinctly modern approach integrates methodology and algorithms with statistical inference. Each chapter ends with class-tested exercises, and the book concludes with speculation on the future direction of statistics and data science.

Distributed Optimization and Statistical Learning Via the Alternating Direction Method of Multipliers

Distributed Optimization and Statistical Learning Via the Alternating Direction Method of Multipliers
Author: Stephen Boyd
Publisher: Now Publishers Inc
Total Pages: 138
Release: 2011
Genre: Computers
ISBN: 160198460X

Download Distributed Optimization and Statistical Learning Via the Alternating Direction Method of Multipliers Book in PDF, Epub and Kindle

Surveys the theory and history of the alternating direction method of multipliers, and discusses its applications to a wide variety of statistical and machine learning problems of recent interest, including the lasso, sparse logistic regression, basis pursuit, covariance selection, support vector machines, and many others.

Statistical Inference for Engineers and Data Scientists

Statistical Inference for Engineers and Data Scientists
Author: Pierre Moulin
Publisher: Cambridge University Press
Total Pages: 423
Release: 2019
Genre: Mathematics
ISBN: 1107185920

Download Statistical Inference for Engineers and Data Scientists Book in PDF, Epub and Kindle

A mathematically accessible textbook introducing all the tools needed to address modern inference problems in engineering and data science.