Hasil untuk "stat.ML"

Menampilkan 20 dari ~159325 hasil Β· dari arXiv, CrossRef, DOAJ

JSON API
arXiv Open Access 2024
Analysis of the rate of convergence of an over-parametrized convolutional neural network image classifier learned by gradient descent

Michael Kohler, Adam Krzyzak, Benjamin Walter

Image classification based on over-parametrized convolutional neural networks with a global average-pooling layer is considered. The weights of the network are learned by gradient descent. A bound on the rate of convergence of the difference between the misclassification risk of the newly introduced convolutional neural network estimate and the minimal possible value is derived.

en stat.ML, cs.LG
arXiv Open Access 2021
Use of High Dimensional Modeling for automatic variables selection: the best path algorithm

Luigi Riso

This paper presents a new algorithm for automatic variables selection. In particular, using the Graphical Models properties it is possible to develop a method that can be used in the contest of large dataset. The advantage of this algorithm is that can be combined with different forecasting models. In this research we have used the OLS method and we have compared the result with the LASSO method.

en stat.ML, cs.LG
arXiv Open Access 2020
Rejoinder: New Objectives for Policy Learning

Nathan Kallus

I provide a rejoinder for discussion of "More Efficient Policy Learning via Optimal Retargeting" to appear in the Journal of the American Statistical Association with discussion by Oliver Dukes and Stijn Vansteelandt; Sijia Li, Xiudi Li, and Alex Luedtkeand; and Muxuan Liang and Yingqi Zhao.

en stat.ML, cs.LG
arXiv Open Access 2020
Minimum Relative Entropy Inference for Normal and Monte Carlo Distributions

Marcello Colasante, Attilio Meucci

We represent affine sub-manifolds of exponential family distributions as minimum relative entropy sub-manifolds. With such representation we derive analytical formulas for the inference from partial information on expectations and covariances of multivariate normal distributions; and we improve the numerical implementation via Monte Carlo simulations for the inference from partial information of generalized expectation type.

en stat.ML, cs.LG
arXiv Open Access 2019
The Blessings of Multiple Causes: A Reply to Ogburn et al. (2019)

Yixin Wang, David M. Blei

Ogburn et al. (2019, arXiv:1910.05438) discuss "The Blessings of Multiple Causes" (Wang and Blei, 2018, arXiv:1805.06826). Many of their remarks are interesting. But they also claim that the paper has "foundational errors" and that its "premise is...incorrect." These claims are not substantiated. There are no foundational errors; the premise is correct.

en stat.ML, cs.LG
arXiv Open Access 2019
Non-asymptotic error bounds for scaled underdamped Langevin MCMC

Tim Zajic

Recent works have derived non-asymptotic upper bounds for convergence of underdamped Langevin MCMC. We revisit these bound and consider introducing scaling terms in the underlying underdamped Langevin equation. In particular, we provide conditions under which an appropriate scaling allows to improve the error bounds in terms of the condition number of the underlying density of interest.

en stat.ML, cs.LG
arXiv Open Access 2018
From Principal Subspaces to Principal Components with Linear Autoencoders

Elad Plaut

The autoencoder is an effective unsupervised learning model which is widely used in deep learning. It is well known that an autoencoder with a single fully-connected hidden layer, a linear activation function and a squared error cost function trains weights that span the same subspace as the one spanned by the principal component loading vectors, but that they are not identical to the loading vectors. In this paper, we show how to recover the loading vectors from the autoencoder weights.

en stat.ML, cs.LG
arXiv Open Access 2017
Kernel Conditional Exponential Family

Michael Arbel, Arthur Gretton

A nonparametric family of conditional distributions is introduced, which generalizes conditional exponential families using functional parameters in a suitable RKHS. An algorithm is provided for learning the generalized natural parameter, and consistency of the estimator is established in the well specified case. In experiments, the new method generally outperforms a competing approach with consistency guarantees, and is competitive with a deep conditional density model on datasets that exhibit abrupt transitions and heteroscedasticity.

en stat.ML
arXiv Open Access 2017
Comparison of PCA with ICA from data distribution perspective

Miron Ivanov

We performed an empirical comparison of ICA and PCA algorithms by applying them on two simulated noisy time series with varying distribution parameters and level of noise. In general, ICA shows better results than PCA because it takes into account higher moments of data distribution. On the other hand, PCA remains quite sensitive to the level of correlations among signals.

en stat.ML, cs.LG
arXiv Open Access 2017
Structured Sparse Modelling with Hierarchical GP

Danil Kuzin, Olga Isupova, Lyudmila Mihaylova

In this paper a new Bayesian model for sparse linear regression with a spatio-temporal structure is proposed. It incorporates the structural assumptions based on a hierarchical Gaussian process prior for spike and slab coefficients. We design an inference algorithm based on Expectation Propagation and evaluate the model over the real data.

en stat.ML
arXiv Open Access 2016
A note on the statistical view of matrix completion

Tianxi Li

A very simple interpretation of matrix completion problem is introduced based on statistical models. Combined with the well-known results from missing data analysis, such interpretation indicates that matrix completion is still a valid and principled estimation procedure even without the missing completely at random (MCAR) assumption, which almost all of the current theoretical studies of matrix completion assume.

en stat.ML
arXiv Open Access 2016
Gradient Descent Converges to Minimizers

Jason D. Lee, Max Simchowitz, Michael I. Jordan et al.

We show that gradient descent converges to a local minimizer, almost surely with random initialization. This is proved by applying the Stable Manifold Theorem from dynamical systems theory.

en stat.ML, cs.LG
arXiv Open Access 2015
k-Means Clustering Is Matrix Factorization

Christian Bauckhage

We show that the objective function of conventional k-means clustering can be expressed as the Frobenius norm of the difference of a data matrix and a low rank approximation of that data matrix. In short, we show that k-means clustering is a matrix factorization problem. These notes are meant as a reference and intended to provide a guided tour towards a result that is often mentioned but seldom made explicit in the literature.

en stat.ML
arXiv Open Access 2014
Scalable Variational Gaussian Process Classification

James Hensman, Alex Matthews, Zoubin Ghahramani

Gaussian process classification is a popular method with a number of appealing properties. We show how to scale the model within a variational inducing point framework, outperforming the state of the art on benchmark datasets. Importantly, the variational formulation can be exploited to allow classification in problems with millions of data points, as we demonstrate in experiments.

en stat.ML

Halaman 4 dari 7967