Hasil untuk "stat.ML"

Menampilkan 20 dari ~159327 hasil · dari arXiv, DOAJ, CrossRef

JSON API
CrossRef Open Access 2024
Why "ML in Health Science"

Yury Rusinovich

This is the first editorial of the journal, discussing the balance between humans and AI in healthcare and emphasizing the need for a human-centric approach in AI and ML applications.

1 sitasi en
arXiv Open Access 2023
Tensor Products and Hyperdimensional Computing

Frank Qiu

Following up on a previous analysis of graph embeddings, we generalize and expand some results to the general setting of vector symbolic architectures (VSA) and hyperdimensional computing (HDC). Importantly, we explore the mathematical relationship between superposition, orthogonality, and tensor product. We establish the tensor product representation as the central representation, with a suite of unique properties. These include it being the most general and expressive representation, as well as being the most compressed representation that has errorrless unbinding and detection.

en stat.ML, cs.LG
arXiv Open Access 2023
Quantile and moment neural networks for learning functionals of distributions

Xavier Warin

We study news neural networks to approximate function of distributions in a probability space. Two classes of neural networks based on quantile and moment approximation are proposed to learn these functions and are theoretically supported by universal approximation theorems. By mixing the quantile and moment features in other new networks, we develop schemes that outperform existing networks on numerical test cases involving univariate distributions. For bivariate distributions, the moment neural network outperforms all other networks.

en stat.ML, cs.LG
arXiv Open Access 2022
Optimal Ratio for Data Splitting

V. Roshan Joseph

It is common to split a dataset into training and testing sets before fitting a statistical or machine learning model. However, there is no clear guidance on how much data should be used for training and testing. In this article we show that the optimal splitting ratio is $\sqrt{p}:1$, where $p$ is the number of parameters in a linear regression model that explains the data well.

en stat.ML, cs.LG
arXiv Open Access 2022
Nonlinear MCMC for Bayesian Machine Learning

James Vuckovic

We explore the application of a nonlinear MCMC technique first introduced in [1] to problems in Bayesian machine learning. We provide a convergence guarantee in total variation that uses novel results for long-time convergence and large-particle ("propagation of chaos") convergence. We apply this nonlinear MCMC technique to sampling problems including a Bayesian neural network on CIFAR10.

en stat.ML, cs.LG
arXiv Open Access 2017
An Empirical Bayes Approach for High Dimensional Classification

Yunbo Ouyang, Feng Liang

We propose an empirical Bayes estimator based on Dirichlet process mixture model for estimating the sparse normalized mean difference, which could be directly applied to the high dimensional linear classification. In theory, we build a bridge to connect the estimation error of the mean difference and the misclassification error, also provide sufficient conditions of sub-optimal classifiers and optimal classifiers. In implementation, a variational Bayes algorithm is developed to compute the posterior efficiently and could be parallelized to deal with the ultra-high dimensional case.

en stat.ML, stat.ME
arXiv Open Access 2017
A minimax and asymptotically optimal algorithm for stochastic bandits

Pierre Ménard, Aurélien Garivier

We propose the kl-UCB ++ algorithm for regret minimization in stochastic bandit models with exponential families of distributions. We prove that it is simultaneously asymptotically optimal (in the sense of Lai and Robbins' lower bound) and minimax optimal. This is the first algorithm proved to enjoy these two properties at the same time. This work thus merges two different lines of research with simple and clear proofs.

en stat.ML, cs.LG
arXiv Open Access 2017
On conditional parity as a notion of non-discrimination in machine learning

Ya'acov Ritov, Yuekai Sun, Ruofei Zhao

We identify conditional parity as a general notion of non-discrimination in machine learning. In fact, several recently proposed notions of non-discrimination, including a few counterfactual notions, are instances of conditional parity. We show that conditional parity is amenable to statistical analysis by studying randomization as a general mechanism for achieving conditional parity and a kernel-based test of conditional parity.

en stat.ML, cs.CY
arXiv Open Access 2017
YouTube-8M Video Understanding Challenge Approach and Applications

Edward Chen

This paper introduces the YouTube-8M Video Understanding Challenge hosted as a Kaggle competition and also describes my approach to experimenting with various models. For each of my experiments, I provide the score result as well as possible improvements to be made. Towards the end of the paper, I discuss the various ensemble learning techniques that I applied on the dataset which significantly boosted my overall competition score. At last, I discuss the exciting future of video understanding research and also the many applications that such research could significantly improve.

en stat.ML
arXiv Open Access 2016
Modeling and Estimation of Discrete-Time Reciprocal Processes via Probabilistic Graphical Models

Francesca Paola Carli

Reciprocal processes are acausal generalizations of Markov processes introduced by Bernstein in 1932. In the literature, a significant amount of attention has been focused on developing dynamical models for reciprocal processes. In this paper, we provide a probabilistic graphical model for reciprocal processes. This leads to a principled solution of the smoothing problem via message passing algorithms. For the finite state space case, convergence analysis is revisited via the Hilbert metric.

en stat.ML, math.OC
arXiv Open Access 2016
A Simple Practical Accelerated Method for Finite Sums

Aaron Defazio

We describe a novel optimization method for finite sums (such as empirical risk minimization problems) building on the recently introduced SAGA method. Our method achieves an accelerated convergence rate on strongly convex smooth problems. Our method has only one parameter (a step size), and is radically simpler than other accelerated methods for finite sums. Additionally it can be applied when the terms are non-smooth, yielding a method applicable in many areas where operator splitting methods would traditionally be applied.

en stat.ML, cs.LG
arXiv Open Access 2016
Conditional distribution variability measures for causality detection

José A. R. Fonollosa

In this paper we derive variability measures for the conditional probability distributions of a pair of random variables, and we study its application in the inference of causal-effect relationships. We also study the combination of the proposed measures with standard statistical measures in the the framework of the ChaLearn cause-effect pair challenge. The developed model obtains an AUC score of 0.82 on the final test database and ranked second in the challenge.

en stat.ML, cs.LG
arXiv Open Access 2015
Modularity Component Analysis versus Principal Component Analysis

Hansi Jiang, Carl Meyer

In this paper the exact linear relation between the leading eigenvectors of the modularity matrix and the singular vectors of an uncentered data matrix is developed. Based on this analysis the concept of a modularity component is defined, and its properties are developed. It is shown that modularity component analysis can be used to cluster data similar to how traditional principal component analysis is used except that modularity component analysis does not require data centering.

arXiv Open Access 2015
How to show a probabilistic model is better

Mithun Chakraborty, Sanmay Das, Allen Lavoie

We present a simple theoretical framework, and corresponding practical procedures, for comparing probabilistic models on real data in a traditional machine learning setting. This framework is based on the theory of proper scoring rules, but requires only basic algebra and probability theory to understand and verify. The theoretical concepts presented are well-studied, primarily in the statistics literature. The goal of this paper is to advocate their wider adoption for performance evaluation in empirical machine learning.

en stat.ML, cs.LG
arXiv Open Access 2015
A tree augmented naive Bayesian network experiment for breast cancer prediction

Ping Ren

In order to investigate the breast cancer prediction problem on the aging population with the grades of DCIS, we conduct a tree augmented naive Bayesian network experiment trained and tested on a large clinical dataset including consecutive diagnostic mammography examinations, consequent biopsy outcomes and related cancer registry records in the population of women across all ages. The aggregated results of our ten-fold cross validation method recommend a biopsy threshold higher than 2% for the aging population.

en stat.ML, q-bio.QM
arXiv Open Access 2015
Emphatic TD Bellman Operator is a Contraction

Assaf Hallak, Aviv Tamar, Shie Mannor

Recently, \citet{SuttonMW15} introduced the emphatic temporal differences (ETD) algorithm for off-policy evaluation in Markov decision processes. In this short note, we show that the projected fixed-point equation that underlies ETD involves a contraction operator, with a $\sqrtγ$-contraction modulus (where $γ$ is the discount factor). This allows us to provide error bounds on the approximation error of ETD. To our knowledge, these are the first error bounds for an off-policy evaluation algorithm under general target and behavior policies.

en stat.ML, cs.LG
arXiv Open Access 2015
Relations Between Adjacency and Modularity Graph Partitioning

Hansi Jiang, Carl Meyer

This paper develops the exact linear relationship between the leading eigenvector of the unnormalized modularity matrix and the eigenvectors of the adjacency matrix. We propose a method for approximating the leading eigenvector of the modularity matrix, and we derive the error of the approximation. There is also a complete proof of the equivalence between normalized adjacency clustering and normalized modularity clustering. Numerical experiments show that normalized adjacency clustering can be as twice efficient as normalized modularity clustering.

arXiv Open Access 2013
bartMachine: Machine Learning with Bayesian Additive Regression Trees

Adam Kapelner, Justin Bleich

We present a new package in R implementing Bayesian additive regression trees (BART). The package introduces many new features for data analysis using BART such as variable selection, interaction detection, model diagnostic plots, incorporation of missing data and the ability to save trees for future prediction. It is significantly faster than the current R implementation, parallelized, and capable of handling both large sample sizes and high-dimensional data.

en stat.ML, cs.LG
arXiv Open Access 2013
The Gaussian Radon Transform and Machine Learning

Irina Holmes, Ambar Sengupta

There has been growing recent interest in probabilistic interpretations of kernel-based methods as well as learning in Banach spaces. The absence of a useful Lebesgue measure on an infinite-dimensional reproducing kernel Hilbert space is a serious obstacle for such stochastic models. We propose an estimation model for the ridge regression problem within the framework of abstract Wiener spaces and show how the support vector machine solution to such problems can be interpreted in terms of the Gaussian Radon transform.

en stat.ML, math.FA
arXiv Open Access 2012
A Bayesian Boosting Model

Alexander Lorbert, David M. Blei, Robert E. Schapire et al.

We offer a novel view of AdaBoost in a statistical setting. We propose a Bayesian model for binary classification in which label noise is modeled hierarchically. Using variational inference to optimize a dynamic evidence lower bound, we derive a new boosting-like algorithm called VIBoost. We show its close connections to AdaBoost and give experimental results from four datasets.

en stat.ML

Halaman 7 dari 7967