A survey of dimensionality reduction techniques
C. Sorzano, Javier Vargas, A. Pascual-Montano
Experimental life sciences like biology or chemistry have seen in the recent decades an explosion of the data available from experiments. Laboratory instruments become more and more complex and report hundreds or thousands measurements for a single experiment and therefore the statistical methods face challenging tasks when dealing with such high dimensional data. However, much of the data is highly redundant and can be efficiently brought down to a much smaller number of variables without a significant loss of information. The mathematical procedures making possible this reduction are called dimensionality reduction techniques; they have widely been developed by fields like Statistics or Machine Learning, and are currently a hot research topic. In this review we categorize the plethora of dimension reduction techniques available and give the mathematical insight behind them.
437 sitasi
en
Computer Science, Mathematics
Demand response flexibility and flexibility potential of residential smart appliances: Experiences from large pilot test in Belgium
R. D’hulst, W. Labeeuw, B. Beusen
et al.
371 sitasi
en
Engineering
Improving Quantum Machine Learning via Heat-Bath Algorithmic Cooling
Nayeli A. Rodríguez-Briones, Daniel K. Park
This work introduces an approach rooted in quantum thermodynamics to enhance sampling efficiency in quantum machine learning (QML). We propose conceptualizing quantum supervised learning as a thermodynamic cooling process. Building on this concept, we develop a quantum refrigerator protocol that enhances sample efficiency during training and prediction without the need for Grover iterations or quantum phase estimation. Inspired by heat-bath algorithmic cooling protocols, our method alternates entropy compression and thermalization steps to decrease the entropy of qubits, increasing polarization toward the dominant bias. This technique minimizes the computational overhead associated with estimating classification scores and gradients, presenting a practical and efficient solution for QML algorithms compatible with noisy intermediate-scale quantum devices.
Physics, Computer software
Toggling the Defiers to Relax Monotonicity: The Difference-in-Instrumental-Variables Estimand
Johann Caro-Burnett
Standard instrumental variables (IV) methods identify a Local Average Treatment Effect under monotonicity, which rules out defiers. In many empirical environments, however, distinct instruments may induce heterogeneous and even opposing behavioral responses. This paper introduces the Difference-in-Instrumental-Variables (DIIV) estimand, which exploits two instruments with opposing compliance patterns to recover a point-identified and behaviorally interpretable causal effect without imposing monotonicity. The estimand yields a convex combination of the marginal treatment effects on compliers and defiers, with weights reflecting differential shifts in treatment take-up across instruments. When monotonicity holds, DIIV coincides with the standard IV estimand. The approach can be implemented using simple linear transformations and standard two-stage least squares procedures. Applications using replication data illustrate its applicability in practice.
Double Machine Learning of Continuous Treatment Effects with General Instrumental Variables
Shuyuan Chen, Peng Zhang, Yifan Cui
Estimating causal effects of continuous treatments is a common problem in practice, for example, in studying average dose-response functions. Classical analyses typically assume that all confounders are fully observed, whereas in real-world applications, unmeasured confounding often persists. In this article, we propose a novel framework for the identification of average dose-response functions using instrumental variables, thereby mitigating bias induced by unobserved confounders. We introduce the concept of a uniform regular weighting function and consider covering the treatment space with a finite collection of open sets. On each of these sets, such a weighting function exists, allowing us to identify the average dose-response function locally within the corresponding region. For estimation, we propose an augmented inverse probability weighted score for continuous treatments with instrumental variables under a debiased machine learning framework, and provide practical guidance to adaptively establish regular weighting functions from the data. We further establish the asymptotic properties when the average dose-response function is estimated via kernel regression or empirical risk minimization. Finally, we conduct both simulation and empirical studies to assess the finite-sample performance of the proposed methods.
Understanding the role of autoencoders for stiff dynamical systems using information theory
Vijayamanikandan Vijayarangan, Harshavardhana A. Uranakara, Francisco E. Hernández–Pérez
et al.
Using information theory, this study provides insights into how the construction of latent space of autoencoder (AE) using deep neural network (DNN) training finds a smooth (non-stiff) low-dimensional manifold in the stiff dynamical system. Our recent study (Vijayarangan et al. 2023) reported that an AE combined with neural ODE (NODE) as a surrogate reduced order model (ROM) for the integration of stiff chemically reacting systems led to a significant reduction in the temporal stiffness, and the behavior was attributed to the identification of a slow invariant manifold by the nonlinear projection using the AE. The present work offers a fundamental understanding of the mechanism of formation of a non-stiff latent space and stiffness reduction by employing concepts from information theory and better mixing. The learning mechanisms of both the encoder and the decoder are explained by plotting the evolution of mutual information and identifying two different phases. Subsequently, the density distribution is plotted for the physical and latent variables, which shows the transformation of the rare event in the physical space to a highly likely (more probable) event in the latent space provided by the nonlinear autoencoder. Finally, the nonlinear transformation leading to density redistribution is explained using concepts from information theory and probability.
Electrical engineering. Electronics. Nuclear engineering, Computer software
Dynamic local differential privacy location protection method based on critical path
YAN Yan, LIU Kun, ZHANG Yanli
et al.
Location information was recognized as a critical personal data asset in the digital age, offering convenient services while simultaneously posing significant risks of privacy breaches. Local differential privacy models, which do not rely on trusted third parties, had garnered widespread attention. However, significant challenges were identified in existing location protection methods, including the difficulty of adapting spatial partitioning to complex location distributions, along with high communication and computational overhead that limited system efficiency. To address these challenges, a dynamic local differential privacy location protection method based on a critical path was proposed. A spatial index adapted to user distribution density was constructed through non-uniform quadtree spatial partitioning and Hilbert curve traversal, which effectively improved data usability. Subsequently, the proposed critical path encoding mechanism was executed on the server side to compress the complex partition structure into concise path information, thereby reducing communication overhead during parameter transmission. On the user side, the Hilbert index encoding of the user’s region was perturbed using a randomized response mechanism under the local differential privacy model to protect the privacy of the original location. On the server side, the collected perturbed location encodings from users were aggregated and analyzed. Based on the spatiotemporal continuity of location distribution, the proposed spatial partition structure dynamic adjustment strategy was then implemented to efficiently adapt to dynamic changes in user distribution at an extremely low computational cost. Experiments conducted on real-world location datasets demonstrate that the proposed method provides improved location data availability and algorithm runtime efficiency while achieving local differential privacy protection for user locations.
Electronic computers. Computer science
Bayesian Model Averaging in Causal Instrumental Variable Models
Gregor Steiner, Mark Steel
Instrumental variables are a popular tool to infer causal effects under unobserved confounding, but choosing suitable instruments is challenging in practice. We propose gIVBMA, a Bayesian model averaging procedure that addresses this challenge by averaging across different sets of instrumental variables and covariates in a structural equation model. This allows for data-driven selection of valid and relevant instruments and provides additional robustness against invalid instruments. Our approach extends previous work through a scale-invariant prior structure and accommodates non-Gaussian outcomes and treatments, offering greater flexibility than existing methods. The computational strategy uses conditional Bayes factors to update models separately for the outcome and treatments. We prove that this model selection procedure is consistent. In simulation experiments, gIVBMA outperforms current state-of-the-art methods. We demonstrate its usefulness in two empirical applications: the effects of malaria and institutions on income per capita and the returns to schooling. A software implementation of gIVBMA is available in Julia.
FABRIC: A National-Scale Programmable Experimental Network Infrastructure
I. Baldin, A. Nikolich, J. Griffioen
et al.
FABRIC is a unique national research infrastructure to enable cutting-edge and exploratory research at-scale in networking, cybersecurity, distributed computing and storage systems, machine learning, and science applications. It is an everywhere-programmable nationwide instrument comprised of novel extensible network elements equipped with large amounts of compute and storage, interconnected by high speed, dedicated optical links. It will connect a number of specialized testbeds for cloud research (NSF Cloud testbeds CloudLab and Chameleon), for research beyond 5G technologies (Platforms for Advanced Wireless Research or PAWR), as well as production high-performance computing facilities and science instruments to create a rich fabric for a wide variety of experimental activities.
169 sitasi
en
Computer Science
Analisis Sentimen Ulasan Game Stumble Guys Pada Playstore Menggunakan Algoritma Naïve Bayes
Awang Herjunie Nurdy, Abdul Rahim, Arbansyah
Perkembangan teknologi yang pesat mempermudah akses ke berbagai hiburan digital, termasuk game online seperti Stumble Guys, yang telah diunduh lebih dari 163 juta kali dan mendapatkan ulasan beragam di Google Play Store. Penelitian ini bertujuan untuk menganalisis sentimen ulasan pengguna Stumble Guys menggunakan algoritma Naïve Bayes. Metode penelitian melibatkan tahapan Knowledge Discovery in Databases (KDD), meliputi pemilihan data, preprocessing, transformasi dengan CountVectorizer dan TF-IDF, serta pengklasifikasian dengan Naïve Bayes. Dengan menggunakan 1.500 ulasan dari Google Play Store, model Naïve Bayes mencapai akurasi 86%, dengan precision, recall, dan f1 score masing-masing sebesar 86%. Hasil penelitian menunjukkan bahwa Naïve Bayes efektif dalam mengklasifikasikan sentimen ulasan game Stumble Guys.
Information technology, Computer software
Claude 2.0 large language model: Tackling a real-world classification problem with a new iterative prompt engineering approach
Loredana Caruccio, Stefano Cirillo, Giuseppe Polese
et al.
In the last year, Large Language Models (LLMs) have transformed the way of tackling problems, opening up new perspectives in various works and research fields, due to their ability to generate and understand human languages. In this regard, the recent release of Claude 2.0 has contributed to the processing of more complex prompts. In this scenario, the goal of this paper is to evaluate the effectiveness of Claude 2.0 in a specific classification task. In particular, we considered the Forest cover-type problem, concerning the prediction of a cover-type value according to the geospatial characterization of target worldwide areas. To this end, we propose a novel iterative prompt template engineering approach, which integrates files by exploiting prompts and evaluates the quality of responses provided by the LLM. Moreover, we conducted several comparative analyses to evaluate the effectiveness of Claude 2.0 with respect to online and batch learning models. The results demonstrated that, although some online and batch models performed better than Claude 2.0, the new iterative prompt engineering approach improved the quality of responses, leading to better performance with increases ranging from 14% to 32% in terms of accuracy, precision, recall, and F1-score.
Cybernetics, Electronic computers. Computer science
Instrument-tissue Interaction Detection Framework for Surgical Video Understanding
Wenjun Lin, Yan Hu, Huazhu Fu
et al.
Instrument-tissue interaction detection task, which helps understand surgical activities, is vital for constructing computer-assisted surgery systems but with many challenges. Firstly, most models represent instrument-tissue interaction in a coarse-grained way which only focuses on classification and lacks the ability to automatically detect instruments and tissues. Secondly, existing works do not fully consider relations between intra- and inter-frame of instruments and tissues. In the paper, we propose to represent instrument-tissue interaction as <instrument class, instrument bounding box, tissue class, tissue bounding box, action class> quintuple and present an Instrument-Tissue Interaction Detection Network (ITIDNet) to detect the quintuple for surgery videos understanding. Specifically, we propose a Snippet Consecutive Feature (SCF) Layer to enhance features by modeling relationships of proposals in the current frame using global context information in the video snippet. We also propose a Spatial Corresponding Attention (SCA) Layer to incorporate features of proposals between adjacent frames through spatial encoding. To reason relationships between instruments and tissues, a Temporal Graph (TG) Layer is proposed with intra-frame connections to exploit relationships between instruments and tissues in the same frame and inter-frame connections to model the temporal information for the same instance. For evaluation, we build a cataract surgery video (PhacoQ) dataset and a cholecystectomy surgery video (CholecQ) dataset. Experimental results demonstrate the promising performance of our model, which outperforms other state-of-the-art models on both datasets.
A comparison of next-generation turbulence profiling instruments at Paranal
Ryan Griffiths, Lisa Bardou, Timothy Butterley
et al.
A six-night optical turbulence monitoring campaign has been carried at Cerro Paranal observatory in February and March, 2023 to facilitate the development and characterisation of two novel atmospheric site monitoring instruments - the ring-image next generation scintillation sensor (RINGSS) and 24-hour Shack Hartmann image motion monitor (24hSHIMM) in the context of providing optical turbulence monitoring support for upcoming 20-40m telescopes. Alongside these two instruments, the well-characterised Stereo-SCIDAR and 2016-MASS-DIMM were operated throughout the campaign to provide data for comparison. All instruments obtain estimates of optical turbulence profiles through statistical analysis of intensity and wavefront angle-of-arrival fluctuations from observations of stars. Contemporaneous measurements of the integrated turbulence parameters are compared and the ratios, bias, unbiased root mean square error and correlation of results from each instrument assessed. Strong agreement was observed in measurements of seeing, free atmosphere seeing and coherence time. Less correlation is seen for isoplanatic angle, although the median values agree well. Median turbulence parameters are further compared against long-term monitoring data from Paranal instruments. Profiles from the three small-telescope instruments are compared with the 100-layer profile from the stereo-SCIDAR. It is found that the RINGSS and SHIMM offer improved accuracy in characterisation of the vertical optical turbulence profile over the MASS-DIMM. Finally, the first results of continuous optical turbulence monitoring at Paranal are presented which show a strong diurnal variation and predictable trend in the seeing. A value of 2.65" is found for the median daytime seeing.
An insight into optical metrology in manufacturing
Y. Shimizu, Liang-Chia Chen, Dae Wook Kim
et al.
Optical metrology is one of the key technologies in today’s manufacturing industry. In this article, we provide an insight into optical measurement technologies for precision positioning and quality assessment in today’s manufacturing industry. First, some optical measurement technologies for precision positioning are explained, mainly focusing on those with a multi-axis positioning system composed of linear slides, often employed in machine tools or measuring instruments. Some optical measurement technologies for the quality assessment of products are then reviewed, focusing on technologies for form measurement of products with a large metric structure, from a telescope mirror to a nanometric structure such as a semiconductor electrode. Furthermore, we also review the state-of-the-art optical technique that has attracted attention in recent years, optical coherence tomography for the non-destructive inspection of the internal structures of a fabricated component, as well as super-resolution techniques for improving the lateral resolution of optical imaging beyond the diffraction limit of light. This review article provides insights into current and future technologies for optical measurement in the manufacturing industry, which are expected to become even more important to meet the industry’s continuing requirements for high-precision and high-efficiency machining.
128 sitasi
en
Engineering, Physics
Mobile detection of autism through machine learning on home video: A development and prospective validation study
Qandeel Tariq, Jena Daniels, J. Schwartz
et al.
Background The standard approaches to diagnosing autism spectrum disorder (ASD) evaluate between 20 and 100 behaviors and take several hours to complete. This has in part contributed to long wait times for a diagnosis and subsequent delays in access to therapy. We hypothesize that the use of machine learning analysis on home video can speed the diagnosis without compromising accuracy. We have analyzed item-level records from 2 standard diagnostic instruments to construct machine learning classifiers optimized for sparsity, interpretability, and accuracy. In the present study, we prospectively test whether the features from these optimized models can be extracted by blinded nonexpert raters from 3-minute home videos of children with and without ASD to arrive at a rapid and accurate machine learning autism classification. Methods and findings We created a mobile web portal for video raters to assess 30 behavioral features (e.g., eye contact, social smile) that are used by 8 independent machine learning models for identifying ASD, each with >94% accuracy in cross-validation testing and subsequent independent validation from previous work. We then collected 116 short home videos of children with autism (mean age = 4 years 10 months, SD = 2 years 3 months) and 46 videos of typically developing children (mean age = 2 years 11 months, SD = 1 year 2 months). Three raters blind to the diagnosis independently measured each of the 30 features from the 8 models, with a median time to completion of 4 minutes. Although several models (consisting of alternating decision trees, support vector machine [SVM], logistic regression (LR), radial kernel, and linear SVM) performed well, a sparse 5-feature LR classifier (LR5) yielded the highest accuracy (area under the curve [AUC]: 92% [95% CI 88%–97%]) across all ages tested. We used a prospectively collected independent validation set of 66 videos (33 ASD and 33 non-ASD) and 3 independent rater measurements to validate the outcome, achieving lower but comparable accuracy (AUC: 89% [95% CI 81%–95%]). Finally, we applied LR to the 162-video-feature matrix to construct an 8-feature model, which achieved 0.93 AUC (95% CI 0.90–0.97) on the held-out test set and 0.86 on the validation set of 66 videos. Validation on children with an existing diagnosis limited the ability to generalize the performance to undiagnosed populations. Conclusions These results support the hypothesis that feature tagging of home videos for machine learning classification of autism can yield accurate outcomes in short time frames, using mobile devices. Further work will be needed to confirm that this approach can accelerate autism diagnosis at scale.
178 sitasi
en
Medicine, Computer Science
Business Model Contributions to Bank Profit Performance: A Machine Learning Approach
BolivarFernando Bolívar, Miguel A. Duran, Ana Lozano-Vivas
This paper analyzes the relation between bank profit performance and business models. Using a machine learning-based approach, we propose a methodological strategy in which balance sheet components' contributions to profitability are the identification instruments of business models. We apply this strategy to the European Union banking system from 1997 to 2021. Our main findings indicate that the standard retail-oriented business model is the profile that performs best in terms of profitability, whereas adopting a non-specialized business profile is a strategic decision that leads to poor profitability. Additionally, our findings suggest that the effect of high capital ratios on profitability depends on the business profile. The contributions of business models to profitability decreased during the Great Recession. Although the situation showed signs of improvement afterward, the European Union banking system's ability to yield returns is still problematic in the post-crisis period, even for the best-performing group.
Recent developments in comprehensive analytical instruments for the culture heritage objects-A review
Yuanjun Xu, Zhu An, Ning Huang
et al.
This paper introduces the necessity and significance of the investigation of cultural heritage objects. The multi-technique method is useful for the study of cultural heritage objects, but a comprehensive analytical instrument is a better choice since it can guarantee that different types of information are always obtained from the same analytical point on the surface of cultural heritage objects, which may be crucial for some situations. Thus, the X-ray fluorescence (XRF)/X-ray diffraction (XRD) and X-ray fluorescence (XRF)/Raman spectroscopy (RS) comprehensive analytical instruments are more and more widely used to study cultural heritage objects. The two types of comprehensive analytical instruments are discussed in detail and the XRF/XRD instruments are further classified into different types on the basis of structure, type and number of detectors. A new comprehensive analytical instrument prototype that can perform XRF, XRD and RS measurements simultaneously has been successfully developed by our team and the preliminary application has shown the analysis performance and application potential. This overview contributes to better understand the research progress and development tendency of comprehensive analytical instruments for the study of cultural heritage objects. The new comprehensive instruments will make researchers obtain more valuable information on cultural heritage objects and further promote the study on cultural heritage objects.
en
physics.ins-det, physics.app-ph
Ghosting the Machine: Judicial Resistance to a Recidivism Risk Assessment Instrument
Dasha Pruss
Recidivism risk assessment instruments are presented as an 'evidence-based' strategy for criminal justice reform - a way of increasing consistency in sentencing, replacing cash bail, and reducing mass incarceration. In practice, however, AI-centric reforms can simply add another layer to the sluggish, labyrinthine machinery of bureaucratic systems and are met with internal resistance. Through a community-informed interview-based study of 23 criminal judges and other criminal legal bureaucrats in Pennsylvania, I find that judges overwhelmingly ignore a recently-implemented sentence risk assessment instrument, which they disparage as "useless," "worthless," "boring," "a waste of time," "a non-thing," and simply "not helpful." I argue that this algorithm aversion cannot be accounted for by individuals' distrust of the tools or automation anxieties, per the explanations given by existing scholarship. Rather, the instrument's non-use is the result of an interplay between three organizational factors: county-level norms about pre-sentence investigation reports; alterations made to the instrument by the Pennsylvania Sentencing Commission in response to years of public and internal resistance; and problems with how information is disseminated to judges. These findings shed new light on the important role of organizational influences on professional resistance to algorithms, which helps explain why algorithm-centric reforms can fail to have their desired effect. This study also contributes to an empirically-informed argument against the use of risk assessment instruments: they are resource-intensive and have not demonstrated positive on-the-ground impacts.
The Terzina instrument onboard the NUSES space mission
R. Aloisio, L. Burmistrov, A. Di Giovanni
et al.
In this paper we will introduce the Terzina instrument, which is one of the two scientific payloads of the NUSES satellite mission. NUSES serves as a technological pathfinder, hosting a suite of innovative instruments designed for the in-orbit detection of cosmic rays, neutrinos, and gamma rays across various energy ranges. The Terzina instrument itself is a compact telescope equipped with Schmidt-Cassegrain optics. Its primary objective is to detect Cherenkov radiation emitted by Extensive Air Showers generated by the interaction of high-energy (> 100 PeV) cosmic rays with the Earth's atmosphere. Terzina represents a critical step forward in the development of future space-based instruments aimed at detecting upward-moving showers induced by tau-leptons and muons resulting from the interaction of high-energy astrophysical neutrinos with the Earth. In this paper, we will delve into the key technical aspects of the Terzina instrument, its capabilities, and its potential for detection.
Learning Features of Music from Scratch
John Thickstun, Z. Harchaoui, S. Kakade
This paper introduces a new large-scale music dataset, MusicNet, to serve as a source of supervision and evaluation of machine learning methods for music research. MusicNet consists of hundreds of freely-licensed classical music recordings by 10 composers, written for 11 instruments, together with instrument/note annotations resulting in over 1 million temporal labels on 34 hours of chamber music performances under various studio and microphone conditions. The paper defines a multi-label classification task to predict notes in musical recordings, along with an evaluation protocol, and benchmarks several machine learning architectures for this task: i) learning from spectrogram features; ii) end-to-end learning with a neural net; iii) end-to-end learning with a convolutional neural net. These experiments show that end-to-end models trained for note prediction learn frequency selective filters as a low-level representation of audio.
227 sitasi
en
Mathematics, Computer Science