Wojciech Kuryłek
Hasil untuk "machine learning"
Menampilkan 20 dari ~2962868 hasil · dari CrossRef, DOAJ, arXiv
A. M. Mutawa, Khalid Sabti, Bibin Shalini Sundaram Thankaleela et al.
Diabetic macular edema (DME) is a primary contributor to visual impairment in diabetic patients, necessitating precise and prompt analysis for optimal treatment. Recent breakthroughs in deep learning (DL) and machine learning (ML) have yielded promising outcomes in ophthalmic image analysis. However, researchers often overlook the significance of optimization algorithms in enhancing the efficacy of their models for DME-related tasks. This review aims to consolidate, seek, discover, assess, and integrate existing work on the application of DL and ML, with emphasis on the integration and impact of optimization algorithms in enhancing their efficacy, robustness, and performance for DME in the fields of computer science and engineering. The population, intervention, comparison, and outcome framework was employed in this study to facilitate a clear and comprehensive analysis. The procedural superiority of the included investigations was evaluated using the Joanna Briggs Institute Critical Appraisal Tools for assessing methodological quality. The Auto-Metric Graph Neural Network achieved the greatest accuracy of 99.57% for combined diabetic retinopathy-DME grading, illustrating the higher efficacy of hybrid architectures augmented by meta-heuristic optimizers, such as Jaya and ant colony optimization. Successful deployment, however, depends on overcoming hurdles, such as the low mean average precision in lesion identification (0.1540) in YOLO-based models in the test set performance, and improved clinical interpretability to foster clinician trust. A Sankey diagram visually analyzes the flow of quantities between different entities of the survey.Systematic review registrationB. (2025, November 2). A Review of Optimization Strategies for Deep and Machine Learning in DME. Retrieved from osf.io/qsh4j.
Rok Rajher, Mila Marinković, Polona Rus Prelog et al.
Abstract Schizophrenia is a chronic and severe mental disorder that still relies on time-intensive, clinician-administered assessments. Although several automated approaches have been proposed to support diagnosis, these systems often lack the level of explainability necessary for informed clinical decision-making. In this study, we present a fully automated and explainable pipeline for detecting schizophrenia from audio recordings of verbal fluency tests, collected from 126 Slovene-speaking participants (68 healthy controls, 58 individuals diagnosed with schizophrenia), leveraging recent advancements in automatic speech recognition (ASR) and large language model (LLM) systems. We evaluated three ASR models–Truebar, Whisper, and Soniox–for transcription quality, and selected the best-performing system for further processing. We semantically enriched the transcriptions using the generative capabilities of LLMs and extracted both verbal and non-verbal features grounded in established diagnostic criteria. We assessed the relevance of these features using a Bayesian statistical framework and trained multiple classical machine learning models for automatic classification. Our best-performing model, an Explainable Boosting Machine, achieved a classification accuracy of 0.82 and an AUC of 0.90. We further generated visual explanations for the model’s predictions, establishing the first fully automated and explainable schizophrenia detection framework developed for the Slovene language. Our approach prioritizes explainability through model-transparent outputs, while still achieving performance comparable to existing automated systems for speech-based schizophrenia detection.
Christoph Balada, Max Bondorf, Sheraz Ahmed et al.
Electricity grids have become an essential part of daily life, even if they are often not noticed in everyday life. We usually only become particularly aware of this dependence by the time the electricity grid is no longer available. However, significant changes, such as the transition to renewable energy (photovoltaic, wind turbines, etc.) and an increasing number of energy consumers with complex load profiles (electric vehicles, home battery systems, etc.), pose new challenges for the electricity grid. At the same time, these challenges are usually too complex to be solved with traditional approaches. In this gap, where traditional approaches are reaching their limits, Machine Learning has become a popular tool to bridge this shortcoming through data-driven approaches. To enable novel ML implementations is we propose FiN-2 dataset, the first large-scale real-world broadband powerline communications (PLC) dataset. FiN-2 was collected during real practical use in a part of the German low-voltage grid that supplies energy to over 4.4 million people and shows well over two billion data points collected by more than 5100 sensors. In addition, we present different use cases in asset management, grid state visualization, forecasting, predictive maintenance, and novelty detection to highlight the benefits of these types of data. For these applications, we particularly highlight the use of novel machine learning architectures to extract rich information from real-world data that cannot be captured using traditional approaches. By publishing the first large-scale real-world dataset, we also aim to shed light on the previously largely unrecognized potential of PLC data and emphasize machine-learning-based research in low-voltage distribution networks by presenting a variety of different use cases.
Maher Abuhussain, Ali Hussain Alhamami, Khaled Almazam et al.
This study introduces a comprehensive framework combining building information modeling (BIM), project management body of knowledge (PMBOK), and machine learning (ML) to optimize energy efficiency and reduce environmental impacts in Riyadh’s construction sector. The suggested methodology utilizes BIM for dynamic energy simulations and design visualization, PMBOK for integrating sustainability into project-management processes, and ML for predictive modeling and real-time energy optimization. Implementing an integrated model that incorporates building-management strategies and machine learning for both commercial and residential structures can offer stakeholders a thorough solution for forecasting energy performance and environmental impact. This is particularly essential in arid climates owing to specific conditions and environmental limitations. Using a simulation-based methodology, the framework was evaluated based on two representative case studies: (i) a commercial complex and (ii) a residential building. The neural network (NN), reinforcement learning (RL), and decision tree (DT) were implemented to assess performance in energy prediction and optimization. Results demonstrated notable seasonal energy savings, particularly in spring (15% reduction for commercial buildings) and fall (13% reduction for residential buildings), driven by optimized heating, ventilation, and air conditioning (HVAC) systems, insulation strategies, and window configurations. ML models successfully predicted energy consumption and greenhouse gas (GHG) emissions, enabling targeted mitigation strategies. GHG emissions were reduced by up to 25% in commercial and 20% in residential settings. Among the models, NN achieved the highest predictive accuracy (R<sup>2</sup> = 0.95), while RL proved effective in adaptive operational control. This study highlights the synergistic potential of BIM, PMBOK, and ML in advancing green project management and sustainable construction.
Khaled Hamdaoui, Ali Benzaamia, Billal Sari Ahmed et al.
Abstract This study introduces a novel, interpretable machine learning framework for predicting the compression index (Cc) of clay soils by integrating three advanced gradient boosting algorithms—XGBoost, CatBoost, and LightGBM—with SHapley Additive exPlanations (SHAP). A comprehensive dataset of 1,243 clay samples, compiled from peer-reviewed literature, includes four geotechnical input variables: plastic limit (PL), plasticity index (PI), initial void ratio (e₀) and water content (w). Data were standardized and partitioned into training (70%) and testing (30%) subsets. Model development employed fivefold cross-validation and Optuna-based hyperparameter optimization. Among the models, XGBoost demonstrated the highest generalization capability, achieving an R2 of 0.913, RMSE of 0.197, and MAE of 0.100 on the test set. SHAP analysis revealed that initial void ratio (e₀) and water content (w) were the most influential features, with mean SHAP values of 0.20 and 0.10, respectively, aligning with established geotechnical principles. The proposed framework enhances transparency in machine learning predictions by making the model’s decision process understandable, thereby addressing the limitations of traditional “black-box” AI. It offers a reliable and efficient alternative to conventional oedometer testing, particularly beneficial for preliminary geotechnical design where timely and interpretable predictions are essential. Graphical Abstract
Hubert Baniecki, Przemyslaw Biecek
A common belief is that intrinsically interpretable deep learning models ensure a correct, intuitive understanding of their behavior and offer greater robustness against accidental errors or intentional manipulation. However, these beliefs have not been comprehensively verified, and growing evidence casts doubt on them. In this paper, we highlight the risks related to overreliance and susceptibility to adversarial manipulation of these so-called "intrinsically (aka inherently) interpretable" models by design. We introduce two strategies for adversarial analysis with prototype manipulation and backdoor attacks against prototype-based networks, and discuss how concept bottleneck models defend against these attacks. Fooling the model's reasoning by exploiting its use of latent prototypes manifests the inherent uninterpretability of deep neural networks, leading to a false sense of security reinforced by a visual confirmation bias. The reported limitations of part-prototype networks put their trustworthiness and applicability into question, motivating further work on the robustness and alignment of (deep) interpretable models.
Chungpa Lee, Sehee Lim, Kibok Lee et al.
Contrastive learning operates on a simple yet effective principle: Embeddings of positive pairs are pulled together, while those of negative pairs are pushed apart. In this paper, we propose a unified framework for understanding contrastive learning through the lens of cosine similarity, and present two key theoretical insights derived from this framework. First, in full-batch settings, we show that perfect alignment of positive pairs is unattainable when negative-pair similarities fall below a threshold, and this misalignment can be mitigated by incorporating within-view negative pairs into the objective. Second, in mini-batch settings, smaller batch sizes induce stronger separation among negative pairs in the embedding space, i.e., higher variance in their similarities, which in turn degrades the quality of learned representations compared to full-batch settings. To address this, we propose an auxiliary loss that reduces the variance of negative-pair similarities in mini-batch settings. Empirical results show that incorporating the proposed loss improves performance in small-batch settings.
Valentino F. Foit, David W. Hogg, Soledad Villar
Many machine learning tasks in the natural sciences are precisely equivariant to particular symmetries. Nonetheless, equivariant methods are often not employed, perhaps because training is perceived to be challenging, or the symmetry is expected to be learned, or equivariant implementations are seen as hard to build. Group averaging is an available technique for these situations. It happens at test time; it can make any trained model precisely equivariant at a (often small) cost proportional to the size of the group; it places no requirements on model structure or training. It is known that, under mild conditions, the group-averaged model will have a provably better prediction accuracy than the original model. Here we show that an inexpensive group averaging can improve accuracy in practice. We take well-established benchmark machine learning models of differential equations in which certain symmetries ought to be obeyed. At evaluation time, we average the models over a small group of symmetries. Our experiments show that this procedure always decreases the average evaluation loss, with improvements of up to 37\% in terms of the VRMSE. The averaging produces visually better predictions for continuous dynamics. This short paper shows that, under certain common circumstances, there are no disadvantages to imposing exact symmetries; the ML4PS community should consider group averaging as a cheap and simple way to improve model accuracy.
Tanay Agrawal
Badar Almarri, Gaurav Gupta, Ravinder Kumar et al.
Manal Binkhonain, Liping Zhao
Siddharth Solaiyappan, Yuxin Wen
Alok Singh Chauhan, H Mary Henrietta
The domain of machine learning has experienced an unparalleled increase in attention and implementation, becoming an essential component of diverse businesses. This review paper provides a thorough analysis of the comprehensive handbook named "Machine Learning Basics: A Comprehensive Guide." Written by [Dr. Jane Doe], this guide has become a vital reference for those at all levels of expertise seeking to comprehend and traverse the intricate realm of machine learning.
Paolo Graniero, Paolo Graniero, Mark Khenkin et al.
Perovskite solar cells are the most dynamic emerging photovoltaic technology and attracts the attention of thousands of researchers worldwide. Recently, many of them are targeting device stability issues–the key challenge for this technology–which has resulted in the accumulation of a significant amount of data. The best example is the “Perovskite Database Project,” which also includes stability-related metrics. From this database, we use data on 1,800 perovskite solar cells where device stability is reported and use Random Forest to identify and study the most important factors for cell stability. By applying the concept of learning curves, we find that the potential for improving the models’ performance by adding more data of the same quality is limited. However, a significant improvement can be made by increasing data quality by reporting more complete information on the performed experiments. Furthermore, we study an in-house database with data on more than 1,000 solar cells, where the entire aging curve for each cell is available as opposed to stability metrics based on a single number. We show that the interpretation of aging experiments can strongly depend on the chosen stability metric, unnaturally favoring some cells over others. Therefore, choosing universal stability metrics is a critical question for future databases targeting this promising technology.
Michael Banf, Michael Banf, Kangmei Zhao et al.
Iris R. Stone, Yotam Sagiv, Il Memming Park et al.
Latent linear dynamical systems with Bernoulli observations provide a powerful modeling framework for identifying the temporal dynamics underlying binary time series data, which arise in a variety of contexts such as binary decision-making and discrete stochastic processes (e.g., binned neural spike trains). Here we develop a spectral learning method for fast, efficient fitting of probit-Bernoulli latent linear dynamical system (LDS) models. Our approach extends traditional subspace identification methods to the Bernoulli setting via a transformation of the first and second sample moments. This results in a robust, fixed-cost estimator that avoids the hazards of local optima and the long computation time of iterative fitting procedures like the expectation-maximization (EM) algorithm. In regimes where data is limited or assumptions about the statistical structure of the data are not met, we demonstrate that the spectral estimate provides a good initialization for Laplace-EM fitting. Finally, we show that the estimator provides substantial benefits to real world settings by analyzing data from mice performing a sensory decision-making task.
Xiaoling Wei, Yongbao Feng, Xiaoxia Han et al.
At present, with the continuous development and great improvement of mechanical manufacturing, processing, and assembly technology, mechanical flow-induced vibration (FIV) with a relatively concentrated frequency domain can be controlled by active and passive noise reduction methods. However, whether it is active noise reduction or passive noise reduction, they all focus on how to suppress the transmission of sound waves and cannot solve the problems of flow leakage, obvious temperature rise, and noise excitation from the root cause. Therefore, it is necessary to determine the location of the primary and secondary excitation sound sources of FIV, the identification of true and false sounds, and the characteristic relationship between flow and noise. This provides a theoretical basis and engineering application direction for the mechanism of noise reduction of FIV. The numerical calculation part of the acoustics in this paper is solved by the hybrid method, and the flow field is discretely calculated by the large eddy simulation (LES) module in the Fluent software. When the calculated flow field is stable, the velocity field of one impeller rotation period is selected to be output as the iterative value of the sound field and imported into ACTRAN for Fourier transform. Then, the sound field calculation is carried out, and the result of the spatial and temporal variation of the sound field is finally obtained. Through experiments, it was found that when the load of the gear pump is 8 MPa, the volumetric efficiency of the optimized circular-arc helical gear pump of the sliding bearing was improved by about 4%. When the rotation speed is 2100°r/min, the arc helical gear pump reduced the surface temperature rise by 2.5°C. This verified that the optimized performance of the sliding bearing in the arc helical gear pump is significantly improved. Through the theoretical model of the temperature rise of the sliding bearing, the phenomenon that the surface temperature of the prototype gear pump was not significantly increased with the loading in the low pressure region is explained.
Nazish Shahid
Abstract A synthesized investigation, employing graphical and analytical approach, has been conducted to examine inadequacy of electronic education and limitations posed by transformative mode of learning from students’ perspective. Moreover, the breadth of subject understanding through digital mode and students’ preference for physical or electronic mode of learning in the future were examined. A descriptive analysis was executed through R programming for the obtained numeric-characteristic statistics. For computational analysis of the data to determine proportion of deteriorating virtual-assessment performance attributed to conditioned subject-command, a machine learning approach of interaction-regression is adopted. It is implied through the obtained results that a majority of students felt discontented at not being able to achieve optimized learning outcomes post-virtual-attendance of study programs. It is also concluded that blended influence of online learning and partial subject-command resulted in insufficient assessment performance. Additionally, the current study highlights the importance of need-based adaptations to facilitate automated mode of learning and virtual platforms’ uniform access to students.
Halaman 7 dari 148144