Hasil untuk "Computer Science"

Menampilkan 20 dari ~22629294 hasil · dari DOAJ, arXiv, Semantic Scholar, CrossRef

JSON API
arXiv Open Access 2026
The Science Data Lake: A Unified Open Infrastructure Integrating 293 Million Papers Across Eight Scholarly Sources with Embedding-Based Ontology Alignment

Jonas Wilinski

Scholarly data are largely fragmented across siloed databases with divergent metadata and missing linkages among them. We present the Science Data Lake, a locally-deployable infrastructure built on DuckDB and simple Parquet files that unifies eight open sources - Semantic Scholar, OpenAlex, SciSciNet, Papers with Code, Retraction Watch, Reliance on Science, a preprint-to-published mapping, and Crossref - via DOI normalization while preserving source-level schemas. The resource comprises approximately 960GB of Parquet files spanning ~293 million uniquely identifiable papers across ~22 schemas and ~153 SQL views. An embedding-based ontology alignment using BGE-large sentence embeddings maps 4,516 OpenAlex topics to 13 scientific ontologies (~1.3 million terms), yielding 16,150 mappings covering 99.8% of topics ($\geq 0.65$ threshold) with $F1 = 0.77$ at the recommended $\geq 0.85$ operating point, outperforming TF-IDF, BM25, and Jaro-Winkler baselines on a 300-pair gold-standard evaluation. We validate through 10 automated checks, cross-source citation agreement analysis (pairwise Pearson $r = 0.76$ - $0.87$), and stratified manual annotation. Four vignettes demonstrate cross-source analyses infeasible with any single database. The resource is open source, deployable on a single drive or queryable remotely via HuggingFace, and includes structured documentation suitable for large language model (LLM) based research agents.

en cs.DL, cs.DB
arXiv Open Access 2025
The Science Fiction Science Method

Iyad Rahwan, Azim Shariff, Jean-François Bonnefon

Predicting the social and behavioral impact of future technologies, before they are achieved, would allow us to guide their development and regulation before these impacts get entrenched. Traditionally, this prediction has relied on qualitative, narrative methods. Here we describe a method which uses experimental methods to simulate future technologies, and collect quantitative measures of the attitudes and behaviors of participants assigned to controlled variations of the future. We call this method 'science fiction science'. We suggest that the reason why this method has not been fully embraced yet, despite its potential benefits, is that experimental scientists may be reluctant to engage in work facing such serious validity threats as science fiction science. To address these threats, we consider possible constraints on the kind of technology that science fiction science may study, as well as the unconventional, immersive methods that science fiction science may require. We seek to provide perspective on the reasons why this method has been marginalized for so long, what benefits it would bring if it could be built on strong yet unusual methods, and how we can normalize these methods to help the diverse community of science fiction scientists to engage in a virtuous cycle of validity improvement.

en cs.HC, cs.AI
S2 Open Access 2010
SciNet: Lessons Learned from Building a Power-efficient Top-20 System and Data Centre

C. Loken, D. Gruner, L. Groer et al.

SciNet, one of seven regional HPC consortia operating under the Compute Canada umbrella, runs Canada's first and third fastest computers (as of June 2010) in a state-of-the-art, highly energy-efficient datacentre with a Power Usage Effectiveness (PUE) design-point of 1.16. Power efficiency, computational "bang for the buck" and system capability for a handful of flagship science projects were important criteria in choosing the nature of the computers and the data centre itself. Here we outline some of the lessons learned in putting together the systems and the data centre that hosts Canada's fastest computer to date.

494 sitasi en Physics, Engineering
DOAJ Open Access 2024
How Did People with Impairments Perceive Public Information During the COVID-19 Pandemic and What Are Their Suggestions for Accessible Crisis Information?

Karl Gummesson, Karin Forsell, Stefan Johansson et al.

The aim of this study was to explore how people with impairments perceived the accessibility of information regarding the COVID-19 pandemic in Sweden and what improvements they suggest to ensure accessibility of information in future societal crises. The study had a descriptive design, involving interviews and focus group discussions with people with impairments and their representative organisations, alongside analysis of public crisis information websites. The results showed that while many people with impairments could use their usual information channels, other found that the adapted information they needed was missing and that the government agencies, regional healthcare organisations and local municipalities were unprepared to produce accessible information. In conclusion, society exhibited shortcomings in providing accessible information to people with impairments during the COVID-19 pandemic. The responsible authorities seemed unprepared to provide accessible information. Proactive planning and training are imperative to ensure the provision of accessible information in future crises.

Social sciences (General)
DOAJ Open Access 2024
Daily natural gas load prediction method based on APSO optimization and Attention-BiLSTM

Xinjing Qi, Huan Wang, Yubo Ji et al.

As the economy continues to develop and technology advances, there is an increasing societal need for an environmentally friendly ecosystem. Consequently, natural gas, known for its minimal greenhouse gas emissions, has been widely adopted as a clean energy alternative. The accurate prediction of short-term natural gas demand poses a significant challenge within this context, as precise forecasts have important implications for gas dispatch and pipeline safety. The incorporation of intelligent algorithms into prediction methodologies has resulted in notable progress in recent times. Nevertheless, certain limitations persist. However, there exist certain limitations, including the tendency to easily fall into local optimization and inadequate search capability. To address the challenge of accurately predicting daily natural gas loads, we propose a novel methodology that integrates the adaptive particle swarm optimization algorithm, attention mechanism, and bidirectional long short-term memory (BiLSTM) neural networks. The initial step involves utilizing the BiLSTM network to conduct bidirectional data learning. Following this, the attention mechanism is employed to calculate the weights of the hidden layer in the BiLSTM, with a specific focus on weight distribution. Lastly, the adaptive particle swarm optimization algorithm is utilized to comprehensively optimize and design the network structure, initial learning rate, and learning rounds of the BiLSTM network model, thereby enhancing the accuracy of the model. The findings revealed that the combined model achieved a mean absolute percentage error (MAPE) of 0.90% and a coefficient of determination (R2) of 0.99. These results surpassed those of the other comparative models, demonstrating superior prediction accuracy, as well as exhibiting favorable generalization and prediction stability.

Electronic computers. Computer science
DOAJ Open Access 2024
Deep learning classification for macrophage subtypes through cell migratory pattern analysis

Manasa Kesapragada, Yao-Hui Sun, Ksenia Zlobina et al.

Macrophages can exhibit pro-inflammatory or pro-reparatory functions, contingent upon their specific activation state. This dynamic behavior empowers macrophages to engage in immune reactions and contribute to tissue homeostasis. Understanding the intricate interplay between macrophage motility and activation status provides valuable insights into the complex mechanisms that govern their diverse functions. In a recent study, we developed a classification method based on morphology, which demonstrated that movement characteristics, including speed and displacement, can serve as distinguishing factors for macrophage subtypes. In this study, we develop a deep learning model to explore the potential of classifying macrophage subtypes based solely on raw trajectory patterns. The classification model relies on the time series of x-y coordinates, as well as the distance traveled and net displacement. We begin by investigating the migratory patterns of macrophages to gain a deeper understanding of their behavior. Although this analysis does not directly inform the deep learning model, it serves to highlight the intricate and distinct dynamics exhibited by different macrophage subtypes, which cannot be easily captured by a finite set of motility metrics. Our study uses cell trajectories to classify three macrophage subtypes: M0, M1, and M2. This advancement holds promising implications for the future, as it suggests the possibility of identifying macrophage subtypes without relying on shape analysis. Consequently, it could potentially eliminate the necessity for high-quality imaging techniques and provide more robust methods for analyzing inherently blurry images.

Biology (General)
arXiv Open Access 2024
A Review of Pulse-Coupled Neural Network Applications in Computer Vision and Image Processing

Nurul Rafi, Pablo Rivas

Research in neural models inspired by mammal's visual cortex has led to many spiking neural networks such as pulse-coupled neural networks (PCNNs). These models are oscillating, spatio-temporal models stimulated with images to produce several time-based responses. This paper reviews PCNN's state of the art, covering its mathematical formulation, variants, and other simplifications found in the literature. We present several applications in which PCNN architectures have successfully addressed some fundamental image processing and computer vision challenges, including image segmentation, edge detection, medical imaging, image fusion, image compression, object recognition, and remote sensing. Results achieved in these applications suggest that the PCNN architecture generates useful perceptual information relevant to a wide variety of computer vision tasks.

en cs.CV, cs.LG
arXiv Open Access 2024
Promising and worth-to-try future directions for advancing state-of-the-art surrogates methods of agent-based models in social and health computational sciences

Atiyah Elsheikh

The execution and runtime performance of model-based analysis tools for realistic large-scale ABMs (Agent-Based Models) can be excessively long. This due to the computational demand exponentially proportional to the model size (e.g. Population size) and the number of model parameters. Even the runtime of a single simulation of a realistic ABM may demand huge computational resources when attempting to employ realistic population size. The main aim of this ad-hoc brief report is to highlight some of surrogate models that were adequate and computationally less demanding for nonlinear dynamical models in various modeling application areas.To the author knowledge, these methods have been not, at least extensively, employed for ABMs within the field of (SHCS) Social Health Computational Sciences, yet. Thus, they might be, but not necessarily, useful in progressing state of the art for establishing surrogate models for ABMs in the field of SHCS.

en cs.CL, cs.AI
arXiv Open Access 2024
PeersimGym: An Environment for Solving the Task Offloading Problem with Reinforcement Learning

Frederico Metelo, Stevo Racković, Pedro Ákos Costa et al.

Task offloading, crucial for balancing computational loads across devices in networks such as the Internet of Things, poses significant optimization challenges, including minimizing latency and energy usage under strict communication and storage constraints. While traditional optimization falls short in scalability; and heuristic approaches lack in achieving optimal outcomes, Reinforcement Learning (RL) offers a promising avenue by enabling the learning of optimal offloading strategies through iterative interactions. However, the efficacy of RL hinges on access to rich datasets and custom-tailored, realistic training environments. To address this, we introduce PeersimGym, an open-source, customizable simulation environment tailored for developing and optimizing task offloading strategies within computational networks. PeersimGym supports a wide range of network topologies and computational constraints and integrates a \textit{PettingZoo}-based interface for RL agent deployment in both solo and multi-agent setups. Furthermore, we demonstrate the utility of the environment through experiments with Deep Reinforcement Learning agents, showcasing the potential of RL-based approaches to significantly enhance offloading strategies in distributed computing settings. PeersimGym thus bridges the gap between theoretical RL models and their practical applications, paving the way for advancements in efficient task offloading methodologies.

en cs.LG, cs.AI

Halaman 39 dari 1131465