Hasil untuk "Language and Literature"

Menampilkan 20 dari ~880404 hasil · dari DOAJ, Semantic Scholar, arXiv

JSON API
S2 Open Access 2020
A Survey on Text Classification: From Traditional to Deep Learning

Qian Li, Hao Peng, Jianxin Li et al.

Text classification is the most fundamental and essential task in natural language processing. The last decade has seen a surge of research in this area due to the unprecedented success of deep learning. Numerous methods, datasets, and evaluation metrics have been proposed in the literature, raising the need for a comprehensive and updated survey. This paper fills the gap by reviewing the state-of-the-art approaches from 1961 to 2021, focusing on models from traditional models to deep learning. We create a taxonomy for text classification according to the text involved and the models used for feature extraction and classification. We then discuss each of these categories in detail, dealing with both the technical developments and benchmark datasets that support tests of predictions. A comprehensive comparison between different techniques, as well as identifying the pros and cons of various evaluation metrics are also provided in this survey. Finally, we conclude by summarizing key implications, future research directions, and the challenges facing the research area.

506 sitasi en Computer Science
S2 Open Access 2019
Setting an Agenda for Positive Psychology in SLA: Theory, Practice, and Research

P. MacIntyre, T. Gregersen, Sarah Mercer

In this article we introduce Positive Psychology (PP), a relatively new subfield of psychology, and outline its development since the year 2000. We describe ways in which PP represents an exciting addition to the Second Language Acquisition (SLA) literature and the ways it is already influencing trends in education generally, thus creating promising expectations of its impact on language teaching and learning. After reviewing the progress made thus far under the rubric of PP in SLA, we offer suggestions for an agenda to move forward with theory, research, and practice.

463 sitasi en Psychology
S2 Open Access 2019
Distilling Task-Specific Knowledge from BERT into Simple Neural Networks

Raphael Tang, Yao Lu, Linqing Liu et al.

In the natural language processing literature, neural networks are becoming increasingly deeper and complex. The recent poster child of this trend is the deep language representation model, which includes BERT, ELMo, and GPT. These developments have led to the conviction that previous-generation, shallower neural networks for language understanding are obsolete. In this paper, however, we demonstrate that rudimentary, lightweight neural networks can still be made competitive without architecture changes, external training data, or additional input features. We propose to distill knowledge from BERT, a state-of-the-art language representation model, into a single-layer BiLSTM, as well as its siamese counterpart for sentence-pair tasks. Across multiple datasets in paraphrasing, natural language inference, and sentiment classification, we achieve comparable results with ELMo, while using roughly 100 times fewer parameters and 15 times less inference time.

454 sitasi en Computer Science
S2 Open Access 1997
Building applied natural language generation systems

Ehud Reiter, R. Dale

In this article, we give an overview of Natural Language Generation (NLG) from an applied system-building perspective. The article includes a discussion of when NLG techniques should be used; suggestions for carrying out requirements analyses; and a description of the basic NLG tasks of content determination, discourse planning, sentence aggregation, lexicalization, referring expression generation, and linguistic realisation. Throughout, the emphasis is on established techniques that can be used to build simple but practical working systems now. We also provide pointers to techniques in the literature that are appropriate for more complicated scenarios.

751 sitasi en Computer Science
S2 Open Access 2023
SciMON: Scientific Inspiration Machines Optimized for Novelty

Qingyun Wang, Doug Downey, Heng Ji et al.

We explore and enhance the ability of neural language models to generate novel scientific directions grounded in literature. Work on literature-based hypothesis generation has traditionally focused on binary link prediction--severely limiting the expressivity of hypotheses. This line of work also does not focus on optimizing novelty. We take a dramatic departure with a novel setting in which models use as input background contexts (e.g., problems, experimental settings, goals), and output natural language ideas grounded in literature. We present SciMON, a modeling framework that uses retrieval of"inspirations"from past scientific papers, and explicitly optimizes for novelty by iteratively comparing to prior papers and updating idea suggestions until sufficient novelty is achieved. Comprehensive evaluations reveal that GPT-4 tends to generate ideas with overall low technical depth and novelty, while our methods partially mitigate this issue. Our work represents a first step toward evaluating and developing language models that generate new ideas derived from the scientific literature

163 sitasi en Computer Science
DOAJ Open Access 2026
The effectiveness of the ECRIF framework in developing EFL students' engagement and learning outcomes: A case study of Moroccan middle school students

Ayad Chraa, Hanane Aqadoh

Recent shifts in education, particularly in English language teaching (ELT), have emphasized learner-centered approaches over teacher-centered ones. The ECRIF framework (Encounter, Clarify, Remember, Internalize, Fluently Use) is widely recognized for providing a practical roadmap for facilitating active, meaningful, and student-centered learning experiences. This study investigates its effectiveness in developing learning outcomes and engagement among Moroccan EFL middle school students. Data were collected from a sample of 67 ninth-grade students, using a pre-test, post-test, and feedback questionnaire. The results revealed that the ECRIF framework significantly enhanced students’ performance in the grammar test compared to traditional instruction. While both instructional methods resulted in learning gains, the improvement observed in the control group was modest compared to the substantial progress made by the experimental group. The structured stages and interactive activities incorporated throughout the instructional session contributed to the positive impact of the ECRIF framework on the experimental group’s achievement. The students expressed high levels of satisfaction and positive perceptions toward the ECRIF approach, as reflected in their responses to the feedback questionnaire. Overall, the present study demonstrated that the ECRIF activities used during the lesson not only improved students’ learning outcomes but also maximized their engagement and participation. These findings suggest that the framework can effectively foster both student learning and engagement in the English classroom. To this end, the study calls for the integration of ECRIF framework to enhance learner-centered practices and improve educational outcomes.

Philology. Linguistics
arXiv Open Access 2026
AfriNLLB: Efficient Translation Models for African Languages

Yasmin Moslem, Aman Kassahun Wassie, Amanuel Gizachew Abebe

In this work, we present AfriNLLB, a series of lightweight models for efficient translation from and into African languages. AfriNLLB supports 15 language pairs (30 translation directions), including Swahili, Hausa, Yoruba, Amharic, Somali, Zulu, Lingala, Afrikaans, Wolof, and Egyptian Arabic, as well as other African Union official languages such as Arabic (MSA), French, Portuguese, and Spanish. Our training data covers bidirectional translation between English and 13 languages, and between French and two languages (Lingala and Wolof). AfriNLLB models are based on NLLB-200 600M, which we compress using iterative layer pruning and quantization. We fine-tune the pruned models on parallel corpora we curated for African languages, employing knowledge distillation from a larger teacher model. Our work aims at enabling efficient deployment of translation models for African languages in resource-constrained settings. Our evaluation results demonstrate that AfriNLLB models achieve performance comparable to the baseline while being significantly faster. We release two versions of the AfriNLLB models, a Transformers version that allows further fine-tuning and a CTranslate2 version for efficient inference. Moreover, we release all the training data that we used for fine-tuning the baseline and pruned models to facilitate further research.

en cs.CL
arXiv Open Access 2026
Infusion of Blockchain to Establish Trustworthiness in AI Supported Software Evolution: A Systematic Literature Review

Mohammad Naserameri, Juergen Rilling

Context: Blockchain and AI are increasingly explored to enhance trustworthiness in software engineering (SE), particularly in supporting software evolution tasks. Method: We conducted a systematic literature review (SLR) using a predefined protocol with clear eligibility criteria to ensure transparency, reproducibility, and minimized bias, synthesizing research on blockchain-enabled trust in AI-driven SE tools and processes. Results: Most studies focus on integrating AI in SE, with only 31% explicitly addressing trustworthiness. Our review highlights six recent studies exploring blockchain-based approaches to reinforce reliability, transparency, and accountability in AI-assisted SE tasks. Conclusion: Blockchain enhances trust by ensuring data immutability, model transparency, and lifecycle accountability, including federated learning with blockchain consensus and private data verification. However, inconsistent definitions of trust and limited real-world testing remain major challenges. Future work must develop measurable, reproducible trust frameworks to enable reliable, secure, and compliant AI-driven SE ecosystems, including applications involving large language models.

en cs.SE
arXiv Open Access 2025
Conversational Lexicography: Querying Lexicographic Data on Knowledge Graphs with SPARQL through Natural Language

Kilian Sennrich, Sina Ahmadi

Knowledge graphs offer an excellent solution for representing the lexical-semantic structures of lexicographic data. However, working with the SPARQL query language represents a considerable hurdle for many non-expert users who could benefit from the advantages of this technology. This paper addresses the challenge of creating natural language interfaces for lexicographic data retrieval on knowledge graphs such as Wikidata. We develop a multidimensional taxonomy capturing the complexity of Wikidata's lexicographic data ontology module through four dimensions and create a template-based dataset with over 1.2 million mappings from natural language utterances to SPARQL queries. Our experiments with GPT-2 (124M), Phi-1.5 (1.3B), and GPT-3.5-Turbo reveal significant differences in model capabilities. While all models perform well on familiar patterns, only GPT-3.5-Turbo demonstrates meaningful generalization capabilities, suggesting that model size and diverse pre-training are crucial for adaptability in this domain. However, significant challenges remain in achieving robust generalization, handling diverse linguistic data, and developing scalable solutions that can accommodate the full complexity of lexicographic knowledge representation.

en cs.CL
arXiv Open Access 2025
From Data to Knowledge: Evaluating How Efficiently Language Models Learn Facts

Daniel Christoph, Max Ploner, Patrick Haller et al.

Sample efficiency is a crucial property of language models with practical implications for training efficiency. In real-world text, information follows a long-tailed distribution. Yet, we expect models to learn and recall frequent and infrequent facts. Sample-efficient models are better equipped to handle this challenge of learning and retaining rare information without requiring excessive exposure. This study analyzes multiple models of varying architectures and sizes, all trained on the same pre-training data. By annotating relational facts with their frequencies in the training corpus, we examine how model performance varies with fact frequency. Our findings show that most models perform similarly on high-frequency facts but differ notably on low-frequency facts. This analysis provides new insights into the relationship between model architecture, size, and factual learning efficiency.

en cs.CL, cs.LG

Halaman 27 dari 44021