A. Morrison, Julie Polisena, D. Husereau et al.
Hasil untuk "English literature"
Menampilkan 20 dari ~7044672 hasil · dari arXiv, DOAJ, Semantic Scholar
D. Jackson, Angela Firtko, Michel A Edenborough
Shelly F. Greenfield, Shelly F. Greenfield, A. Brooks et al.
J. Lacro, L. Dunn, Christian R. Dolder et al.
R. Huddleston, G. Pullum
H. Payne, C. Lister, J. West et al.
Background Several thousand mobile phone apps are available to download to mobile phones for health and fitness. Mobile phones may provide a unique means of administering health interventions to populations. Objective The purpose of this systematic review was to systematically search and describe the literature on mobile apps used in health behavior interventions, describe the behavioral features and focus of health apps, and to evaluate the potential of apps to disseminate health behavior interventions. Methods We conducted a review of the literature in September 2014 using key search terms in several relevant scientific journal databases. Only English articles pertaining to health interventions using mobile phone apps were included in the final sample. Results The 24 studies identified for this review were primarily feasibility and pilot studies of mobile apps with small sample sizes. All studies were informed by behavioral theories or strategies, with self-monitoring as the most common construct. Acceptability of mobile phone apps was high among mobile phone users. Conclusions The lack of large sample studies using mobile phone apps may signal a need for additional studies on the potential use of mobile apps to assist individuals in changing their health behaviors. Of these studies, there is early evidence that apps are well received by users. Based on available research, mobile apps may be considered a feasible and acceptable means of administering health interventions, but a greater number of studies and more rigorous research and evaluations are needed to determine efficacy and establish evidence for best practices.
G. Raghu, Valeria C. Amatto, J. Behr et al.
Yiran Rex Ma
Based on comparable English-Chinese news corpora annotated by Large Language Model (LLM), this paper attempts to explore the differences in constituent order of English-Chinese news from the perspective of functional chunks with adverbial roles, and analyze their typical positional preferences and distribution patterns. It is found that: (1) English news prefers linear narrative of core information first, and functional chunks are mostly post-positioned, while Chinese news prefers overall presentation mode of background first, and functional chunks are often pre-positioned; (2) In SVO structure, both English and Chinese news show differences in the distribution of functional chunks, but the tendency of Chinese pre-positioning is more significant, while that of English post-positioning is relatively mild; (3) When function blocks are co-occurring, both English and Chinese news show high flexibility, and the order adjustment is driven by information and pragmatic purposes. The study reveals that word order has both systematic preference and dynamic adaptability, providing new empirical support for contrastive study of English-Chinese information structure.
Iona Carslaw, Sivan Milton, Nicolas Navarre et al.
For linguists, embedded clauses have been of special interest because of their intricate distribution of syntactic and semantic features. Yet, current research relies on schematically created language examples to investigate these constructions, missing out on statistical information and naturally-occurring examples that can be gained from large language corpora. Thus, we present a methodological approach for detecting and annotating naturally-occurring examples of English embedded clauses in large-scale text data using constituency parsing and a set of parsing heuristics. Our tool has been evaluated on our dataset Golden Embedded Clause Set (GECS), which includes hand-annotated examples of naturally-occurring English embedded clause sentences. Finally, we present a large-scale dataset of naturally-occurring English embedded clauses which we have extracted from the open-source corpus Dolma using our extraction tool.
Sophie Jaffer, Simeon Sayer
As large language models (LLMs) expand multilingual capabilities, questions remain about the equity of their performance across languages. While many communities stand to benefit from AI systems, the dominance of English in training data risks disadvantaging non-English speakers. To test the hypothesis that such data disparities may affect model performance, this study compares two monolingual BERT models: one trained and tested entirely on Swahili data, and another on comparable English news data. To simulate how multilingual LLMs process non-English queries through internal translation and abstraction, we translated the Swahili news data into English and evaluated it using the English-trained model. This approach tests the hypothesis by evaluating whether translating Swahili inputs for evaluation on an English model yields better or worse performance compared to training and testing a model entirely in Swahili, thus isolating the effect of language consistency versus cross-lingual abstraction. The results prove that, despite high-quality translation, the native Swahili-trained model performed better than the Swahili-to-English translated model, producing nearly four times fewer errors: 0.36% vs. 1.47% respectively. This gap suggests that translation alone does not bridge representational differences between languages and that models trained in one language may struggle to accurately interpret translated inputs due to imperfect internal knowledge representation, suggesting that native-language training remains important for reliable outcomes. In educational and informational contexts, even small performance gaps may compound inequality. Future research should focus on addressing broader dataset development for underrepresented languages and renewed attention to multilingual model evaluation, ensuring the reinforcing effect of global AI deployment on existing digital divides is reduced.
O.V. Toikina
Verbs of speaking are widely represented in modern English and Russian. The group of semantically related speaking verbs in English includes: to say, to speak, to tell, and to talk ; in Russian, this group can include the following words: говорить, сказать, рассказывать, and разговаривать . In both languages, the semantic group of verbs of speaking is much broader and is not limited to the verbs listed above. A significant number of synonyms for speaking appear and disappear under the influence of language style and usage in various historical periods. At the same time, the above-mentioned verbs, which appeared at the time of the emergence of the language, have not only preserved their original meanings but have also significantly expanded them. Therefore, the aim of this study is to identify the features of the development of the semantic meaning of verbs of speaking in English and Russian from both diachronic and synchronic perspectives. The relevance of this study lies in the need to examine the semantic meaning of verbs of speaking in both languages, as they are widely used across a wide variety of subject areas. Data from the British National Corpus show that the four synonymous verbs occur in both spoken and written English and are used in the following subject fields: literature, religion, mathematics, economics and commerce, linguistics, law, telecommunications, and others. According to the Russian National Corpus, these Russian speaking verbs occur in areas such as: fiction, journalism, everyday speech, educational and scientific literature, church and theological texts, electronic communications, official and business documents, advertising, and industrial and technical writing. In addition, these verbs are among the first 1,000 most frequently used words in modern English and Russian.
Rricha Jalota, Lyan Verwimp, Markus Nussbaum-Thom et al.
Neural Network Language Models (NNLMs) for Virtual Assistants (VAs) are generally language-, region-, and in some cases, device-dependent, which increases the effort to scale and maintain them. Combining NNLMs for one or more of the categories is one way to improve scalability. In this work, we combine regional variants of English to build a ``World English'' NNLM for on-device VAs. In particular, we investigate the application of adapter bottlenecks to model dialect-specific characteristics in our existing production NNLMs {and enhance the multi-dialect baselines}. We find that adapter modules are more effective in modeling dialects than specializing entire sub-networks. Based on this insight and leveraging the design of our production models, we introduce a new architecture for World English NNLM that meets the accuracy, latency, and memory constraints of our single-dialect models.
Ismael Villegas Molina, Audria Montalvo, Benjamin Ochoa et al.
Computer science has historically presented barriers for non-native English speaking (NNES) students, often due to language and terminology challenges. With the rise of large language models (LLMs), there is potential to leverage this technology to support NNES students more effectively. Recent implementations of LLMs as tutors in classrooms have shown promising results. In this study, we deployed an LLM tutor in an accelerated introductory computing course to evaluate its effectiveness specifically for NNES students. Key insights for LLM tutor use are as follows: NNES students signed up for the LLM tutor at a similar rate to native English speakers (NES); NNES students used the system at a lower rate than NES students -- to a small effect; NNES students asked significantly more questions in languages other than English compared to NES students, with many of the questions being multilingual by incorporating English programming keywords. Results for views of the LLM tutor are as follows: both NNES and NES students appreciated the LLM tutor for its accessibility, conversational style, and the guardrails put in place to guide users to answers rather than directly providing solutions; NNES students highlighted its approachability as they did not need to communicate in perfect English; NNES students rated help-seeking preferences of online resources higher than NES students; Many NNES students were unfamiliar with computing terminology in their native languages. These results suggest that LLM tutors can be a valuable resource for NNES students in computing, providing tailored support that enhances their learning experience and overcomes language barriers.
Maryam Al Ali, Hanan Aldarmaki
This paper introduces Mixat: a dataset of Emirati speech code-mixed with English. Mixat was developed to address the shortcomings of current speech recognition resources when applied to Emirati speech, and in particular, to bilignual Emirati speakers who often mix and switch between their local dialect and English. The data set consists of 15 hours of speech derived from two public podcasts featuring native Emirati speakers, one of which is in the form of conversations between the host and a guest. Therefore, the collection contains examples of Emirati-English code-switching in both formal and natural conversational contexts. In this paper, we describe the process of data collection and annotation, and describe some of the features and statistics of the resulting data set. In addition, we evaluate the performance of pre-trained Arabic and multi-lingual ASR systems on our dataset, demonstrating the shortcomings of existing models on this low-resource dialectal Arabic, and the additional challenge of recognizing code-switching in ASR. The dataset will be made publicly available for research use.
Chris Wendler, Veniamin Veselovsky, Giovanni Monea et al.
We ask whether multilingual language models trained on unbalanced, English-dominated corpora use English as an internal pivot language -- a question of key importance for understanding how language models function and the origins of linguistic bias. Focusing on the Llama-2 family of transformer models, our study uses carefully constructed non-English prompts with a unique correct single-token continuation. From layer to layer, transformers gradually map an input embedding of the final prompt token to an output embedding from which next-token probabilities are computed. Tracking intermediate embeddings through their high-dimensional space reveals three distinct phases, whereby intermediate embeddings (1) start far away from output token embeddings; (2) already allow for decoding a semantically correct next token in the middle layers, but give higher probability to its version in English than in the input language; (3) finally move into an input-language-specific region of the embedding space. We cast these results into a conceptual model where the three phases operate in "input space", "concept space", and "output space", respectively. Crucially, our evidence suggests that the abstract "concept space" lies closer to English than to other languages, which may have important consequences regarding the biases held by multilingual language models.
Amir Hossein Kargaran, Ali Modarressi, Nafiseh Nikeghbal et al.
English-centric large language models (LLMs) often show strong multilingual capabilities. However, their multilingual performance remains unclear and is under-evaluated for many other languages. Most benchmarks for multilinguality focus on classic NLP tasks or cover a minimal number of languages. We introduce MEXA, a method for assessing the multilingual capabilities of pre-trained English-centric LLMs using parallel sentences, which are available for more languages than existing downstream tasks. MEXA leverages that English-centric LLMs use English as a pivot language in their intermediate layers. MEXA computes the alignment between English and non-English languages using parallel sentences to evaluate the transfer of language understanding from English to other languages. This alignment can be used to estimate model performance in different languages. We conduct controlled experiments using various parallel datasets (FLORES-200 and Bible), models (Llama family, Gemma family, Mistral, and OLMo), and established downstream tasks (Belebele, m-MMLU, and m-ARC). We explore different methods to compute embeddings in decoder-only models. Our results show that MEXA, in its default settings, achieves an average Pearson correlation of 0.90 between its predicted scores and actual task performance across languages. This suggests that MEXA is a reliable method for estimating the multilingual capabilities of English-centric LLMs, providing a clearer understanding of their multilingual potential and the inner workings of LLMs. Leaderboard: https://cis-lmu-mexa.hf.space, Code: https://github.com/cisnlp/MEXA.
Manon Reusens, Philipp Borchert, Jochen De Weerdt et al.
Large Language Models (LLMs) excel at providing information acquired during pretraining on large-scale corpora and following instructions through user prompts. This study investigates whether the quality of LLM responses varies depending on the demographic profile of users. Considering English as the global lingua franca, along with the diversity of its dialects among speakers of different native languages, we explore whether non-native English speakers receive lower-quality or even factually incorrect responses from LLMs more frequently. Our results show that performance discrepancies occur when LLMs are prompted by native versus non-native English speakers and persist when comparing native speakers from Western countries with others. Additionally, we find a strong anchoring effect when the model recognizes or is made aware of the user's nativeness, which further degrades the response quality when interacting with non-native speakers. Our analysis is based on a newly collected dataset with over 12,000 unique annotations from 124 annotators, including information on their native language and English proficiency.
Alvin Kuo Jing Teo, Emily Lai-Ho MacLean, Greg J. Fox
Background: This scoping review aimed to characterise definitions used to describe subclinical tuberculosis (TB), estimate the prevalence in different populations and describe the clinical characteristics and treatment outcomes in the scientific literature. Methods: A systematic literature search was conducted using PubMed. We included studies published in English between January 1990 and August 2022 that defined “subclinical” or “asymptomatic” pulmonary TB disease, regardless of age, HIV status and comorbidities. We estimated the weighted pooled proportions of subclinical TB using a random-effects model by World Health Organization reported TB incidence, populations and settings. We also pooled the proportion of subclinical TB according to definitions described in published prevalence surveys. Results: We identified 29 prevalence surveys and 71 other studies. Prevalence survey data (2002–2022) using “absence of cough of any duration” criteria reported higher subclinical TB prevalence than those using the stricter “completely asymptomatic” threshold. Prevalence estimates overlap in studies using other symptoms and cough duration. Subclinical TB in studies was commonly defined as asymptomatic TB disease. Higher prevalence was reported in high TB burden areas, community settings and immunocompetent populations. People with subclinical TB showed less extensive radiographic abnormalities, higher treatment success rates and lower mortality, although studies were few. Conclusion: A substantial proportion of TB is subclinical. However, prevalence estimates were highly heterogeneous between settings. Most published studies incompletely characterised the phenotype of people with subclinical TB. Standardised definitions and diagnostic criteria are needed to characterise this phenotype. Further research is required to enhance case finding, screening, diagnostics and treatment options for subclinical TB.
Xuan-Quy Dao
This paper presents a performance comparison of three large language models (LLMs), namely OpenAI ChatGPT, Microsoft Bing Chat (BingChat), and Google Bard, on the VNHSGE English dataset. The performance of BingChat, Bard, and ChatGPT (GPT-3.5) is 92.4\%, 86\%, and 79.2\%, respectively. The results show that BingChat is better than ChatGPT and Bard. Therefore, BingChat and Bard can replace ChatGPT while ChatGPT is not yet officially available in Vietnam. The results also indicate that BingChat, Bard and ChatGPT outperform Vietnamese students in English language proficiency. The findings of this study contribute to the understanding of the potential of LLMs in English language education. The remarkable performance of ChatGPT, BingChat, and Bard demonstrates their potential as effective tools for teaching and learning English at the high school level.
Hari Prabhat Gupta, Rahul Mishra
This paper presents an end-to-end methodology for collecting datasets to recognize handwritten English alphabets by utilizing Inertial Measurement Units (IMUs) and leveraging the diversity present in the Indian writing style. The IMUs are utilized to capture the dynamic movement patterns associated with handwriting, enabling more accurate recognition of alphabets. The Indian context introduces various challenges due to the heterogeneity in writing styles across different regions and languages. By leveraging this diversity, the collected dataset and the collection system aim to achieve higher recognition accuracy. Some preliminary experimental results demonstrate the effectiveness of the dataset in accurately recognizing handwritten English alphabet in the Indian context. This research can be extended and contributes to the field of pattern recognition and offers valuable insights for developing improved systems for handwriting recognition, particularly in diverse linguistic and cultural contexts.
Halaman 17 dari 352234