Large Language Models and Arabic Content: A Review
Haneh Rhel, Dmitri Roussinov
Over the past three years, the rapid advancement of Large Language Models (LLMs) has had a profound impact on multiple areas of Artificial Intelligence (AI), particularly in Natural Language Processing (NLP) across diverse languages, including Arabic. Although Arabic is considered one of the most widely spoken languages across 27 countries in the Arabic world and used as a second language in some other non-Arabic countries as well, there is still a scarcity of Arabic resources, datasets, and tools. Arabic NLP tasks face various challenges due to the complexities of the Arabic language, including its rich morphology, intricate structure, and diverse writing standards, among other factors. Researchers have been actively addressing these challenges, demonstrating that pre-trained Large Language Models (LLMs) trained on multilingual corpora achieve significant success in various Arabic NLP tasks. This study provides an overview of using large language models (LLMs) for the Arabic language, highlighting early pre-trained Arabic Language models across various NLP applications and their ability to handle diverse Arabic content tasks and dialects. It also provides an overview of how techniques like finetuning and prompt engineering can enhance the performance of these models. Additionally, the study summarizes common Arabic benchmarks and datasets while presenting our observations on the persistent upward trend in the adoption of LLMs.
„Jeśli jeszcze tylko zabijemy starca...” Farsa kryminalna w teatrze grecko-rzymskim w Egipcie (POxy. 413 verso)
Agnieszka Kotlińska-Toma
This paper presents a translatory proposition of the mime preserved on POxy. 413 verso (Moicheutria) into Polish, along with a concise analysis of the theatrical and literary issues that are observed in this ancient genre.
Philology. Linguistics, Greek language and literature. Latin language and literature
WHAT-IF: Exploring Branching Narratives by Meta-Prompting Large Language Models
Runsheng "Anson" Huang, Lara J. Martin, Chris Callison-Burch
WHAT-IF -- Writing a Hero's Alternate Timeline through Interactive Fiction -- is a system that uses zero-shot meta-prompting to create branching narratives from a prewritten story. Played as an interactive fiction (IF) game, WHAT-IF lets the player choose between decisions that the large language model (LLM) GPT-4 generates as possible branches in the story. Starting with an existing linear plot as input, a branch is created at each key decision taken by the main character. By meta-prompting the LLM to consider the major plot points from the story, the system produces coherent and well-structured alternate storylines. WHAT-IF stores the branching plot tree in a graph which helps it to both keep track of the story for prompting and maintain the structure for the final IF system. A demo of WHAT-IF can be found at https://what-if-game.github.io/.
PPTC-R benchmark: Towards Evaluating the Robustness of Large Language Models for PowerPoint Task Completion
Zekai Zhang, Yiduo Guo, Yaobo Liang
et al.
The growing dependence on Large Language Models (LLMs) for finishing user instructions necessitates a comprehensive understanding of their robustness to complex task completion in real-world situations. To address this critical need, we propose the PowerPoint Task Completion Robustness benchmark (PPTC-R) to measure LLMs' robustness to the user PPT task instruction and software version. Specifically, we construct adversarial user instructions by attacking user instructions at sentence, semantic, and multi-language levels. To assess the robustness of Language Models to software versions, we vary the number of provided APIs to simulate both the newest version and earlier version settings. Subsequently, we test 3 closed-source and 4 open-source LLMs using a benchmark that incorporates these robustness settings, aiming to evaluate how deviations impact LLMs' API calls for task completion. We find that GPT-4 exhibits the highest performance and strong robustness in our benchmark, particularly in the version update and the multilingual settings. However, we find that all LLMs lose their robustness when confronted with multiple challenges (e.g., multi-turn) simultaneously, leading to significant performance drops. We further analyze the robustness behavior and error reasons of LLMs in our benchmark, which provide valuable insights for researchers to understand the LLM's robustness in task completion and develop more robust LLMs and agents. We release the code and data at \url{https://github.com/ZekaiGalaxy/PPTCR}.
Decreto ateniese relativo alla concessione della sitesis
Tentori Montalto, Marco, Cardinale, Sandy, Pizzoli, Lorenzo
Il decreto ateniese qui discusso presenta alcuni aspetti problematici, tra cui la datazione, che oscilla tra il 440 e il 424 a.C. e lo stesso contenuto del documento. Il decreto, molto lacunoso, garantisce il pasto a spese pubbliche (sitesis) per alcuni beneficiari, tra cui si annoverano i discendenti dei tirannicidi, Armodio e Aristogitone; forse i pythochrestoi o, meglio, gli indovini scelti da Apollo; i vincitori di almeno uno dei quattro agoni della periodos e probabilmente gli strateghi. Infine si discute l'eventualità della concessione della proedria insieme alla sitesis.
Ancient history, Greek philology and language
Leveraging Large Language Models for Enhanced Product Descriptions in eCommerce
Jianghong Zhou, Bo Liu, Jhalak Nilesh Acharya Yao Hong
et al.
In the dynamic field of eCommerce, the quality and comprehensiveness of product descriptions are pivotal for enhancing search visibility and customer engagement. Effective product descriptions can address the 'cold start' problem, align with market trends, and ultimately lead to increased click-through rates. Traditional methods for crafting these descriptions often involve significant human effort and may lack both consistency and scalability. This paper introduces a novel methodology for automating product description generation using the LLAMA 2.0 7B language model. We train the model on a dataset of authentic product descriptions from Walmart, one of the largest eCommerce platforms. The model is then fine-tuned for domain-specific language features and eCommerce nuances to enhance its utility in sales and user engagement. We employ multiple evaluation metrics, including NDCG, customer click-through rates, and human assessments, to validate the effectiveness of our approach. Our findings reveal that the system is not only scalable but also significantly reduces the human workload involved in creating product descriptions. This study underscores the considerable potential of large language models like LLAMA 2.0 7B in automating and optimizing various facets of eCommerce platforms, offering significant business impact, including improved search functionality and increased sales.
WizardLM: Empowering large pre-trained language models to follow complex instructions
Can Xu, Qingfeng Sun, Kai Zheng
et al.
Training large language models (LLMs) with open-domain instruction following data brings colossal success. However, manually creating such instruction data is very time-consuming and labor-intensive. Moreover, humans may struggle to produce high-complexity instructions. In this paper, we show an avenue for creating large amounts of instruction data with varying levels of complexity using LLM instead of humans. Starting with an initial set of instructions, we use our proposed Evol-Instruct to rewrite them step by step into more complex instructions. Then, we mix all generated instruction data to fine-tune LLaMA. We call the resulting model WizardLM. Human evaluations on a complexity-balanced test bed and Vicuna's testset show that instructions from Evol-Instruct are superior to human-created ones. By analyzing the human evaluation results of the high complexity part, we demonstrate that outputs from our WizardLM are preferred to outputs from OpenAI ChatGPT. In GPT-4 automatic evaluation, WizardLM achieves more than 90\% capacity of ChatGPT on 17 out of 29 skills. Even though WizardLM still lags behind ChatGPT in some aspects, our findings suggest that fine-tuning with AI-evolved instructions is a promising direction for enhancing LLMs. Our code and data are public at https://github.com/nlpxucan/WizardLM
Dedica di Messeni e Naupatti a Olimpia
Tronchin, Davide
Nel complesso monumentale di Olimpia Messeni e Naupatti dedicarono una Nike di marmo a Zeus Olimpio. L’offerta votiva era corredata da un’iscrizione, incisa in alfabeto ‘azzurro scuro’ e iscritta in dialetto dorico, che si compone di quattro linee: le prime due contengono la dedica della Nike da parte di Messeni e Naupatti, le ultime due la ‘firma’ dello scultore, Paionios di Mende. Il motivo della dedica va ricercato negli eventi bellici occorsi nelle fasi conclusive della cosiddetta ‘guerra archidamica’, indicativamente tra 425 e 421 a.C. I rapporti di alleanza tra Messeni e Ateniesi consentono di chiarire la presenza della Nike di Paionios a Olimpia: a questa e a un’analoga offerta a Delfi corrispondono due Nikai dedicate dagli Ateniesi sull’Acropoli, per commemorare la campagna etolica e le imprese di Pilo e Sfacteria.
Ancient history, Greek philology and language
Quantum Natural Language Generation on Near-Term Devices
Amin Karamlou, Marcel Pfaffhauser, James Wootton
The emergence of noisy medium-scale quantum devices has led to proof-of-concept applications for quantum computing in various domains. Examples include Natural Language Processing (NLP) where sentence classification experiments have been carried out, as well as procedural generation, where tasks such as geopolitical map creation, and image manipulation have been performed. We explore applications at the intersection of these two areas by designing a hybrid quantum-classical algorithm for sentence generation. Our algorithm is based on the well-known simulated annealing technique for combinatorial optimisation. An implementation is provided and used to demonstrate successful sentence generation on both simulated and real quantum hardware. A variant of our algorithm can also be used for music generation. This paper aims to be self-contained, introducing all the necessary background on NLP and quantum computing along the way.
Towards Pragmatic Production Strategies for Natural Language Generation Tasks
Mario Giulianelli
This position paper proposes a conceptual framework for the design of Natural Language Generation (NLG) systems that follow efficient and effective production strategies in order to achieve complex communicative goals. In this general framework, efficiency is characterised as the parsimonious regulation of production and comprehension costs while effectiveness is measured with respect to task-oriented and contextually grounded communicative goals. We provide concrete suggestions for the estimation of goals, costs, and utility via modern statistical methods, demonstrating applications of our framework to the classic pragmatic task of visually grounded referential games and to abstractive text summarisation, two popular generation tasks with real-world applications. In sum, we advocate for the development of NLG systems that learn to make pragmatic production decisions from experience, by reasoning about goals, costs, and utility in a human-like way.
Construction of English Resume Corpus and Test with Pre-trained Language Models
Chengguang Gan, Tatsunori Mori
Information extraction(IE) has always been one of the essential tasks of NLP. Moreover, one of the most critical application scenarios of information extraction is the information extraction of resumes. Constructed text is obtained by classifying each part of the resume. It is convenient to store these texts for later search and analysis. Furthermore, the constructed resume data can also be used in the AI resume screening system. Significantly reduce the labor cost of HR. This study aims to transform the information extraction task of resumes into a simple sentence classification task. Based on the English resume dataset produced by the prior study. The classification rules are improved to create a larger and more fine-grained classification dataset of resumes. This corpus is also used to test some current mainstream Pre-training language models (PLMs) performance.Furthermore, in order to explore the relationship between the number of training samples and the correctness rate of the resume dataset, we also performed comparison experiments with training sets of different train set sizes.The final multiple experimental results show that the resume dataset with improved annotation rules and increased sample size of the dataset improves the accuracy of the original resume dataset.
Reinforcement Learning and Bandits for Speech and Language Processing: Tutorial, Review and Outlook
Baihan Lin
In recent years, reinforcement learning and bandits have transformed a wide range of real-world applications including healthcare, finance, recommendation systems, robotics, and last but not least, the speech and natural language processing. While most speech and language applications of reinforcement learning algorithms are centered around improving the training of deep neural networks with its flexible optimization properties, there are still many grounds to explore to utilize the benefits of reinforcement learning, such as its reward-driven adaptability, state representations, temporal structures and generalizability. In this survey, we present an overview of recent advancements of reinforcement learning and bandits, and discuss how they can be effectively employed to solve speech and natural language processing problems with models that are adaptive, interactive and scalable.
Effectiveness of French Language Models on Abstractive Dialogue Summarization Task
Yongxin Zhou, François Portet, Fabien Ringeval
Pre-trained language models have established the state-of-the-art on various natural language processing tasks, including dialogue summarization, which allows the reader to quickly access key information from long conversations in meetings, interviews or phone calls. However, such dialogues are still difficult to handle with current models because the spontaneity of the language involves expressions that are rarely present in the corpora used for pre-training the language models. Moreover, the vast majority of the work accomplished in this field has been focused on English. In this work, we present a study on the summarization of spontaneous oral dialogues in French using several language specific pre-trained models: BARThez, and BelGPT-2, as well as multilingual pre-trained models: mBART, mBARThez, and mT5. Experiments were performed on the DECODA (Call Center) dialogue corpus whose task is to generate abstractive synopses from call center conversations between a caller and one or several agents depending on the situation. Results show that the BARThez models offer the best performance far above the previous state-of-the-art on DECODA. We further discuss the limits of such pre-trained models and the challenges that must be addressed for summarizing spontaneous dialogues.
Die Mär vom Fehlerquotienten
Thomas Doepner
Greek language and literature. Latin language and literature, Philology. Linguistics
„Deine Schrift lateinisch ...“
Julia Großekathöfer, Sabine Holländer, Franziska Weber
Greek language and literature. Latin language and literature, Philology. Linguistics
Sense representations for Portuguese: experiments with sense embeddings and deep neural language models
Jessica Rodrigues da Silva, Helena de Medeiros Caseli
Sense representations have gone beyond word representations like Word2Vec, GloVe and FastText and achieved innovative performance on a wide range of natural language processing tasks. Although very useful in many applications, the traditional approaches for generating word embeddings have a strict drawback: they produce a single vector representation for a given word ignoring the fact that ambiguous words can assume different meanings. In this paper, we explore unsupervised sense representations which, different from traditional word embeddings, are able to induce different senses of a word by analyzing its contextual semantics in a text. The unsupervised sense representations investigated in this paper are: sense embeddings and deep neural language models. We present the first experiments carried out for generating sense embeddings for Portuguese. Our experiments show that the sense embedding model (Sense2vec) outperformed traditional word embeddings in syntactic and semantic analogies task, proving that the language resource generated here can improve the performance of NLP tasks in Portuguese. We also evaluated the performance of pre-trained deep neural language models (ELMo and BERT) in two transfer learning approaches: feature based and fine-tuning, in the semantic textual similarity task. Our experiments indicate that the fine tuned Multilingual and Portuguese BERT language models were able to achieve better accuracy than the ELMo model and baselines.
Sueños religiosos en Lucrecio (DRN., V, 1161-1240)
Antonio Ruiz Castellanos
Analizamos un texto controvertido del De rerum natura (DRN) de Lucrecio (V, 1161-1240). En él se pregunta cómo surgió la idea de dios entre los primitivos ofreciéndose dos etiologías: Una legítima (vv. 1169-1182), que afirma que fue a través de las apariciones religiosas en sueños. Y otra no legítima, la religiosidad astral (1183-1240); la descalificación de ésta la han generalizado muchos editores a todo el texto. No argumentaremos de forma esotérica, ni freudiana, ni filosófica, sólo registramos la autonomía de la primera etiología mediante la sintaxis del texto y la utilización de las fuentes epicúreas, incluidos los papiros de Herculano.
Philology. Linguistics, Greek language and literature. Latin language and literature
6. evropský kongres novořeckých studií, Lund, 4.–7.10.2018
Markéta Kulhánková
History of Greece, Translating and interpreting
Prossenia per Aristotele figlio di Cheilonio
De Luca, Gaia
Si tratta di un decreto con il quale il consiglio e l’assemblea di Eretria attribuiscono i titoli di proxenos e di euergetes ad Aristotele figlio di Cheilonio e ai suoi fratelli o a un suo fratello (ll. 13-14).
Ancient history, Greek philology and language
LSCP: Enhanced Large Scale Colloquial Persian Language Understanding
Hadi Abdi Khojasteh, Ebrahim Ansari, Mahdi Bohlouli
Language recognition has been significantly advanced in recent years by means of modern machine learning methods such as deep learning and benchmarks with rich annotations. However, research is still limited in low-resource formal languages. This consists of a significant gap in describing the colloquial language especially for low-resourced ones such as Persian. In order to target this gap for low resource languages, we propose a "Large Scale Colloquial Persian Dataset" (LSCP). LSCP is hierarchically organized in a semantic taxonomy that focuses on multi-task informal Persian language understanding as a comprehensive problem. This encompasses the recognition of multiple semantic aspects in the human-level sentences, which naturally captures from the real-world sentences. We believe that further investigations and processing, as well as the application of novel algorithms and methods, can strengthen enriching computerized understanding and processing of low resource languages. The proposed corpus consists of 120M sentences resulted from 27M tweets annotated with parsing tree, part-of-speech tags, sentiment polarity and translation in five different languages.