{"results":[{"id":"ss_b5ef0f91663f0cbd6910dec9a890c138f7ec10e0","title":"Oscar: Object-Semantics Aligned Pre-training for Vision-Language Tasks","authors":[{"name":"Xiujun Li"},{"name":"Xi Yin"},{"name":"Chunyuan Li"},{"name":"Xiaowei Hu"},{"name":"Pengchuan Zhang"},{"name":"Lei Zhang"},{"name":"Lijuan Wang"},{"name":"Houdong Hu"},{"name":"Li Dong"},{"name":"Furu Wei"},{"name":"Yejin Choi"},{"name":"Jianfeng Gao"}],"abstract":"Large-scale pre-training methods of learning cross-modal representations on image-text pairs are becoming popular for vision-language tasks. While existing methods simply concatenate image region features and text features as input to the model to be pre-trained and use self-attention to learn image-text semantic alignments in a brute force manner, in this paper, we propose a new learning method Oscar (Object-Semantics Aligned Pre-training), which uses object tags detected in images as anchor points to significantly ease the learning of alignments. Our method is motivated by the observation that the salient objects in an image can be accurately detected, and are often mentioned in the paired text. We pre-train an Oscar model on the public corpus of 6.5 million text-image pairs, and fine-tune it on downstream tasks, creating new state-of-the-arts on six well-established vision-language understanding and generation tasks.","source":"Semantic Scholar","year":2020,"language":"en","subjects":["Computer Science"],"doi":"10.1007/978-3-030-58577-8_8","url":"https://www.semanticscholar.org/paper/b5ef0f91663f0cbd6910dec9a890c138f7ec10e0","is_open_access":true,"citations":2191,"published_at":"","score":94},{"id":"ss_f98788f32b0d33d200c9bc7d900d0ef39519c927","title":"Multi-task Learning Using Uncertainty to Weigh Losses for Scene Geometry and Semantics","authors":[{"name":"Alex Kendall"},{"name":"Y. Gal"},{"name":"R. Cipolla"}],"abstract":"Numerous deep learning applications benefit from multitask learning with multiple regression and classification objectives. In this paper we make the observation that the performance of such systems is strongly dependent on the relative weighting between each task's loss. Tuning these weights by hand is a difficult and expensive process, making multi-task learning prohibitive in practice. We propose a principled approach to multi-task deep learning which weighs multiple loss functions by considering the homoscedastic uncertainty of each task. This allows us to simultaneously learn various quantities with different units or scales in both classification and regression settings. We demonstrate our model learning per-pixel depth regression, semantic and instance segmentation from a monocular input image. Perhaps surprisingly, we show our model can learn multi-task weightings and outperform separate models trained individually on each task.","source":"Semantic Scholar","year":2017,"language":"en","subjects":["Computer Science"],"doi":"10.1109/CVPR.2018.00781","url":"https://www.semanticscholar.org/paper/f98788f32b0d33d200c9bc7d900d0ef39519c927","pdf_url":"https://arxiv.org/pdf/1705.07115","is_open_access":true,"citations":3850,"published_at":"","score":91},{"id":"ss_5966d7c7f60898d610812e24c64d4d57855ad86a","title":"Semantics derived automatically from language corpora contain human-like biases","authors":[{"name":"Aylin Caliskan"},{"name":"J. Bryson"},{"name":"Arvind Narayanan"}],"abstract":"Machines learn what people know implicitly AlphaGo has demonstrated that a machine can learn how to do things that people spend many years of concentrated study learning, and it can rapidly learn how to do them better than any human can. Caliskan et al. now show that machines can learn word associations from written texts and that these associations mirror those learned by humans, as measured by the Implicit Association Test (IAT) (see the Perspective by Greenwald). Why does this matter? Because the IAT has predictive value in uncovering the association between concepts, such as pleasantness and flowers or unpleasantness and insects. It can also tease out attitudes and beliefs—for example, associations between female names and family or male names and career. Such biases may not be expressed explicitly, yet they can prove influential in behavior. Science, this issue p. 183; see also p. 133 Computers can learn which words go together more or less often and can thus mimic human performance on a test of implicit bias. Machine learning is a means to derive artificial intelligence by discovering patterns in existing data. Here, we show that applying machine learning to ordinary human language results in human-like semantic biases. We replicated a spectrum of known biases, as measured by the Implicit Association Test, using a widely used, purely statistical machine-learning model trained on a standard corpus of text from the World Wide Web. Our results indicate that text corpora contain recoverable and accurate imprints of our historic biases, whether morally neutral as toward insects or flowers, problematic as toward race or gender, or even simply veridical, reflecting the status quo distribution of gender with respect to careers or first names. Our methods hold promise for identifying and addressing sources of bias in culture, including technology.","source":"Semantic Scholar","year":2016,"language":"en","subjects":["Computer Science","Medicine"],"doi":"10.1126/science.aal4230","url":"https://www.semanticscholar.org/paper/5966d7c7f60898d610812e24c64d4d57855ad86a","pdf_url":"https://researchportal.bath.ac.uk/files/168480066/CaliskanEtAl_authors_full.pdf","is_open_access":true,"citations":3048,"published_at":"","score":90},{"id":"ss_2dcfa0ab3d301f0358f46c280f0bf8614efdfd2c","title":"Beyond Transmitting Bits: Context, Semantics, and Task-Oriented Communications","authors":[{"name":"Deniz Gündüz"},{"name":"Zhijin Qin"},{"name":"Iñaki Estella Aguerri"},{"name":"Harpreet S. Dhillon"},{"name":"Zhaohui Yang"},{"name":"A. Yener"},{"name":"Kai‐Kit Wong"},{"name":"C. Chae"}],"abstract":"Communication systems to date primarily aim at reliably communicating bit sequences. Such an approach provides efficient engineering designs that are agnostic to the meanings of the messages or to the goal that the message exchange aims to achieve. Next generation systems, however, can be potentially enriched by folding message semantics and goals of communication into their design. Further, these systems can be made cognizant of the context in which communication exchange takes place, thereby providing avenues for novel design insights. This tutorial summarizes the efforts to date, starting from its early adaptations, semantic-aware and task-oriented communications, covering the foundations, algorithms and potential implementations. The focus is on approaches that utilize information theory to provide the foundations, as well as the significant role of learning in semantics and task-aware communications.","source":"Semantic Scholar","year":2022,"language":"en","subjects":["Computer Science","Mathematics"],"doi":"10.1109/JSAC.2022.3223408","url":"https://www.semanticscholar.org/paper/2dcfa0ab3d301f0358f46c280f0bf8614efdfd2c","pdf_url":"http://arxiv.org/pdf/2207.09353","is_open_access":true,"citations":723,"published_at":"","score":87.69},{"id":"ss_3a0e788268fafb23ab20da0e98bb578b06830f7d","title":"From Frequency to Meaning: Vector Space Models of Semantics","authors":[{"name":"Peter D. Turney"},{"name":"Patrick Pantel"}],"abstract":"Computers understand very little of the meaning of human language. This profoundly limits our ability to give instructions to computers, the ability of computers to explain their actions to us, and the ability of computers to analyse and process text. Vector space models (VSMs) of semantics are beginning to address these limits. This paper surveys the use of VSMs for semantic processing of text. We organize the literature on VSMs according to the structure of the matrix in a VSM. There are currently three broad classes of VSMs, based on term-document, word-context, and pair-pattern matrices, yielding three classes of applications. We survey a broad range of applications in these three categories and we take a detailed look at a specific open source project in each category. Our goal in this survey is to show the breadth of applications of VSMs for semantics, to provide a new perspective on VSMs for those who are already familiar with the area, and to provide pointers into the literature for those who are less familiar with the field.","source":"Semantic Scholar","year":2010,"language":"en","subjects":["Computer Science"],"doi":"10.1613/jair.2934","url":"https://www.semanticscholar.org/paper/3a0e788268fafb23ab20da0e98bb578b06830f7d","pdf_url":"https://jair.org/index.php/jair/article/download/10640/25440","is_open_access":true,"citations":3067,"published_at":"","score":84},{"id":"ss_dc0092d06ab76465431edfd51b08d823b7d1ff3f","title":"Closed-Form Factorization of Latent Semantics in GANs","authors":[{"name":"Yujun Shen"},{"name":"Bolei Zhou"}],"abstract":"A rich set of interpretable dimensions has been shown to emerge in the latent space of the Generative Adversarial Networks (GANs) trained for synthesizing images. In order to identify such latent dimensions for image editing, previous methods typically annotate a collection of synthesized samples and train linear classifiers in the latent space. However, they require a clear definition of the target attribute as well as the corresponding manual annotations, limiting their applications in practice. In this work, we examine the internal representation learned by GANs to reveal the underlying variation factors in an unsupervised manner. In particular, we take a closer look into the generation mechanism of GANs and further propose a closedform factorization algorithm for latent semantic discovery by directly decomposing the pre-trained weights. With a lightning-fast implementation, our approach is capable of not only finding semantically meaningful dimensions comparably to the state-of-the-art supervised methods, but also resulting in far more versatile concepts across multiple GAN models trained on a wide range of datasets.1","source":"Semantic Scholar","year":2020,"language":"en","subjects":["Computer Science"],"doi":"10.1109/CVPR46437.2021.00158","url":"https://www.semanticscholar.org/paper/dc0092d06ab76465431edfd51b08d823b7d1ff3f","pdf_url":"https://arxiv.org/pdf/2007.06600","is_open_access":true,"citations":644,"published_at":"","score":83.32},{"id":"ss_35b5ec70577bba9c37b76d2f11372f14e7fe6c6c","title":"The semantics of microglia activation: neuroinflammation, homeostasis, and stress","authors":[{"name":"S. Woodburn"},{"name":"J. Bollinger"},{"name":"E. Wohleb"}],"abstract":"Microglia are emerging as critical regulators of neuronal function and behavior in nearly every area of neuroscience. Initial reports focused on classical immune functions of microglia in pathological contexts, however, immunological concepts from these studies have been applied to describe neuro-immune interactions in the absence of disease, injury, or infection. Indeed, terms such as ‘microglia activation’ or ‘neuroinflammation’ are used ubiquitously to describe changes in neuro-immune function in disparate contexts; particularly in stress research, where these terms prompt undue comparisons to pathological conditions. This creates a barrier for investigators new to neuro-immunology and ultimately hinders our understanding of stress effects on microglia. As more studies seek to understand the role of microglia in neurobiology and behavior, it is increasingly important to develop standard methods to study and define microglial phenotype and function. In this review, we summarize primary research on the role of microglia in pathological and physiological contexts. Further, we propose a framework to better describe changes in microglia1 phenotype and function in chronic stress. This approach will enable more precise characterization of microglia in different contexts, which should facilitate development of microglia-directed therapeutics in psychiatric and neurological disease.","source":"Semantic Scholar","year":2021,"language":"en","subjects":["Medicine"],"doi":"10.1186/s12974-021-02309-6","url":"https://www.semanticscholar.org/paper/35b5ec70577bba9c37b76d2f11372f14e7fe6c6c","pdf_url":"https://jneuroinflammation.biomedcentral.com/track/pdf/10.1186/s12974-021-02309-6","is_open_access":true,"citations":585,"published_at":"","score":82.55},{"id":"ss_eb1cc1a15764746bf26e03b0ffd3112c1cbe268b","title":"Lexical Semantics","authors":[{"name":"David Smith"},{"name":"Jason Eisner"}],"abstract":"General rights Unless other specific re-use rights are stated the following general rights apply: Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal","source":"Semantic Scholar","year":2020,"language":"en","subjects":null,"doi":"10.1007/978-3-642-41464-0_15","url":"https://www.semanticscholar.org/paper/eb1cc1a15764746bf26e03b0ffd3112c1cbe268b","pdf_url":"https://doi.org/10.1093/obo/9780199772810-0078","is_open_access":true,"citations":617,"published_at":"","score":82.50999999999999},{"id":"ss_c200a54eed60bad1c7ed8d7de2b2c1ff749f5d78","title":"The Lambda Calculus: Its Syntax and Semantics","authors":[{"name":"H. Barendregt"}],"abstract":"","source":"Semantic Scholar","year":1985,"language":"en","subjects":["Mathematics","Computer Science"],"url":"https://www.semanticscholar.org/paper/c200a54eed60bad1c7ed8d7de2b2c1ff749f5d78","is_open_access":true,"citations":3979,"published_at":"","score":80},{"id":"ss_0e3e3c3d8ae5cb7c4636870d69967c197484d3bb","title":"Verb Semantics and Lexical Selection","authors":[{"name":"Zhibiao Wu"},{"name":"Martha Palmer"}],"abstract":"This paper will focus on the semantic representation of verbs in computer systems and its impact on lexical selection problems in machine translation (MT). Two groups of English and Chinese verbs are examined to show that lexical selection must be based on interpretation of the sentences as well as selection restrictions placed on the verb arguments. A novel representation scheme is suggested, and is compared to representations with selection restrictions used in transfer-based MT. We see our approach as closely aligned with knowledge-based MT approaches (KBMT), and as a separate component that could be incorporated into existing systems. Examples and experimental results will show that, using this scheme, inexact matches can achieve correct lexical selection.","source":"Semantic Scholar","year":1994,"language":"en","subjects":["Computer Science"],"doi":"10.3115/981732.981751","url":"https://www.semanticscholar.org/paper/0e3e3c3d8ae5cb7c4636870d69967c197484d3bb","pdf_url":"https://dl.acm.org/doi/pdf/10.3115/981732.981751","is_open_access":true,"citations":4013,"published_at":"","score":80},{"id":"ss_0dc0d0cfebc2ea1f88c64a159a76d0cb8258bff9","title":"Semantics in generative grammar","authors":[{"name":"I. Heim"},{"name":"A. Kratzer"}],"abstract":"","source":"Semantic Scholar","year":1998,"language":"en","subjects":["Computer Science"],"url":"https://www.semanticscholar.org/paper/0dc0d0cfebc2ea1f88c64a159a76d0cb8258bff9","is_open_access":true,"citations":2991,"published_at":"","score":80},{"id":"ss_d42a29e6977c28f7bf23d63b00c48f2e9100403e","title":"The Semantics of a Simple Language for Parallel Programming","authors":[{"name":"G. Kahn"}],"abstract":"","source":"Semantic Scholar","year":1974,"language":"en","subjects":["Computer Science"],"url":"https://www.semanticscholar.org/paper/d42a29e6977c28f7bf23d63b00c48f2e9100403e","is_open_access":true,"citations":2774,"published_at":"","score":80},{"id":"ss_2a41681900f295a3098592c1040b35b6e3767c96","title":"SIMPLIcity: Semantics-Sensitive Integrated Matching for Picture LIbraries","authors":[{"name":"James Ze Wang"},{"name":"Jia Li"},{"name":"G. Wiederhold"}],"abstract":"We present here SIMPLIcity (semantics-sensitive integrated matching for picture libraries), an image retrieval system, which uses semantics classification methods, a wavelet-based approach for feature extraction, and integrated region matching based upon image segmentation. An image is represented by a set of regions, roughly corresponding to objects, which are characterized by color, texture, shape, and location. The system classifies images into semantic categories. Potentially, the categorization enhances retrieval by permitting semantically-adaptive searching methods and narrowing down the searching range in a database. A measure for the overall similarity between images is developed using a region-matching scheme that integrates properties of all the regions in the images. The application of SIMPLIcity to several databases has demonstrated that our system performs significantly better and faster than existing ones. The system is fairly robust to image alterations.","source":"Semantic Scholar","year":2000,"language":"en","subjects":["Computer Science"],"doi":"10.1109/34.955109","url":"https://www.semanticscholar.org/paper/2a41681900f295a3098592c1040b35b6e3767c96","pdf_url":"http://www-db.stanford.edu/~wangz/project/imsearch/SIMPLIcity/TPAMI/wang2.pdf","is_open_access":true,"citations":2556,"published_at":"","score":80},{"id":"ss_99b1804689ec67dbc794e6707262901dd76085a8","title":"The semantics of definite and indefinite noun phrases : a dissertation","authors":[{"name":"I. Heim"}],"abstract":"","source":"Semantic Scholar","year":1982,"language":"en","subjects":["Computer Science"],"url":"https://www.semanticscholar.org/paper/99b1804689ec67dbc794e6707262901dd76085a8","is_open_access":true,"citations":3032,"published_at":"","score":80},{"id":"ss_620ce06151f6a7789d9e597ef2b0df73636c8c6b","title":"A structural approach to operational semantics","authors":[{"name":"G. Plotkin"}],"abstract":"","source":"Semantic Scholar","year":2004,"language":"en","subjects":["Computer Science"],"doi":"10.1016/j.jlap.2004.05.001","url":"https://www.semanticscholar.org/paper/620ce06151f6a7789d9e597ef2b0df73636c8c6b","pdf_url":"https://doi.org/10.1016/j.jlap.2004.05.001","is_open_access":true,"citations":2632,"published_at":"","score":80},{"id":"ss_6e07619398452c82839503cd25c1fc014f5dada7","title":"The Stable Model Semantics for Logic Programming","authors":[{"name":"M. Gelfond"},{"name":"V. Lifschitz"}],"abstract":"","source":"Semantic Scholar","year":1988,"language":"en","subjects":["Mathematics","Computer Science"],"url":"https://www.semanticscholar.org/paper/6e07619398452c82839503cd25c1fc014f5dada7","is_open_access":true,"citations":2441,"published_at":"","score":80},{"id":"ss_72951c265944bc8da0386eff0c978e0915177ace","title":"Toward a Cognitive Semantics","authors":[{"name":"L. Talmy"}],"abstract":"","source":"Semantic Scholar","year":2003,"language":"en","subjects":["Psychology"],"url":"https://www.semanticscholar.org/paper/72951c265944bc8da0386eff0c978e0915177ace","is_open_access":true,"citations":2083,"published_at":"","score":80},{"id":"ss_ffd87206430e4d7657a77d8eff484d335ac4056c","title":"3D-FRONT: 3D Furnished Rooms with layOuts and semaNTics","authors":[{"name":"Huan Fu"},{"name":"Bowen Cai"},{"name":"Lin Gao"},{"name":"Ling-Xiao Zhang"},{"name":"Cao Li"},{"name":"Zengqi Xun"},{"name":"Chengyue Sun"},{"name":"Yiyun Fei"},{"name":"Yu-qiong Zheng"},{"name":"Ying Li"},{"name":"Yi Liu"},{"name":"Peng Liu"},{"name":"Lin Ma"},{"name":"Le Weng"},{"name":"Xiaohang Hu"},{"name":"Xin Ma"},{"name":"Qian Qian"},{"name":"Rongfei Jia"},{"name":"Binqiang Zhao"},{"name":"H. Zhang"}],"abstract":"We introduce 3D-FRONT (3D Furnished Rooms with layOuts and semaNTics), a new, large-scale, and comprehensive repository of synthetic indoor scenes highlighted by professionally designed layouts and a large number of rooms populated by high-quality textured 3D models with style compatibility. From layout semantics down to texture details of individual objects, our dataset is freely available to the academic community and beyond. Currently, 3D-FRONT contains 6,813 CAD houses, where 18,968 rooms diversely furnished by 3D objects, far surpassing all publicly available scene datasets. The 13,151 furniture objects all come with high-quality textures. While the floorplans and layout designs (i.e., furniture arrangements) are directly sourced from professional creations, the interior designs in terms of furniture styles, color, and textures have been carefully curated based on a recommender system we develop to attain consistent styles as expert designs. Furthermore, we release Trescope, a light-weight rendering tool, to support benchmark rendering of 2D images and annotations from 3D-FRONT. We demonstrate two applications, interior scene synthesis and texture synthesis, that are especially tailored to the strengths of our new dataset.","source":"Semantic Scholar","year":2020,"language":"en","subjects":["Computer Science"],"doi":"10.1109/ICCV48922.2021.01075","url":"https://www.semanticscholar.org/paper/ffd87206430e4d7657a77d8eff484d335ac4056c","pdf_url":"http://arxiv.org/pdf/2011.09127","is_open_access":true,"citations":402,"published_at":"","score":76.06},{"id":"ss_9d2c9a7c3ff8f760b43fdd6a7082069a467c8956","title":"Adapting Large Language Models by Integrating Collaborative Semantics for Recommendation","authors":[{"name":"Bowen Zheng"},{"name":"Yupeng Hou"},{"name":"Hongyu Lu"},{"name":"Yu Chen"},{"name":"Wayne Xin Zhao"},{"name":"Ji-rong Wen"}],"abstract":"Recently, large language models (LLMs) have shown great potential in recommender systems, either improving existing recommendation models or serving as the backbone. However, there exists a large semantic gap between LLMs and recommender systems, since items to be recommended are often indexed by discrete identifiers (item ID) out of the LLM's vocabulary. In essence, LLMs capture language semantics while recommender systems imply collaborative semantics, making it difficult to sufficiently leverage the model capacity of LLMs for recommendation. To address this challenge, in this paper, we propose a new LLM-based recommendation model called LC-Rec, which can better integrate language and collaborative semantics for recommender systems. Our approach can directly generate items from the entire item set for recommendation, without relying on candidate items. Specifically, we make two major contributions in our approach. For item indexing, we design a learning-based vector quantization method with uniform semantic mapping, which can assign meaningful and non-conflicting IDs (called item indices) for items. For alignment tuning, we propose a series of specially designed tuning tasks to enhance the integration of collaborative semantics in LLMs. Our fine-tuning tasks enforce LLMs to deeply integrate language and collaborative semantics (characterized by the learned item indices), so as to achieve an effective adaptation to recommender systems. Extensive experiments demonstrate the effectiveness of our method, showing that our approach can outperform a number of competitive baselines including traditional recommenders and existing LLM-based recommenders. Our code is available at https://github.com/RUCAIBox/LC-Rec/.","source":"Semantic Scholar","year":2023,"language":"en","subjects":["Computer Science"],"doi":"10.1109/ICDE60146.2024.00118","url":"https://www.semanticscholar.org/paper/9d2c9a7c3ff8f760b43fdd6a7082069a467c8956","pdf_url":"https://arxiv.org/pdf/2311.09049","is_open_access":true,"citations":291,"published_at":"","score":75.72999999999999},{"id":"ss_5744f56d3253bd7c4341d36de40a93fceaa266b3","title":"Semantics-aware BERT for Language Understanding","authors":[{"name":"Zhuosheng Zhang"},{"name":"Yuwei Wu"},{"name":"Zhao Hai"},{"name":"Z. Li"},{"name":"Shuailiang Zhang"},{"name":"Xi Zhou"},{"name":"Xiang Zhou"}],"abstract":"The latest work on language representations carefully integrates contextualized features into language model training, which enables a series of success especially in various machine reading comprehension and natural language inference tasks. However, the existing language representation models including ELMo, GPT and BERT only exploit plain context-sensitive features such as character or word embeddings. They rarely consider incorporating structured semantic information which can provide rich semantics for language representation. To promote natural language understanding, we propose to incorporate explicit contextual semantics from pre-trained semantic role labeling, and introduce an improved language representation model, Semantics-aware BERT (SemBERT), which is capable of explicitly absorbing contextual semantics over a BERT backbone. SemBERT keeps the convenient usability of its BERT precursor in a light fine-tuning way without substantial task-specific modifications. Compared with BERT, semantics-aware BERT is as simple in concept but more powerful. It obtains new state-of-the-art or substantially improves results on ten reading comprehension and language inference tasks.","source":"Semantic Scholar","year":2019,"language":"en","subjects":["Computer Science"],"doi":"10.1609/AAAI.V34I05.6510","url":"https://www.semanticscholar.org/paper/5744f56d3253bd7c4341d36de40a93fceaa266b3","pdf_url":"https://ojs.aaai.org/index.php/AAAI/article/download/6510/6366","is_open_access":true,"citations":395,"published_at":"","score":74.85}],"total":331733,"page":1,"page_size":20,"sources":["DOAJ","CrossRef","arXiv","Semantic Scholar"],"query":"Semantics"}