AAAI 2018 分析
AAAI 2018 分析
word embedding
Learning Sentiment-Specific Word Embedding via Global Sentiment Representation
Context-based word embedding learning approaches can model rich semantic and syntactic information.
However, it is problematic for sentiment analysis because the words with similar contexts but opposite sentiment polarities, such as good and bad, are mapped into close word vectors in the embedding space.
Recently, some sentiment embedding learning methods have been proposed, but most of them are designed to work well on sentence-level texts.
Directly applying those models to document-level texts often leads to unsatisfied results.
To address this issue, we present a sentiment-specific word embedding learning architecture that utilizes local context informationas well as global sentiment representation.
The architecture is applicable for both sentence-level and document-level texts.
We take global sentiment representation as a simple average of word embeddings in the text, and use a corruption strategy as a sentiment-dependent regularization.
Extensive experiments conducted on several benchmark datasets demonstrate that the proposed architecture outperforms the state-of-the-art methods for sentiment classification.
《通过全局情绪表示学习特定情绪词的嵌入》
基于上下文的词嵌入学习方法可以对丰富的语义和句法信息进行建模。
然而,对于情感分析却存在问题,因为在嵌入空间中,具有相似语境但对立情感极性的词(如好词和坏词)被映射到封闭词向量中。
近年来,人们提出了一些情绪嵌入学习方法,但大多数方法都是为了在句子级的文本中发挥作用。
直接将这些模型应用于文档级文本通常会导致不满意的结果。
为了解决这个问题,我们提出了一种情绪特定的词嵌入学习架构,它利用了局部的背景信息以及全局情绪表示。
该体系结构适用于句子级和文档级文本。
我们将全球情绪表征作为文字嵌入的简单平均值,并将腐败策略作为情绪依赖的规范化。
在多个基准数据集上进行的大量实验表明,所提出的架构优于最先进的情感分类方法。
Using k-Way Co-Occurrences for Learning Word Embeddings
Co-occurrences between two words provide useful insights into the semantics of those words. Consequently, numerous prior work on word embedding learning has used co-occurrences between two words as the training signal for learning word embeddings. However, in natural language texts it is common for multiple words to be related and co-occurring in the same context. We extend the notion of co-occurrences to cover k(≥2)-way co-occurrences among a set of k-words. Specifically, we prove a theoretical relationship between the joint probability of k(≥2) words, and the sum of l_2 norms of their embeddings. Next, we propose a learning objective motivated by our theoretical result that utilize k-way co-occurrences for learning word embeddings. Our experimental results show that the derived theoretical relationship does indeed hold empirically, and despite data sparsity, for some smaller k(≤5) values, k-way embeddings perform comparably or better than 2-way embeddings in a range of tasks.
Semantic Structure-Based Word Embedding by Incorporating Concept Convergence and Word Divergence
Representing the semantics of words is a fundamental task in text processing.
Several research studies have shown that text and knowledge bases (KBs) are complementary sources for word embedding learning.
Most existing methods only consider relationships within word-pairs in the usage of KBs.
We argue that the structural information of well-organized words within the KBs is able to convey more effective and stable knowledge in capturing semantics of words.
In this paper, we propose a semantic structure-based word embedding method, and introduce concept convergence and word divergence to reveal semantic structures in the word embedding learning process.
To assess the effectiveness of our method, we use WordNet for training and conduct extensive experiments on word similarity, word analogy, text classification and query expansion.
The experimental results show that our method outperforms state-of-the-art methods, including the methods trained solely on the corpus, and others trained on the corpus and the KBs.
《基于语义结构的概念收敛和词嵌入发散》
表示词的语义是文本处理中的一项基本任务。
一些研究表明,文本和知识库(KBS)是词嵌入学习的补充源。
在使用KBS时,大多数现有方法只考虑单词对儿内部的关系。
我们认为,知识库中组织良好的单词的结构信息能够更有效、更稳定地传递捕获单词语义的知识。
本文提出了一种基于语义结构的嵌入方法,并引入概念收敛和单词发散来揭示嵌入学习过程中的语义结构。
为了评估该方法的有效性,我们使用WordNet进行了训练,并在单词相似性、单词类比、文本分类和查询扩展等方面进行了广泛的实验。
实验结果表明,我们的方法优于最先进的方法,包括只在语料库上训练的方法,以及在语料库和KBS上训练的方法。
Spectral Word Embedding with Negative Sampling
In this work, we investigate word embedding algorithms in the context of natural language processing. In particular, we examine the notion of ``negative examples'', the unobserved or insignificant word-context co-occurrences, in spectral methods. we provide a new formulation for the word embedding problem by proposing a new intuitive objective function that perfectly justifies the use of negative examples. In fact, our algorithm not only learns from the important word-context co-occurrences, but also it learns from the abundance of unobserved or insignificant co-occurrences to improve the distribution of words in the latent embedded space. We analyze the algorithm theoretically and provide an optimal solution for the problem using spectral analysis. We have trained various word embedding algorithms on articles of Wikipedia with 2.1 billion tokens and show that negative sampling can boost the quality of spectral methods. Our algorithm provides results as good as the state-of-the-art but in a much faster and efficient way.
Chinese LIWC Lexicon Expansion via Hierarchical Classification of Word Embeddings with Sememe Attention
Linguistic Inquiry and Word Count (LIWC) is a word counting software tool which has been used for quantitative text analysis in many fields.
Due to its success and popularity, the core lexicon has been translated into Chinese and many other languages.
However, the lexicon only contains several thousand of words, which is deficient compared with the number of common words in Chinese.
Current approaches often require manually expanding the lexicon, but it often takes too much time and requires linguistic experts to extend the lexicon.
To address this issue, we propose to expand the LIWC lexicon automatically.
Specifically, we consider it as a hierarchical classification problem and utilize the Sequence-to-Sequence model to classify words in the lexicon.
Moreover, we use the sememe information with the attention mechanism to capture the exact meanings of a word, so that we can expand a more precise and comprehensive lexicon.
The experimental results show that our model has a better understanding of word meanings with the help of sememes and achieves significant and consistent improvements compared with the state-of-the-art methods.
The source code of this paper can be obtained from https://github.com/thunlp/Auto_CLIWC.
《基于词组嵌入层次分类的汉语词汇扩展》
语言查询和字数统计(LIWC)是一种在许多领域用于定量文本分析的字数统计软件工具。
由于它的成功和普及,核心词汇已被翻译成汉语和许多其他语言。
然而,词汇中只有几千个词,与汉语常用词相比,这是一个不足之处。
目前的方法通常需要手动扩展词汇,但通常需要花费大量的时间,并且需要语言专家来扩展词汇。
为了解决这个问题,我们提出自动扩展LIWC词典。
具体地说,我们将其视为一个层次分类问题,并利用序列-序列模型对词汇进行分类。
此外,我们还利用具有注意机制的义位信息来捕获一个词的确切含义,从而扩展出一个更精确、更全面的词汇。
实验结果表明,我们的模型在义原的帮助下,对词义有了更好的理解,与目前最先进的方法相比,得到了显著和一致的改进。
本文的源代码可从 https://github.com/thunlp/auto_cliwc 获得。
Training and Evaluating Improved Dependency-Based Word Embeddings
Word embedding has been widely used in many natural language processing tasks. In this paper, we focus on learning word embeddings through selective higher-order relationships in sentences to improve the embeddings to be less sensitive to local context and more accurate in capturing semantic compositionality. We present a novel multi-order dependency-based strategy to composite and represent the context under several essential constraints. In order to realize selective learning from the word contexts, we automatically assign the strengths of different dependencies between co-occurred words in the stochastic gradient descent process. We evaluate and analyze our proposed approach using several direct and indirect tasks for word embeddings. Experimental results demonstrate that our embeddings are competitive to or better than state-of-the-art methods and significantly outperform other methods in terms of context stability. The output weights and representations of dependencies obtained in our embedding model conform to most of the linguistic characteristics and are valuable for many downstream tasks.
word representation
Learning Multimodal Word Representation via Dynamic Fusion Methods
Multimodal models have been proven to outperform text-based models on learning semantic word representations. Almost all previous multimodal models typically treat the representations from different modalities equally. However, it is obvious that information from different modalities contributes differently to the meaning of words. This motivates us to build a multimodal model that can dynamically fuse the semantic representations from different modalities according to different types of words. To that end, we propose three novel dynamic fusion methods to assign importance weights to each modality, in which weights are learned under the weak supervision of word association pairs. The extensive experiments have demonstrated that the proposed methods outperform strong unimodal baselines and state-of-the-art multimodal models.
Learning Multi-Modal Word Representation Grounded in Visual Context
Representing the semantics of words is a long-standing problem for the natural language processing community.
Most methods compute word semantics given their textual context in large corpora.
More recently, researchers attempted to integrate perceptual and visual features.
Most of these works consider the visual appearance of objects to enhance word representations but they ignore the visual environment and context in which objects appear.
We propose to unify text-based techniques with vision-based techniques by simultaneously leveraging textual and visual context to learn multimodal word embeddings.
We explore various choices for what can serve as a visual context and present an end-to-end method to integrate visual context elements in a multimodal skip-gram model.
We provide experiments and extensive analysis of the obtained results.
《基于视觉语境的多模态词表示学习》
词的语义表示是自然语言处理界长期存在的问题。
大多数方法根据大语料库中的文本上下文计算单词语义。
最近,研究人员试图整合感知和视觉特征。
这些作品大多考虑对象的视觉外观以增强单词表示,但它们忽略了对象出现的视觉环境和上下文。
我们建议将基于文本的技术与基于视觉的技术统一起来,同时利用文本和视觉上下文来学习多模态单词嵌入。
我们探讨了作为可视上下文的各种选择,并提出了一种将可视上下文元素集成到多模式跳图模型中的端到端方法。
我们对所得结果进行了实验和广泛的分析。
AAAI 2018 分析的更多相关文章
- AAAI 2018 论文 | 蚂蚁金服公开最新基于笔画的中文词向量算法
AAAI 2018 论文 | 蚂蚁金服公开最新基于笔画的中文词向量算法 2018-01-18 16:13蚂蚁金服/雾霾/人工智能 导读:词向量算法是自然语言处理领域的基础算法,在序列标注.问答系统和机 ...
- AAAI 2019 分析
AAAI 2019 分析 Google Scholar 订阅 CoKE : Word Sense Induction Using Contextualized Knowledge Embeddings ...
- 回首2018 | 分析型数据库AnalyticDB: 不忘初心 砥砺前行
题记 分析型数据库AnalyticDB(下文简称ADB),是阿里巴巴自主研发.唯一经过超大规模以及核心业务验证的PB级实时数据仓库.截止目前,现有外部支撑客户既包括传统的大中型企业和政府机构,也包括众 ...
- EAS:基于网络转换的神经网络结构搜索 | AAAI 2018
论文提出经济实惠且高效的神经网络结构搜索算法EAS,使用RL agent作为meta-controller,学习通过网络变换进行结构空间探索.从指定的网络开始,通过function-preservin ...
- cw2vec理论及其实现
导读 本文对AAAI 2018(Association for the Advancement of Artificial Intelligence 2018)高分录用的一篇中文词向量论文(cw2ve ...
- AAAI2018中的自注意力机制(Self-attention Mechanism)
近年来,注意力(Attention)机制被广泛应用到基于深度学习的自然语言处理(NLP)各个任务中.随着注意力机制的深入研究,各式各样的attention被研究者们提出,如单个.多个.交互式等等.去年 ...
- 『高性能模型』HetConv: HeterogeneousKernel-BasedConvolutionsforDeepCNNs
论文地址:HetConv 一.现有网络加速技术 1.卷积加速技术 作者对已有的新型卷积划分如下:标准卷积.Depthwise 卷积.Pointwise 卷积.群卷积(相关介绍见『高性能模型』深度可分离 ...
- zz【清华NLP】图神经网络GNN论文分门别类,16大应用200+篇论文最新推荐
[清华NLP]图神经网络GNN论文分门别类,16大应用200+篇论文最新推荐 图神经网络研究成为当前深度学习领域的热点.最近,清华大学NLP课题组Jie Zhou, Ganqu Cui, Zhengy ...
- Paper Mark
BigCowPeking的CSDN博客 https://blog.csdn.net/wfei101/article/category/7120809 Low Rank Structure of Lea ...
随机推荐
- MWPhotoBrowser.bundle: bundle format unrecognized, invalid, or unsuitable
今天在github下载了MWPhotoBrowser的demo想跑一下,却发现报了MWPhotoBrowser.bundle: bundle format unrecognized, invalid, ...
- cassandra基本操作
basic operate1. 创建keyspace基本语句:CREATE KEYSPACE <identifier> WITH <properties>案例:CREATE K ...
- 自动化运维——HelloWorld(一)
1.HelloWorld vi first_shell.sh #!/bin/bash #Filename: first_shell.sh #auto echo hello world! #by aut ...
- YARN组件详细介绍
一.ResourceManager 内部主要有两个组件: 1.Scheduler:这个组件完全是插拔式的,用户可以根据自己的需求实现不同的调度器,目前YARN提供了FIFO.容量以及公平调度器.这个组 ...
- Codeforces Round #575 (Div. 3) B. Odd Sum Segments (构造,数学)
B. Odd Sum Segments time limit per test3 seconds memory limit per test256 megabytes inputstandard in ...
- java动态代理(JDK和cglib)(转载)
原文地址:http://www.cnblogs.com/jqyp/archive/2010/08/20/1805041.html 高亮部分是我的理解. JAVA的动态代理 代理模式 代理模式是常用的j ...
- vue和cordova项目整合打包,并实现vue调用android的相机的demo
经过网上查找很多资料,发现很多只有vue+cordova的项目整合,但是vue使用cordova插件的文章很少,现在把从创建cordova和创建vue到vue使用插件到项目打包到android手机运行 ...
- Eclipse/MyEclipse超全常用快捷键汇总,绝对实用
[MyEclipse CI 2019.4.0安装包下载] Eclipse的编辑功能非常强大,掌握了Eclipse快捷键功能,能大大提高开发效率.Eclipse中有如下一些和编辑相关的快捷键. 常用快捷 ...
- swiper保存
//初始化swiper var index=0; if(sessionStorage.getItem("index")){ index=sessionStorage.getItem ...
- Redis如何实现高可用【主从复制+哨兵机制+keepalived】
实现redis高可用机制的一些方法: 保证redis高可用机制需要redis主从复制.redis持久化机制.哨兵机制.keepalived等的支持. 主从复制的作用:数据备份.读写分离.分布式集群.实 ...