[2017 - 2018 ACL] 对话系统论文研究点整理
(论文编号及摘要见 [2017 ACL] 对话系统. [2018 ACL Long] 对话系统. 论文标题[]中最后的数字表示截止2019.1.21 google被引次数)
1. Domain Adaptation:
challenges:
(a) data shifts (syn -> live user data; stale -> current) cause distribution mismatch bet train and eval. -> 2017.1
(b) reestimate a global model from scratch each time a new domain with potentially new intents and slots is added. -> 2017.4
papers:
2017.1 adversarial training
[Adversarial Adaptation of Synthetic or Stale Data. Young-Bum Kim. 14]
2017.4 model(k + 1) = weighted_combination[model(1),...,model(k)]
[Domain Attention with an Ensemble of Experts. Young-Bum Kim. 17]
2. NLG:
challenges:
(a) integrate LM + Affect. -> 2017.2
(b) refering expression misunderstand -> 2017.5
(c) neural encoder-decoder models in open-domain: generate dull and generic responses. -> 2017.8
(d) multi-turn: lose relationships among utterances or important contextual information. -> 2017.11
(e) automatically evaluating the quality of dialogue responses for unstructured domains: biased and correlate very poorly with human judgements of response quality. -> 2017.12
(f) deep latent variable models used in open-domain: highly randomized, leading to uncontrollable generated responses. -> 2017.14
(g) does not employ knowledge to guide the generation -> tends to generate short, general, and meaningless responses. -> 2018.L1
(h) encoder-decoder dialog model is limited because it cannot output interpretable actions as in traditional systems, which hinders(阻碍) humans from understanding its generation process. -> 2018.L6
(i) translate natural language questions ->structured queries: further improvement hard. -> 2018.L8
papers:
2017.2 language model + affect info
[Affect-LM: A Neural Language Model for Customizable Affective Text Generation. 27]
2017.5 refering expression misunderstand correction - alg: contrastive focus
[Generating Contrastive Referring Expressions. 0]
2017.8 open-domain - Framework - conditional variaional autoencoders
pre: word-level decoder
cur: discourse-level encoder
[Learning Discourse-level Diversity for Neural Dialog Models using Conditional Variational Autoencoders. CMU. 69]
2017.11 muli-turn response selection - sequential matching network (SMN)
pre: concatenates utterances in context
matches a response with a highly abstract context vector
=> lose relationships among utterances or important contextual information
current: matches a response with each utterance on multiple levels of granularity
distills important matching information -> vector -> conv + pooling
accumulate vector -> RNN (models relationships among utterances)
final matching score (calcu with hid of rnn)
[Sequential Matching Network: A New Architecture for Multi-turn Response Selection in Retrieval-based Chatbots. 北航. 南开.微软. 48]
2017.12 auto eval Metric - ADEM
[Towards an Automatic Turing Test: Learning to Evaluate Dialogue Responses. 47]
2017.14 Framework - generation based on specific attributes(manually + auto detected) - both speakers diag states modeled -> personal features
[A Conditional Variational Framework for Dialog Generation. 20]
2017.16 Open-domain - Engine - generation (info retrieval + Seq2Seq) - AliMe chat
[AliMe Chat: A Sequence to Sequence and Rerank based Chatbot Engine. 27]
2018.L1 knowledge guide generation - neural knowledge diffusion (NKD) model - both fact + chi-chats
match the relevant facts for the input utterance + diffuse them to similar entities
[Knowledge Diffusion for Neural Dialogue Generation. 3]
2018.L6 encoder-decoder model - interprete- unsup discrete sent representation learning
DI-VAE + DI-VST - discover interpretable semantics via either auto encoding or context predicting
[Unsupervised Discrete Sentence Representation Learning for Interpretable Neural Dialog Generation. 8]
2018.L8 Framework - DialSQL + human intelligence
identify potential error of SQL -> ask for validation -> feedback to revise query
[ DialSQL: Dialogue Based Structured Query Generation. 4]
3. Task + Non-task hybrid
2017.3 whether to have a chat - dataset
[Chat Detection in Intelligent Assistant: Combining Task-oriented and Non-task-oriented Spoken Dialogue Systems. McGill University. Montreal. 7]
4. E2E
challenges:
(a) data-intensive -> 2017.6
(b) task - interact with KB -> pre: issuing a symbolic query to the KB to retrieve entries based on their attributes. -> 2017.13
disadvantages:
(1) such symbolic operations break the differentiability(可辨性) of the system
(2) prevent end-to-end training of neural dialogue agents
(c) only consider user semantic inputs and under-utilize other user info. -> 2018.L4
(b) incorporating knowledge bases. -> 2018.L7
papers:
2017.6 Framework - HCNs : RNN + knowledge(software/sys action templates) - reduce train data - opt (sup + RL) - bAbI dialog dataset - 2 commercial diag sys
[Hybrid Code Networks: practical and efficient end-to-end dialog control with supervised and reinforcement learning. Microsoft Research. 87]
2017.13 KB-InfoBot - E2E - task -multi-turn - interact with KB - present a agent
replacing symbolic queries -> induced "soft" posterior distribution over the KB
integrate soft retrival process + RL
[Towards End-to-End Reinforcement Learning of Dialogue Agents for Information Access. CMU. MS. 国立台北. 82]
2018.L4 multimodel info (sup + RL) - user adaptive - reduce diag length + improve success rate
[Sentiment Adaptive End-to-End Dialog Systems. 2]
2018.L7 Mem2Seq - first neural generative model: combines [ multi-hop attention over memories + idea of pointer network]
[Mem2Seq: Effectively Incorporating Knowledge Bases into End-to-End Task-Oriented Dialog Systems. UST.8]
5. NLU
challenges:
(a) no systematic comparison to analyze how to use context effectively. -> 2017.15
papers:
2017.7 identity discussion points + discourse relations
[Joint Modeling of Content and Discourse Relations in Dialogues. 7]
2017.15 context utiliztion eval -empirical study and compare models - variant: weights context vectors by context-query relevance
[How to Make Contexts More Useful? An Empirical Study to Context-Aware Neural Conversation Models. 18]
6. Dialogue state tracking
challenges:
(a) have difficulty scaling to larger, more complex dialogue domains. -> 2017.10
(1) Spoken Language Understanding models that require large amounts of annotated training data
(2) hand-crafted lexicons for capturing some of the linguistic variation in users' language.
(b) handling unknown slot values -> Pre: assume predefined candidate lists and thus are not designed to output unknown values. especially in E2E, SLU is absent. -> 2018.L10
papers:
2017.10 Framework - Neural Belief Tracking (NBT) - representation learning (compose pre-trained word vector -> utterances and context)
[Neural Belief Tracker: Data-Driven Dialogue State Tracking. 63]
2018.L9 Global-Locally Self-Attentive Dialogue State Tracker (GLAD)
global modules: shares parameters between estimators for different types (called slots) of dialogue states
local modules: learn slot-specific features
[Global-Locally Self-Attentive Encoder for Dialogue State Tracking. 0]
2018.L10 E2E + pointer nerwork (PtrNet)
[An End-to-end Approach for Handling Unknown Slot Values in Dialogue State Tracking. 2]
7. Framework
challenges:
(a) pipeline: introduces architectural complexity and fragility. -> 2018.L2
papers:
2018.L2 Seq2Seq + opt (sup / RL) - Task
design text spans named belief spans -> track dialogue believes -> allow task-oriented sys be modeled in Seq2Seq
Two Stage CopyNet instantiation -> reduce para, train time + better than pipeline on large dataset + OOV
[Sequicity: Simplifying Task-oriented Dialogue Systems with Single Sequence-to-Sequence Architectures. 新加坡国立. 复旦. 京东. 9]
8. RL
challenges:
(a) Training a task-completion dialogue agent via reinforcement learning (RL) is costly: requires many interactions with real users.
(b) use a user simulator: lacks the language complexity + biases
papers:
2018.L3 RL - policy learning - Deep Dyna-Q
first deep RL framework that integrates planning for task-completion dialogue policy learning
world model update with real user experience + agent opt using real and simulated experience
[Deep Dyna-Q: Integrating Planning for Task-Completion Dialogue Policy Learning. 3]
8. Chi-chat
challenges:
(a) lack specificity
(b) do not display a consistent personality. -> 2018.L5
papers:
2018.L5 add profile info[i. given + ii.partner] train to engage ii with personal topics -> used to predict profile
[Personalizing Dialogue Agents: I have a dog, do you have pets too? 31.]
9. Others:
challenges:
(a) open-ended dialogue state. -> 2017.9
papers:
2017.9 Symmetric Collaborative Dialogue - two agents to achieve a common goal
[Learning Symmetric Collaborative Dialogue Agents with Dynamic Knowledge Graph Embeddings. 21]
[2017 - 2018 ACL] 对话系统论文研究点整理的更多相关文章
- R语言重要数据集分析研究——需要整理分析阐明理念
1.R语言重要数据集分析研究需要整理分析阐明理念? 上一节讲了R语言作图,本节来讲讲当你拿到一个数据集的时候如何下手分析,数据分析的第一步,探索性数据分析. 统计量,即统计学里面关注的数据集的几个指标 ...
- MyEclips 2017/2018 (mac 版)安装与破解
MyEclips 2017/2018 (mac 版)安装与破解 现在在学J2EE,然后使用的工具就是 MyEclipse,现在就抛弃 Eclipse 了,我就不多说它俩的区别了,但是 MyEclips ...
- MyEclipse 2017/2018 安装与破解 图文教程
SSM 框架-02-MyEclipse 2017/2018 安装与破解 现在在学J2EE,然后使用的工具就是 MyEclipse,现在就抛弃 Eclipse 了,我就不多说它俩的区别了,但是 MyEc ...
- Hadoop是原Yahoo的Doug Cutting根据Google发布的学术论文研究而来
Hadoop是原Yahoo的Doug Cutting根据Google发布的学术论文研究而来.Doug Cutting给这个Project起了个名字,就叫Hadoop. Doug Cutting在Clo ...
- </2017><2018>
>>> Blog 随笔原始文档及源代码 -> github: https://github.com/StackLike/Python_Note >>> 统计信 ...
- 转:2018最全Redis面试题整理
Java面试----2018最全Redis面试题整理 1.什么是Redis? 答:Redis全称为:Remote Dictionary Server(远程数据服务),是一个基于内存的高性能key-va ...
- [2017 ACL] 对话系统
Long Papers [Domain adaptation ] 1. Adversarial Adaptation of Synthetic or Stale Data ( Cited by 14 ...
- [2018 ACL Short and System] 对话系统
Short Paper(s) 1. Task-oriented Dialogue System for Automatic Diagnosis. (Cited by 0) Zhongyu Wei, ...
- [2018 ACL Long] 对话系统
[NLG - E2E - knowledge guide generation] 1. Knowledge Diffusion for Neural Dialogue Generation ( Ci ...
随机推荐
- Java Bean与Map之间相互转化的实现
目录树 概述 Apache BeanUtils将Bean转Map Apache BeanUtils将Map转Bean 理解BeanUtils将Bean转Map的实现之手写Bean转Map 概述 Apa ...
- Java关于NIO类的详解
一.IO与NIO的区别: 前提我们先说一说java IO: Java中使用IO(输入输出)来读取和写入,读写设备上的数据.硬盘文件.内存.键盘......,根据数据的走向可分为输入流和输出流,这个走向 ...
- forEach for...in for...of
forEach orEach 方法为数组中含有有效值的每一项执行一次 callback 函数,那些已删除(使用 delete 方法等情况)或者从未赋值的项将被跳过(不包括那些值为 undefined ...
- Asp.Net Core使用Log4Net优化日志【项目开源】
我在前一篇文章中介绍了一种使用Log4Net的方法,但是那种方法打出来的日志不是很直观 然后我前不久阅读了一篇非常不错的博客:https://www.cnblogs.com/guolianyu/p/9 ...
- Hadoop2学习路程-HDFS
什么是Hadoop HDFS? Hadoop 分布式文件系统是世界上最可靠的文件系统.HDFS可以再大量硬件组成的集群中存储大文件. 它的设计原则是趋向于存储少量的大文件,而不是存储大量的小文件. 即 ...
- 我的名字叫hadoop
第一回 新入环境 我的名字是hadoop,我一出生我的爸爸雅虎就给我取了这样一个名字:hadoop,我也不知道为什么叫这个名字,刚出生没多久,雅虎爸爸就把我领进一个黑暗的屋子里面,屋里堆满了黑色的 ...
- ruby中的return方法及class实例方法的initialize方法
return是函数的返回值 class Mtring def initialize(str) @name = str end def aa ary = @name.split(/:/) return ...
- python代理爬取存入csv文件
爬取高匿代理 from urllib import request import re import time f = open('西1.csv','w',encoding='GBK') header ...
- 理解golang中的channel
channel是goroutine之间的通信机制.可以类比线程间的通信,线程间的通信有多种方式,比如线程上下文.共享内存.IPC通信.socket实现不同机器间的通信. channel用起来很简单,绑 ...
- POJ-2155:Matrix(二维树状数祖)
Matrix Time Limit: 3000MS Memory Limit: 65536K Total Submissions: 31892 Accepted: 11594 Descript ...