VIP内容

题目: Integrating Deep Learning with Logic Fusion for Information Extraction

摘要:

信息抽取(Information extraction, IE)旨在从输入文本中产生结构化的信息,例如命名实体识别和关系抽取。通过特征工程或深度学习为IE提出了各种尝试。然而,他们中的大多数人并没有将任务本身所固有的复杂关系联系起来,而这一点已被证明是特别重要的。例如,两个实体之间的关系高度依赖于它们的实体类型。这些依赖关系可以看作是复杂的约束,可以有效地表示为逻辑规则。为了将这种逻辑推理能力与深度神经网络的学习能力相结合,我们提出将一阶逻辑形式的逻辑知识集成到深度学习系统中,以端到端方式联合训练。该集成框架通过逻辑规则对神经输出进行知识正则化增强,同时根据训练数据的特点更新逻辑规则的权值。我们证明了该模型在多个IE任务上的有效性和泛化性。

作者:

Sinno Jialin Pan是南洋理工大学计算机科学与工程学院院长兼副教授,研究方向是迁移学习、数据挖掘、人工智能、机器学习。

成为VIP会员查看完整内容
0
47

热门内容

The quest of `can machines think' and `can machines do what human do' are quests that drive the development of artificial intelligence. Although recent artificial intelligence succeeds in many data intensive applications, it still lacks the ability of learning from limited exemplars and fast generalizing to new tasks. To tackle this problem, one has to turn to machine learning, which supports the scientific study of artificial intelligence. Particularly, a machine learning problem called Few-Shot Learning (FSL) targets at this case. It can rapidly generalize to new tasks of limited supervised experience by turning to prior knowledge, which mimics human's ability to acquire knowledge from few examples through generalization and analogy. It has been seen as a test-bed for real artificial intelligence, a way to reduce laborious data gathering and computationally costly training, and antidote for rare cases learning. With extensive works on FSL emerging, we give a comprehensive survey for it. We first give the formal definition for FSL. Then we point out the core issues of FSL, which turns the problem from "how to solve FSL" to "how to deal with the core issues". Accordingly, existing works from the birth of FSL to the most recent published ones are categorized in a unified taxonomy, with thorough discussion of the pros and cons for different categories. Finally, we envision possible future directions for FSL in terms of problem setup, techniques, applications and theory, hoping to provide insights to both beginners and experienced researchers.

0
311
下载
预览

最新内容

Few-shot dataset generalization is a challenging variant of the well-studied few-shot classification problem where a diverse training set of several datasets is given, for the purpose of training an adaptable model that can then learn classes from new datasets using only a few examples. To this end, we propose to utilize the diverse training set to construct a universal template: a partial model that can define a wide array of dataset-specialized models, by plugging in appropriate components. For each new few-shot classification problem, our approach therefore only requires inferring a small number of parameters to insert into the universal template. We design a separate network that produces an initialization of those parameters for each given task, and we then fine-tune its proposed initialization via a few steps of gradient descent. Our approach is more parameter-efficient, scalable and adaptable compared to previous methods, and achieves the state-of-the-art on the challenging Meta-Dataset benchmark.

0
0
下载
预览

最新论文

Few-shot dataset generalization is a challenging variant of the well-studied few-shot classification problem where a diverse training set of several datasets is given, for the purpose of training an adaptable model that can then learn classes from new datasets using only a few examples. To this end, we propose to utilize the diverse training set to construct a universal template: a partial model that can define a wide array of dataset-specialized models, by plugging in appropriate components. For each new few-shot classification problem, our approach therefore only requires inferring a small number of parameters to insert into the universal template. We design a separate network that produces an initialization of those parameters for each given task, and we then fine-tune its proposed initialization via a few steps of gradient descent. Our approach is more parameter-efficient, scalable and adaptable compared to previous methods, and achieves the state-of-the-art on the challenging Meta-Dataset benchmark.

0
0
下载
预览
子主题
Top