ing extra module or placing constraint on model
output, we propose novel schema-awared denois-
ing objectives trained along with the original S2S
task. These denoising objectives deal with the in-
trinsic attribute of logical form and therefore facili-
tate schema linking required for text-to-SQL task.
The inductive schema-awared noises can be cate-
gorized into two types: erosion and shuffle. Ero-
sion acts on schema input by randomly permute,
drop and add columns into the current schema set.
The related schema entity in target SQL query will
be jointly modified according to the erosion re-
sult. Shuffle is applied via randomly re-ordering
the mentioned entity and values in NL or SQL with
respect to the schema columns. During training pro-
cedure, shuffle is performed during monolingual
self-supervision that trains model to recover orig-
inal text given the noised one. Erosion is applied
to S2S task that trains model to generate corrupted
SQL sequence, given NL and eroded schema as
input. These proposed denoising objectives are
combined along with the origin S2S task to train a
SeaD model. In addition, to deal with the limitation
of execution-guided (EG) decoding, we propose a
clause-sensitive EG strategy that decide beam size
with respect to the clause token that is predicted.
We compare the proposed method with other
top-performing models on WikiSQL benchmark.
The results show that the performance of our model
surpasses previous work and establish new state-of-
the-art for WikiSQL. It demonstrate the effective-
ness of the schema-aware denoising approach and
also shad lights on the importance of task-oriented
denoising objective.
2 Related Work
Semantic Parsing
The problem of mapping natu-
ral language to meaningful executable programs
has been widely studied in natural language pro-
cessing research. Logic forms (Zettlemoyer and
Collins, 2012; Artzi and Zettlemoyer, 2011, 2013;
Cai and Yates, 2013; Reddy et al., 2014; Liang
et al., 2013; Quirk et al., 2015; Chen et al., 2016)
can be considered as a special instance to the more
generic semantic parsing problem. As a sub-task
of semantic parsing, the text-to-SQL problem has
been studied for decades. (Warren and Pereira,
1982; Popescu et al., 2003; Li et al., 2006; Giordani
and Moschitti, 2012; Bodik). Slot-filling model
(Hwang et al., 2019; He et al., 2019a; Lyu et al.,
2020) translates the clauses of SQL into subtasks,
(Ma et al., 2020) treat this task as a two-stage se-
quence labeling model. However, the convergence
rate between subtasks is inconsistent or the inter-
action between multiple subtasks may lead to the
model may not converge well. Like lots of previ-
ous work (Dong and Lapata, 2016; Lin et al., 2018;
Zhong et al., 2017; Suhr et al., 2020; Raffel et al.,
2019), we treat text-to-SQL as a translation prob-
lem, and taking both the natural language question
and the DB as input.
Hybrid Pointer Networks
Proposed by (Vinyals
et al., 2015), copying mechanism (CM) uses atten-
tion as a pointer to copy several discrete tokens
from input sequence as the output and have been
successfully used in machine reading comprehen-
sion (Wang and Jiang, 2016; Trischler et al., 2016;
Kadlec et al., 2016; Xiong et al., 2016), interactive
conversation (Gu et al., 2016; Yu and Joty, 2020;
He et al., 2019b), geometric problems (Vinyals
et al., 2015) and program generation (Zhong et al.,
2017; Xu et al., 2017; Dong and Lapata, 2016; Yu
et al., 2018; McCann et al., 2018; Hwang et al.,
2019). In text-to-SQL, CM can not only facilitate
the condition value extraction from source input,
but also help to protect the privacy of the database.
In this paper, We use a Hybrid Pointer Generator
Network which is similar to (Jia and Liang, 2016;
Rongali et al., 2020) to generate next step token.
Denoising Self-training
Language model pretrain-
ing (Devlin et al., 2018; Yang et al., 2019; Liu et al.,
2019; Lan et al., 2019) has been shown to improve
the downstream performance on many NLP tasks
and brought significant gains. (Radford et al., 2018;
Peters et al., 2018; Song et al., 2019) are benefi-
cial to S2S task, while they are problematic for
some tasks. While (Lewis et al., 2019) is a denois-
ing S2S pre-training model, which is effective for
both generative and discriminative tasks, reduces
the mismatch between pre-training and generation
tasks. Inspired by this, we propose a denosing self-
training architecture in training to learn mapping
corrupted documents to the original.
3 Methodology
Given natural language question
Q
and a schema
S
,
our goal is to obtain the corresponding SQL query
Y
. Here the natural question
Q = {q
1
, ..., q
|Q|
}
denotes a word sequence, the schema
S =
{c
1
, ..., c
|S|
}
is composed of a set of columns,
where each column
c
i
= {c
1
, ..., c
|c
i
|
}
is a se-
quence of words.
Y = y
1
, ..., y
|Y |
denotes the
评论