Searchable Hidden Intermediates for End-to-End Models of Decomposable Sequence Tasks

Page created by Rene Sandoval
 
CONTINUE READING
Searchable Hidden Intermediates for End-to-End Models of
                                                                 Decomposable Sequence Tasks

                                            Siddharth Dalmia Brian Yan Vikas Raunak Florian Metze Shinji Watanabe
                                                      Language Technologies Institute, Carnegie Mellon University, USA
                                                                   {sdalmia,byan}@cs.cmu.edu

                                                               Abstract                                  Similarly, many sequence-to-sequence tasks that
                                                                                                      convert one sequence into another (Sutskever et al.,
                                             End-to-end approaches for sequence tasks are             2014) can be decomposed to simpler sequence sub-
                                             becoming increasingly popular. Yet for com-
                                                                                                      tasks in order to reduce the overall complexity.
arXiv:2105.00573v1 [cs.CL] 2 May 2021

                                             plex sequence tasks, like speech translation,
                                             systems that cascade several models trained
                                                                                                      For example, speech translation systems, which
                                             on sub-tasks have shown to be superior, sug-             seek to process speech in one language and output
                                             gesting that the compositionality of cascaded            text in another language, can be naturally decom-
                                             systems simplifies learning and enables so-              posed into the transcription of source language au-
                                             phisticated search capabilities. In this work,           dio through automatic speech recognition (ASR)
                                             we present an end-to-end framework that ex-              and translation into the target language through ma-
                                             ploits compositionality to learn searchable hid-         chine translation (MT). Such cascaded approaches
                                             den representations at intermediate stages of a
                                                                                                      have been widely used to build practical systems
                                             sequence model using decomposed sub-tasks.
                                             These hidden intermediates can be improved               for a variety of sequence tasks like hybrid ASR
                                             using beam search to enhance the overall per-            (Hinton et al., 2012), phrase-based MT (Koehn
                                             formance and can also incorporate external               et al., 2007), and cascaded ASR-MT systems for
                                             models at intermediate stages of the network to          speech translation (ST) (Pham et al., 2019).
                                             re-score or adapt towards out-of-domain data.               End-to-end sequence models like encoder-
                                             One instance of the proposed framework is
                                                                                                      decoder models (Bahdanau et al., 2015; Vaswani
                                             a Multi-Decoder model for speech translation
                                             that extracts the searchable hidden intermedi-           et al., 2017), are attractive in part due to their sim-
                                             ates from a speech recognition sub-task. The             plistic design and the reduced need for hand-crafted
                                             model demonstrates the aforementioned bene-              features. However, studies have shown mixed re-
                                             fits and outperforms the previous state-of-the-          sults compared to cascaded models particularly for
                                             art by around +6 and +3 BLEU on the two test             complex sequence tasks like speech translation (In-
                                             sets of Fisher-CallHome and by around +3 and             aguma et al., 2020) and spoken language under-
                                             +4 BLEU on the English-German and English-
                                                                                                      standing (Coucke et al., 2018). Although direct
                                             French test sets of MuST-C.1
                                                                                                      target sequence prediction avoids the issue of er-
                                        1    Introduction                                             ror propagation from one system to another in cas-
                                                                                                      caded approaches (Tzoukermann and Miller, 2018),
                                        The principle of compositionality loosely states that         there are many attractive properties of cascaded sys-
                                        a complex whole is composed of its parts and the              tems, missing in end-to-end approaches, that are
                                        rules by which those parts are combined (Lake and             useful in complex sequence tasks.
                                        Baroni, 2018). This principle is present in engineer-            In particular, we are interested in (1) the strong
                                        ing, where task decomposition of a complex system             search capabilities of the cascaded systems that
                                        is required to assess and optimize task allocations           compose the final task output from individual sys-
                                        (Levis et al., 1994), and in natural language, where          tem predictions (Mohri et al., 2002; Kumar et al.,
                                        paragraph coherence and discourse analysis rely               2006; Beck et al., 2019), (2) the ability to incor-
                                        on decomposition into sentences (Johnson, 1992;               porate external models to re-score each individual
                                        Kuo, 1995) and sentence level semantics relies on             system (Och and Ney, 2002; Huang and Chiang,
                                        decomposition into lexical units (Liu et al., 2020b).         2007), (3) the ability to easily adapt individual com-
                                           1
                                             All code and models are released as part of the ESPnet   ponents towards out-of-domain data (Koehn and
                                        toolkit: https://github.com/espnet/espnet.                    Schroeder, 2007; Peddinti et al., 2015), and finally
(4) the ability to monitor performance of the indi-        sequence tasks to learn next word prediction, which
vidual systems towards the decomposed sub-task             outputs a distribution over the next target token
(Tillmann and Ney, 2003; Meyer et al., 2016).              yl given the previous tokens y1:l91 and the input
   In this paper, we seek to incorporate these proper-     sequence x = (x1 , xt , . . . , xT ), where T is the
ties of cascaded systems into end-to-end sequence          input sequence length. In the next sub-section we
models. We first propose a generic framework               detail the training and inference of these models.
to learn searchable hidden intermediates using an
auto-regressive encoder-decoder model for any de-          2.2    Auto-regressive Encoder-Decoder Models
composable sequence task (§3). We then apply               Training: In an auto-regressive encoder-decoder
this approach to speech translation, where the in-         model, the E NCODER maps the input sequence x
termediate stage is the output of ASR, by passing          to a sequence of continuous hidden representations
continuous hidden representations of discrete tran-        hE = (hE        E            E             E
                                                                      1 , ht , . . . , hT ), where ht ∈ R . The
                                                                                                                 d

                                                           D ECODER then auto-regressively maps h and the     E
script sequences from the ASR sub-net decoder to
the MT sub-net encoder. By doing so, we gain               preceding ground-truth output tokens, ŷ1:l91 , to hD      l ,
the ability to use beam search with optional ex-           where hD l  ∈  R d . The sequence of decoder hidden

ternal model re-scoring on the hidden intermedi-           representations form hD = (hD            D            D
                                                                                               1 , hl , . . . , hL ) and
ates, while maintaining end-to-end differentiability.      the likelihood of each output token yl is given by
Next, we suggest mitigation strategies for the error       S OFTMAX O UT, which denotes an affine projection
propagation issues inherited from decomposition.           of hDl to V followed by a softmax function.
   We show the efficacy of searchable intermediate
representations in our proposed model, called the                              hE = E NCODER(x)
Multi-Decoder, on speech translation with a 5.4                                ĥD              E
                                                                                 l = D ECODER (h , ŷ1:l91 ) (1)
and 2.8 BLEU score improvement over the previ-
                                                             P (yl | ŷ1:l91 , hE ) = S OFTMAX O UT(ĥD
                                                                                                      l )            (2)
ous state-of-the-arts for Fisher and CallHome test
sets respectively (§6). We extend these improve-           During training, the D ECODER performs token clas-
ments by an average of 0.5 BLEU score through              sification for next word prediction by considering
the aforementioned benefit of re-scoring the inter-        only the ground truth sequences for previous to-
mediate search with external models trained on the         kens ŷ. We refer to this ĥD as oracle decoder
same dataset. We also show a method for monitor-           representations, which will be discussed later.
ing sub-net performance using oracle intermediates         Inference: During inference, we can maximize the
that are void of search errors (§6.1). Finally, we         likelihood of the entire sequence from the output
show how these models can adapt to out-of-domain           space S by composing the conditional probabilities
speech translation datasets, how our approach can          of each step for the L tokens in the sequence.
be generalized to other sequence tasks like speech
recognition, and how the benefits of decomposition                          hD              E
                                                                             l = D ECODER (h , y1:l91 )              (3)
persist even for larger corpora like MuST-C (§6.2).         P (yl | x, y1:l91 ) = S OFTMAX O UT(hD
                                                                                                 l )
2     Background and Motivation                                                            L
                                                                                           Y
                                                                           ỹ = argmax           P (yi | x, y1:i91 ) (4)
2.1    Compositionality in Sequences Models                                        y∈S     i=1
The probabilistic space of a sequence is combinato-
                                                           This is an intractable search problem and it can be
rial in nature, such that a sentence of L words from
                                                           approximated by either greedily choosing argmax
a fixed vocabulary V would have an output space S
                                                           at each step or using a search algorithm like beam
of size |V|L . In order to deal with this combinato-
                                                           search to approximate ỹ. Beam search (Reddy,
rial output space, an output sentence is decomposed
                                                           1988) generates candidates at each step and prunes
into labeled target tokens, y = (y1 , y2 , . . . , yL ),
                                                           the search space to a tractable beam size of B most
where yl ∈ V.
                                                           likely sequences. As B → ∞, the beam search
                        L
                        Y                                  result would be equivalent to equation 4.
          P (y | x) =         P (yi | x, y1:i91 )
                        i=1                                  G REEDY S EARCH := argmax P (yl | x, y1:l91 )
                                                                                            yl
An auto-regressive encoder-decoder model uses the
above probabilistic decomposition in sequence-to-                B EAM S EARCH := B EAM(P (yl | x, y1:l91 ))
(a) Multi-Decoder ST Model                                   (b) Multi-Sequence Attention

Figure 1: The left side present the schematics and the information flow of our proposed framework applied to ST, in
a model we call the Multi-Decoder. Our model decomposes ST into ASR and MT sub-nets, each of which consist
of an encoder and decoder. The right side displays a Multi-Sequence Attention variant of the DECODER ST that is
conditioned on both speech information via the ENCODER ASR and transcription information via the ENCODER ST .

In approximate search for auto-regressive models,                S UBA→B N ET:
like beam search, the D ECODER receives alternate
candidates of previous tokens to find candidates                                   hE = E NCODERA (A)
with a higher likelihood as an overall sequence.                                 ĥD
                                                                                   l
                                                                                     B                     B
                                                                                       = D ECODERB (hE , ŷ1:l91 )
This also allows for the use of external models like
                                                                  P (ylB | ŷ1:l91
                                                                             B
                                                                                   , hE ) = S OFTMAX O UT(ĥDB
                                                                                                            l ) (5)
Language Models (LM) or Connectionist Temporal
Classification Models (CTC) for re-scoring candi-                S UBB→C N ET:
dates (Hori et al., 2017).
                                                                        P (C | ĥD B                    DB
                                                                                 l ) = S UB B→C N ET (ĥl )           (6)

3    Proposed Framework                                          Note that the final prediction, given by equation
                                                                 6, does not need to be a sequence and can be a
In this section, we present a general framework to               categorical class like in spoken language under-
exploit natural decompositions in sequence tasks                 standing tasks. Next we will show how the hidden
which seek to predict some output C from an input                intermediates become searchable during inference.
sequence A. If there is an intermediate sequence B
                                                                 3.1   Searchable Hidden Intermediates
for which A → B sequence transduction followed
by B → C prediction achieves the original task,                  As stated in section §2.2, approximate search algo-
then the original A → C task is decomposable.                    rithms maximize the likelihood, P (y | x), of the
                                                                 entire sequence by considering different candidates
   In other words, if we can learn P (B | A) then
                                                                 yl at each step. Candidate-based search, particu-
we can learn the overall task of P (C | A) through
                                                                 larly in auto-regressive encoder-decoder models,
maxB (P (C | A, B)P (B | A)), approximated
                                                                 also affects the decoder hidden representation, hD ,
using Viterbi search. We define a first encoder-
                                                                 as these are directly dependent on the previous can-
decoder S UBA→B N ET to map an input sequence
                                                                 didate (refer to equations 1 and 3). This implies that
A to a sequence of decoder hidden states, hDB .
                                                                 by searching for better approximations of the pre-
Then we define a subsequent S UBB→C N ET to map
                                                                 vious predicted tokens, yl91 = (yBEAM )l91 , we also
hDB to the final probabilistic output space of C.
                                                                 improve the decoder hidden representations for the
Therefore, we call hDB hidden intermediates. The
                                                                 next token, hD          D
                                                                                 l = (hBEAM )l . As yBEAM → ŷ, the
following equations shows the two sub-networks of
                                                                 decoder hidden representations tend to the oracle
our framework, S UBA→B N ET and S UBB→C N ET,
                                                                 decoder representations that have only errors from
which can be trained end-to-end while also exploit-
                                                                 next word prediction, hD               D
                                                                                            BEAM → ĥ . A perfect
ing compositionality in sequence tasks. 2
                                                                 search is analogous to choosing the ground truth ŷ
                                                                 at each step, which would yield ĥD .
    2
      Note that this framework does not use locally-normalized      We apply this beam search of hidden interme-
softmax distributions but rather the hidden representations,
thereby avoiding label bias issues when combining multiple       diates, thereby approximating ĥDB with hD         B
                                                                                                                 BEAM .
sub-systems (Bottou et al., 1997; Wiseman and Rush, 2016).       This process is illustrated in algorithm 1, which
shows beam search for hD  B
                         BEAM that are subsequently             quence of speech x and uses a sequence of text
passed to the S UBB→C N ET.3 In line 7, we show                 transcriptions y ASR as an intermediate. In this case,
how an external model like an LM or a CTC model                 the S UBA→B N ET in equation 5 is specified as the
can be used to generate an alternate sequence like-             ASR sub-net and the S UBB→C N ET in equation 6 is
lihood, PEXT (ylB ), which can be combined with                 specified as the MT sub-net. Since the MT sub-net
the S UBA→B N ET likelihood, PB (ylB | x) , with a              is also a sequence prediction task, both sub-nets are
tunable parameter λ.                                            encoder-decoder models in our architecture (Bah-
                                                                danau et al., 2015; Vaswani et al., 2017). In Figure
Algorithm 1 Beam Search for Hidden Interme-                     1 we illustrate the schematics of our transformer
diates: We perform beam search to approximate                   based Multi-Decoder ST model which can also be
the most likely sequence for the sub-task A →                   summarized as follows:
B, yB BEAM , while collecting the corresponding
D ECODERB hidden representations, hD       B                              hEASR = E NCODER ASR (x)                         (7)
                                          BEAM . The
         DB
output hBEAM , is passed to the final sub-network to                      ĥD
                                                                            l
                                                                              ASR
                                                                                  = D ECODER ASR (hEASR , ŷ1:l91
                                                                                                            ASR
                                                                                                                  )        (8)
predict final output C and yB BEAM is used for moni-                      hEST = E NCODER ST (ĥDASR )                     (9)
toring performance on predicting B.
                                                                          ĥD
                                                                            l
                                                                              ST
                                                                                   = D ECODER ST (h   EST      ST
                                                                                                            , ŷ1:l91 )   (10)
 1:    Initialize: BEAM ← {sos}; k ← beam size;
 2:    hEA ← E NCODERA (x)                                      As we can see from Equations 9 and 10, the MT
 3:    for l=1 to maxSTEPS do                                   sub-network attends only to the decoder representa-
 4:      for yl91B ∈ BEAM do                                    tions, ĥDASR , of the ASR sub-network, which could
 5:          hD   B                       B )
                    ← D ECODERB (hEA , yl91                     lead to the error propagation issues from the ASR
               l
 6:                 B      B
             for yl ∈ yl91 + {V} do                             sub-network to the MT sub-network similar to the
 7:              sl ← PA→B (ylB | x)19λ PEXT (ylB )λ            cascade systems, as mentioned in §1. To allevi-
                 H ← (sl , ylB , hDB                            ate this problem, we modify equation 10 such that
 8:                               l )
 9:          end for                                            D ECODER ST attends to both hEST and hEASR :
                                                                       SA
                                                                      DST                    EST
10:      end for                                                    ĥl     = D ECODER SA
                                                                                       ST (h     , hEASR , ŷ1:l91
                                                                                                             ST
                                                                                                                   ) (11)
11:      BEAM ← argk max(H)
                                                                We use the multi-sequence cross-attention dis-
12:    end for
                        DB                                      cussed by Helcl et al. (2018), shown on the right
13:    (sB , yB
              BEAM , h BEAM ) ← argmax( BEAM )                  side of Figure 1, to condition the final outputs gen-
14:    Return yB    BEAM → S UB A→B N ET Monitoring             erated by ĥD ST
                                                                                 on both speech and transcript in-
15:    Return hD      B
                    BEAM → Final S UB B→C N ET
                                                                            l
                                                                formation in an attempt to allow our network to
                                                                recover from intermediate mistakes during infer-
We can monitor the performance of the                           ence. We call this model the Multi-Decoder w/
S UBA→B N ET by comparing the decoded in-                       Speech-Attention.
termediate sequence yB   BEAM to the ground truth
  B
ŷ . We can also monitor the S UBB→C N ET                       4     Baseline Encoder-Decoder Model
performance by using the aforementioned oracle
                                                                For our baseline model, we use an end-to-end
representations of the intermediates, ĥDB , which
                                                                encoder-decoder (Enc-Dec) ST model with ASR
can be obtained by feeding the ground truth ŷB
                                                                joint training (Inaguma et al., 2020) as an aux-
to D ECODERB . By passing ĥDB to S UBB→C N ET,
                                                                iliarly loss to the speech encoder. In other
we can observe its performance in a vacuum, i.e.
                                                                words, the model consumes speech input using
void of search errors in the hidden intermediates.
                                                                the E NCODER ASR , to produce hEASR , which is
3.2     Multi-Decoder Model                                     used for cross-attention by D ECODER ASR and the
                                                                D ECODER ST . Using the decomposed ASR task as
In order to show the applicability of our end-to-end
                                                                an auxiliary loss also helps the baseline Enc-Dec
framework we propose our Multi-Decoder model
                                                                model and provide strong baseline performance, as
for speech translation. This model predicts a se-
                                                                we will see in Section 6.
quence of text translations y ST from an input se-
   3
     The algorithm shown only considers a single top approxi-   5     Data and Experimental Setup
mation of the search; however, with added time-complexity,
the final task prediction improves with the n-best hD  B
                                                     BEAM for
                                                                Data: We demonstrate the efficacy of our pro-
selecting the best resultant C.                                 posed approach on ST in the Fisher-CallHome cor-
pus (Post et al., 2013) which contains 170 hours of     has an E NCODER ST consisting of 2 transformer
Spanish conversational telephone speech, transcrip-     encoder blocks with the same configuration as
tions, and English translations. All punctuations       E NCODER ASR , giving a total of 40.5M trainable
except apostrophes were removed and results are         parameters. The training configuration is also the
reported in terms of detokenized case-insensitive       same as for the baseline. For the Multi-Decoder w/
BLEU (Papineni et al., 2002; Post, 2018). We com-       Speech-Attention model (42.1M trainable parame-
pute BLEU using the 4 references in Fisher (dev,        ters), we increase the attention dropout of the ST
dev2, and test) and the single reference in Call-       decoder to 0.4 and dropout on all other components
Home (dev and test) (Post et al., 2013; Kumar et al.,   of the ST decoder to 0.2 while keeping dropout on
2014; Weiss et al., 2017). We use a joint source and    the remaining components at 0.1. We verified that
target vocabulary of 1K byte pair encoding (BPE)        increasing the dropout does not help the vanilla
units (Kudo and Richardson, 2018).                      multi-decoder ST model.
   We prepare the corpus using the ESPnet library          During inference, we perform beam search on
and we follow the standard data preparation, where      both the ASR and ST output sequences, as dis-
inputs are globally mean-variance normalized log-       cussed in §3. The ST beam search is identical
mel filterbank and pitch features from up-sampled       to that of the baseline. For the intermediate ASR
16kHz audio (Watanabe et al., 2018). We also ap-        beam search, we use a beam size of 16, length
ply speed perturbations of 0.9 and 1.1 and the SS       penalty of 0.2, max length ratio of 0.3. In some of
SpecAugment policy (Park et al., 2019).                 our experiments, we also include fusion of a source
                                                        language LM with a 0.2 weight and CTC with a
Baseline Configuration: All of our models are           0.3 weight to re-score the intermediate ASR beam
implemented using the ESPnet library and trained        search (Watanabe et al., 2017). For the Speech-
on 3 NVIDIA Titan 2080Ti GPUs for ≈12 hours.            Attention variant, we increase LM weight to 0.4.
For the Baseline Enc-Dec baseline, discussed in            Note that the ST beam search configuration
§4, we use an E NCODER ASR consisting of a con-         remains constant across our baseline and Multi-
volutional sub-sampling by a factor of 4 (Watan-        Decoder experiments as our focus is on improving
abe et al., 2018) and 12 transformer encoder            overall performance through searchable intermedi-
blocks with 2048 feed-forward dimension, 256            ate representations. Thus, the various re-scoring
attention dimension, and 4 attention heads. The         techniques applied to the ASR beam search are op-
D ECODER ASR and D ECODER ST both consist of 6          tions newly enabled by our proposed architecture
transformer decoder blocks with the same configu-       and are not used in the ST beam search.
ration as E NCODER ASR . There are 37.9M trainable
parameters. We apply dropout of 0.1 for all com-        6   Results
ponents, detailed in the Appendix (A.1).
   We train our models using an effective batch-        Table 1 presents the overall ST performance
size of 384 utterances and use the Adam optimizer       (BLEU) of our proposed Multi-Decoder
(Kingma and Ba, 2015) with inverse square root          model.     Our model improves by +2.9/+0.3
decay learning rate schedule. We set learning rate      (Fisher/CallHome) over the best cascaded baseline
to 12.5, warmup steps to 25K, and epochs to 50. We      and by +5.6/+1.5 BLEU over the best published
use joint training with hybrid CTC/attention ASR        end-to-end baselines. With Speech-Attention,
(Watanabe et al., 2017) by setting mtl-alpha to 0.3     our model improves by +3.4/+1.6 BLEU over
and asr-weight to 0.5 as defined by Watanabe et al.     the cascaded baselines and +7.1/+2.8 BLEU
(2018). During inference, we perform beam search        over encoder-decoder baselines. Both the Multi-
(Seki et al., 2019) on the ST sequences, using a        Decoder and Multi-Decoder w/ Speech-Attention
beam size of 10, length penalty of 0.2, max length      on average are further improved by +0.9/+0.4
ratio of 0.3 (Watanabe et al., 2018).                   BLEU through ASR re-scoring.4
                                                           Table 1 also includes our implementation of the
Multi-Decoder Configuration: For the Multi-             Baseline Enc-Dec model discussed in §4. In this
Decoder ST model, discussed in §3, we use               way, we are able to make a fair comparison with our
the same transformer configuration as the base-         framework as we control the model and inference
line for the E NCODER ASR , D ECODER ASR , and             4
                                                             We also evaluate our models using other MT metrics to
D ECODER ST . Additionally, the Multi-Decoder           supplement these results, as shown in the Appendix (A.2).
Uses Speech                                 Fisher                CallHome
  Model Type       Model Name                    Transcripts                    dev(↑)     dev2(↑)    test(↑)   dev(↑)     test(↑)
  Cascade          Inaguma et al. (2020)             3                           41.5        43.5      42.2      19.6       19.8
  Cascade          ESPnet ASR+MT (2018)              3                           50.4        51.2      50.7      19.6       19.2
  Enc-Dec          Weiss et al. (2017) ♦             7                           46.5        47.3      47.3      16.4       16.6
  Enc-Dec          Weiss et al. (2017) ♦             3                           48.3        49.1      48.7      16.8       17.4
  Enc-Dec          Inaguma et al. (2020)             3                           46.6        47.6      46.5      16.8       16.8
  Enc-Dec          Guo et al. (2021)                 3                           48.7        49.6      47.0      18.5       18.6
  Enc-Dec          Our Implementation                3                           49.6        50.9      49.5      19.1       18.2
  Multi-Decoder    Our Proposed Model                3                           52.7        53.3      52.6      20.5       20.1
  Multi-Decoder     +ASR Re-scoring                  3                           53.3        54.2      53.7      21.1       20.8
  Multi-Decoder     +Speech-Attention                3                           54.6        54.6      54.1      21.7       21.4
  Multi-Decoder      +ASR Re-scoring                 3                           55.2        55.2      55.0      21.7       21.5
Table 1: Results presenting the overall performance (BLEU) of our proposed multi-decoder model. Cascade and
Enc-Dec results from previous papers and our own implementation of the Enc-Dec are shown for comparison. The
best performing models are highlighted. ♦ Implemented with LSTM, while all others are Transformer-based.

                                                                                                                           23.8
                       Overall   Sub-Net   Sub-Net
 Model                 ST(↑)     ASR(↓)     MT(↑)                             52.6                                         23.6
                                                                                                           Multi-Decoder
                                                          ST BLEU Score (↑)

 Multi-Decoder           52.7      22.6      64.9                             52.4

                                                                                                                                   ASR % WER (↓)
                                                                                                                  BLEU     23.4
  +Speech-Attention      54.6      22.4      66.6                                                                % WER
                                                                              52.2                                         23.2
Table 2: Results presenting the overall ST performance                         52
(BLEU) of our Multi-Decoder models, along with their                                                                       23
sub-net ASR (% WER) and MT (BLEU) performances.                               51.8
                                                                                                                           22.8
All results are from the Fisher dev set.
                                                                              51.6
                                                                                                                           22.6
                                                                                       1    4       8 10             16
configurations to be analagous. For instance, we                                                ASR Beam Size
keep the same search parameters for the final output
in the baseline and the Multi-Decoder to demon-          Figure 2: Results studying the effect of the differ-
                                                         ent ASR beam sizes in the intermediate representa-
strate impact of the intermediate beam search.
                                                         tion search on the overall ST performance (BLEU) and
6.1     Benefits                                         the ASR sub-net performance (% WER) for our multi-
                                                         decoder model. Beam of 1 is same as greedy search.
6.1.1    Sub-network performance monitoring
An added benefit of our proposed approach over the
Baseline Enc-Dec is the ability to monitor the indi-     beam size of 1, which is a greedy search, results in
vidual performances of the ASR (% WER) and MT            lower ASR sub-net and overall ST performances.
(BLEU) sub-nets as shown in Table 2. The Multi-          As beam sizes become larger, gains taper off as can
Decoder w/ Speech-Attention shows a greater MT           be seen between beam sizes of 10 and 16.
sub-net performance than the Multi-Decoder as
                                                         6.1.3                       External models for better search
well as a slight improvement of the ASR sub-net,
suggesting that ST can potentially help ASR.             External models like CTC acoustic models and lan-
                                                         guage models are commonly used for re-scoring
6.1.2    Beam search for better intermediates            encoder-decoder models (Hori et al., 2017), due to
The overall ST performance improves when a               the difference in their modeling capabilities. CTC
higher beam size is used in the intermediate ASR         directly models transcripts while being condition-
search, and this increase can be attributed to the im-   ally independent on the other outputs given the in-
proved ASR sub-net performance. Figure 1 shows           put, and LMs predict the next token in a sequence.
this trend across ASR beam sizes of 1, 4, 8, 10, 16         Both variants of the Multi-Decoder improve due
while fixing the ST decoding beam size to 10. A          to improved ASR sub-net performance using exter-
Overall   Sub-Net                           40
 Model                            ST(↑)     ASR(↓)                                                          Baseline Enc-Dec
                                                                                           32.1 33.2        Multi-Decoder
 Multi-Decoder                     52.7      22.6                                   29.9                    Multi-Decoder w/ SA

                                                          ST BLEU Score (↑)
                                                                              30
  +ASR Re-scoring w/ LM            53.2      22.6
  +ASR Re-scoring w/ CTC           52.8      22.1
   +ASR Re-scoring w/ LM           53.3      21.7                                                      20.1 19.1 21.2
                                                                              20
 Multi-Decoder w/ Speech-Attn.     54.6      22.4
  +ASR Re-scoring w/ LM            55.1      22.4
  +ASR Re-scoring w/ CTC           54.7      22.0
   +ASR Re-scoring w/ LM           55.2      21.9                             10
                                                                                                                        5.4 5.8 5.6
Table 3: Results presenting the overall ST performance
(BLEU) and the sub-net ASR (% WER) of our Multi-                                       < 40%            [40, 80)%         ≥ 80%
Decoder models with external CTC and LM re-scoring
                                                                                                   ASR % WER (↓)
in the ASR intermediate representation search. All re-
sults are from the Fisher dev set.
                                                         Figure 3: Results comparing the ST performances
                                                         (BLEU) of our Baseline Enc-Dec, Multi-Decoder, and
                                                         Multi-Decoder w/ Speech-Attention across different
nal CTC and LM models for re-scoring, as shown
                                                         ASR difficulties measured using % WER on the Fisher
in Table 3. We use a recurrent neural network LM         dev set (1-ref). The buckets on the x-axis are de-
trained on the Fisher-CallHome Spanish transcripts       termined using the utterance level % WER using the
with a dev perplexity of 18.8 and the CTC model          Multi-Decoder ASR sub-net performance.
from joint loss applied during training. Neither
external model incorporates additional data. Al-
though the impact of the LM-only re-scoring is not       6.2.1                     Robustness through Decomposition
shown in the ASR % WER, it reduces substitution          Like cascaded systems, searchable intermediates
and deletion rates in the ASR and this is observed       provide our model adaptability in individual sub-
to help the overall ST performance.                      systems towards out-of-domain data using external
                                                         in-domain language model, thereby giving access
6.1.4    Error propagation avoidance
                                                         to more in-domain data. Specifically for speech
As discussed in §3, our Multi-Decoder model in-          translation systems, this means we can use in-
herits the error propagation issue as can be seen        domain language models in both source and target
in Figure 3. For the easiest bucket of utterances        languages. We test the robustness of our Multi-
with < 40% WER in Multi-Decoder’s ASR sub-               Decoder model trained on Fisher-CallHome con-
net, our model’s ST performance, as measured by          versational speech dataset on read speech CoVost-2
the corpus BLEU of the bucket, exceeds that of           dataset (Wang et al., 2020b). In Table 4 we show
the Baseline Enc-Dec. The inverse is true for the        that re-scoring the ASR sub-net with an in-domain
more difficult bucket of [40, 80)%, showing that         LM improves ASR with around 10.0% lower WER,
error propagation is limiting the performance of         improving the overall ST performance by around
our model; however, we show that multi-sequence          +2.5 BLEU. Compared to an in-domain ST base-
attention can alleviate this issue. For extremely        line (Wang et al., 2020a), our out-of-domain Multi-
difficult utterances in the ≥ 80% bucket, ST perfor-     Decoder with in-domain ASR re-scoring demon-
mance for all three approaches is suppressed. We         strates the robustness of our approach.
also provide qualitative examples of error propaga-
tion avoidance in the Appendix (A.3).                    6.2.2                     Decomposing Speech Transcripts
                                                         We apply our generic framework to another de-
6.2     Generalizability
                                                         composable sequence task, speech recognition, and
In this section, we discuss the generalizability of      show the results of various levels of decomposition
our framework towards out-of-domain data. We             in Table 5. We show that with phoneme, character,
also extend our Multi-Decoder model to other se-         or byte-pair encoding (BPE) sequences as interme-
quence tasks like speech recognition. Finally, we        diates, the Multi-Decoder presents strong results
apply our ST models to a larger corpus with more         on both Fisher and CallHome test sets. We also
language pairs and a different domain of speech.         observe that the BPE intermediates perform bet-
Overall   Sub-Net                                          En→De    En→Fr
 Model                                    ST(↑)     ASR(↓)        Model                               ST(↑)    ST(↑)
 I N - DOMAIN ST M ODEL                                           NeurST (Zhao et al., 2020)          22.9     33.3
 Baseline (Wang et al., 2020b)             12.0       -
                                                                  Fairseq S2T (Wang et al., 2020a)    22.7     32.9
   +ASR Pretrain (Wang et al., 2020b) ♦    23.0      16.0
                                                                  ESPnet-ST (Inaguma et al., 2020)    22.9     32.7
 O UT- OF - DOMAIN ST M ODEL                                      Dual-Decoder (Le et al., 2020)      23.6     33.5
 Multi-Decoder                             11.8      46.8
  +ASR Re-scoring w/ in-domain LM          14.4      36.7         Multi-Decoder w/ Speech-Attn.       26.3     37.0
 Multi-Decoder w/ Speech-Attention         12.6      46.5          +ASR Re-scoring                    26.4     37.4
  +ASR Re-scoring w/ in-domain LM          15.0      36.7
                                                              Table 6: Results presenting the overall ST performance
Table 4: Results presenting the overall ST perfor-            (BLEU) of our Multi-Decoder w/ Speech-Attention
mance (BLEU) and the sub-net ASR (% WER) of our               models with ASR re-scoring across two language-
Multi-Decoder models when tested on out-of-domain             pairs, English-German (En→De) and English-French
data. All models were trained on the Fisher-CallHome          (En→Fr). All results are from the MuST-C tst-
Es→En corpus and tested on CoVost2 Es→En corpus.              COMMON sets. All models use speech transcripts.
♦
  Pretrained with 364 hours of in-domain ASR data.

                                                              approach across several dimensions of ST tasks.
                                    Fisher        CallHome
 Model              Intermediate    ASR(↓)         ASR(↓)     First, our approach consistently improves over base-
           ♦
                                                              lines across multiple language-pairs. Second, our
 Enc-Dec                  -           23.2          45.3
                                                              approach is robust to the distinct domains of tele-
 Multi-Decoder       Phoneme          20.7          40.0      phone conversations from Fisher-CallHome and
 Multi-Decoder       Character        20.4          39.9
                                                              the TED-Talks from MuST-C. Finally, by scaling
 Multi-Decoder       BPE100           19.7          38.9
                                                              from 170 hours of Fisher-CallHome data to 500
Table 5: Results presenting the % WER ASR perfor-             hours of MuST-C data, we show that the benefits
mance when using the Multi-Decoder model on de-               of decomposing sequence tasks with searchable
composed ASR task with phoneme, character, and                hidden intermediates persist even with more data.
BPE100 as intermediates. All results are from the
                                                                 Furthermore, the performance of our Multi-
Fisher-CallHome Spanish corpus. ♦ (Weiss et al., 2017)
                                                              Decoder models trained with only English-German
                                                              or English-French ST data from MuST-C is com-
ter than phoneme/character variants, which could              parable to other methods which incorporate larger
be attributed to the reduced search capabilities              external ASR and MT data in various ways. For in-
of encoder-decoder models using beam search on                stance, Zheng et al. (2021) use 4700 hours of ASR
longer sequences (Sountsov and Sarawagi, 2016)                data and 2M sentences of MT data for pretrain-
like in phoneme/character sequences.                          ing and multi-task learning. Similarly, Bahar et al.
                                                              (2021) use 2300 hours of ASR data and 27M sen-
6.2.3    Extending to MuST-C Language Pairs                   tences of MT data for pretraining. Our competitive
In addition to our results using the 170 hours of the         performance without the use of any additional data
Spanish-English Fisher-CallHome corpus, in Ta-                highlights the data-efficient nature of our proposed
ble 6 we show that our decompositional framework              end-to-end framework as opposed to the baseline
is also effective on larger ST corpora. In particu-           encoder-decoder model, as pointed out by Sperber
lar, we use 400 hours of English-German and 500               and Paulik (2020).
hours of English-French ST from the MuST-C cor-
pus (Di Gangi et al., 2019). Our Multi-Decoder                7     Discussion and Relation to Prior Work
model improves by +2.7 and +1.5 BLEU, in Ger-
                                                              Compositionality: A number of recent works
man and French respectively, over end-to-end base-
                                                              have constructed composable neural network mod-
lines from prior works that do not use additional
                                                              ules for tasks such as visual question answering
training data. We show that ASR re-scoring gives
                                                              (Andreas et al., 2016), neural MT (Raunak et al.,
an additional +0.1 and +0.4 BLEU improvement. 5
                                                              2019), and synthetic sequence-to-sequence tasks
   By extending our Multi-Decoder models to this              (Lake, 2019). Modules that are first trained sepa-
MuST-C study, we show the generalizability of our             rately can subsequently be tightly integrated into a
   5
     Details of the MuST-C data preparation and model pa-     single end-to-end trainable model by passing differ-
rameters are detailed in Appendix (A.4).                      entiable soft decisions instead of discrete decisions
in the intermediate stage (Bahar et al., 2021). Fur-    impacting the performance of both the task at hand
ther, even a single encoder-decoder model can be        and any downstream tasks. Our approach allevi-
decomposed into modular components where the            ates these problems through intermediate search,
encoder and decoder modules have explicit func-         external models for intermediate re-scoring, and
tions (Dalmia et al., 2019).                            multi-sequence attention.

Joint Training with Sub-Tasks: End-to-end se-           8   Conclusion and Future Work
quence models been shown to benefit from intro-
ducing joint training with sub-tasks as auxiliary       We present searchable hidden intermediates for end-
loss functions for a variety of tasks like ASR (Kim     to-end models of decomposable sequence tasks.
et al., 2017), ST (Salesky et al., 2019; Liu et al.,    We show the efficacy of our Multi-Decoder model
2020a; Dong et al., 2020; Le et al., 2020), SLU         on the Fisher-CallHome Es→En and MuST-C
(Haghani et al., 2018). They have been shown to in-     En→De and En→Fr speech translation corpora,
duce structure (Belinkov et al., 2020) and improve      achieving state-of-the-art results. We present var-
the model performance (Toshniwal et al., 2017),         ious benefits in our framework, including sub-net
but this joint training may reduce data efficiency      performance monitoring, beam search for better
if some sub-nets are not included in the final end-     hidden intermediates, external models for better
to-end model (Sperber et al., 2019; Wang et al.,        search, and error propagation avoidance. Further,
2020c). Our framework avoids this sub-net waste         we demonstrate the flexibility of our framework
at the cost of computational load during inference.     towards out-of-domain tasks with the ability to
                                                        adapt our sequence model at intermediate stages of
Speech Translation Decoders: Prior works                decomposition. Finally, we show generalizability
have used ASR/MT decoding to improve the over-          by training Multi-Decoder models for the speech
all ST decoding through synchronous decoding            recognition task at various levels of decomposition.
(Liu et al., 2020a), dual decoding (Le et al., 2020),      We hope insights derived from our study stim-
and successive decoding (Dong et al., 2020). These      ulate research on tighter integrations between the
works partially or fully decode ASR transcripts and     benefits of cascaded and end-to-end sequence mod-
use discrete intermediates to assist MT decoding.       els. Exploiting searchable intermediates through
Tu et al. (2017) and Anastasopoulos and Chiang          beam search is just the tip of the iceberg for search
(2018) are closest to our multi-decoder ST model,       algorithms, as numerous approximate search tech-
however the benefits of our proposed framework          niques like diverse beam search (Vijayakumar et al.,
are not entirely explored in these works.               2018) and best-first beam search (Meister et al.,
                                                        2020) have been recently proposed to improve di-
Two-Pass Decoding: Two-pass decoding in-                versity and approximation of the most-likely se-
volves first predicting with one decoder and then       quence. Incorporating differentiable lattice based
re-evaluating with another decoder (Geng et al.,        search (Hannun et al., 2020) can also allow the sub-
2018; Sainath et al., 2019; Hu et al., 2020; Rijh-      sequent sub-net to digest n-best representations.
wani et al., 2020). The two decoders iterate on the
same sequence, so there is no decomposition into        9   Acknowledgements
sub-tasks in this method. On the other hand, our
approach provides the subsequent decoder with a         This work started while Vikas Raunak was a stu-
more structured representation than the input by de-    dent at CMU, he is now working as a Research Sci-
composing the complexity of the overall task. Like      entist at Microsoft. We thank Pengcheng Guo, Hi-
two-pass decoding, our approach provides a sense        rofumi Inaguma, Elizabeth Salesky, Maria Ryskina,
of the future to the second decoder which allows it     Marta Méndez Simón and Vijay Viswanathan for
to correct mistakes from the previous first decoder.    their helpful discussion during the course of this
                                                        project. We also thank the anonymous reviewers
Auto-Regressive Decoding: As auto-regressive            for their valuable feedback. This work used the
decoders inherently learn a language model along        Extreme Science and Engineering Discovery En-
with the task at hand, they tend to be domain spe-      vironment (XSEDE) (Towns et al., 2014), which
cific (Samarakoon et al., 2018; Müller et al., 2020).   is supported by National Science Foundation grant
This can cause generalizability issues during infer-    number ACI-1548562. Specifically, it used the
ence (Murray and Chiang, 2018; Yang et al., 2018),      Bridges system (Nystrom et al., 2015), which is
supported by NSF award number ACI-1445606,                 Yonatan Belinkov, Nadir Durrani, Fahim Dalvi, Has-
at the Pittsburgh Supercomputing Center (PSC).               san Sajjad, and James Glass. 2020. On the linguistic
                                                             representational power of neural machine translation
The work was supported in part by an AWS Ma-
                                                             models. Computational Linguistics, 46(1):1–52.
chine Learning Research Award. This research
was also supported in part the DARPA KAIROS                Léon Bottou, Yoshua Bengio, and Yann Le Cun. 1997.
program from the Air Force Research Laboratory               Global training of document processing systems us-
under agreement number FA8750-19-2-0200. The                 ing graph transformer networks. In Proceedings of
                                                             IEEE Computer Society Conference on Computer Vi-
U.S. Government is authorized to reproduce and               sion and Pattern Recognition, pages 489–494. IEEE.
distribute reprints for Governmental purposes not
withstanding any copyright notation there on. The          Alice Coucke, Alaa Saade, Adrien Ball, Théodore
views and conclusions contained herein are those of          Bluche, Alexandre Caulier, David Leroy, Clément
the authors and should not be interpreted as neces-          Doumouro, Thibault Gisselbrecht, Francesco Calt-
                                                             agirone, Thibaut Lavril, et al. 2018. Snips voice
sarily representing the official policies or endorse-        platform: an embedded spoken language understand-
ments, either expressed or implied, of the Air Force         ing system for private-by-design voice interfaces. In
Research Laboratory or the U.S. Government.                  Privacy in Machine Learning and Artificial Intelli-
                                                             gence workshop, ICML.

References                                                 Siddharth Dalmia, Abdelrahman Mohamed, Mike
                                                              Lewis, Florian Metze, and Luke Zettlemoyer.
Antonios Anastasopoulos and David Chiang. 2018.               2019. Enforcing encoder-decoder modularity in
  Tied multitask learning for neural speech translation.      sequence-to-sequence models.   arXiv preprint
  In Proceedings of the 2018 Conference of the North          arXiv:1911.03782.
  American Chapter of the Association for Computa-
  tional Linguistics: Human Language Technologies,         Mattia A. Di Gangi, Roldano Cattoni, Luisa Bentivogli,
  Volume 1 (Long Papers), pages 82–91, New Orleans,         Matteo Negri, and Marco Turchi. 2019. MuST-
  Louisiana. Association for Computational Linguis-         C: a Multilingual Speech Translation Corpus. In
  tics.                                                     Proceedings of the 2019 Conference of the North
                                                            American Chapter of the Association for Computa-
Jacob Andreas, Marcus Rohrbach, Trevor Darrell, and
                                                            tional Linguistics: Human Language Technologies,
   Dan Klein. 2016. Neural module networks. In 2016
                                                            pages 2012–2017, Minneapolis, Minnesota. Associ-
   IEEE Conference on Computer Vision and Pattern
                                                            ation for Computational Linguistics.
   Recognition, CVPR 2016, Las Vegas, NV, USA, June
   27-30, 2016, pages 39–48. IEEE Computer Society.
                                                           Qianqian Dong, Mingxuan Wang, Hao Zhou, Shuang
Parnia Bahar, Tobias Bieschke, Ralf Schlüter, and            Xu, Bo Xu, and Lei Li. 2020. SDST: Successive de-
  Hermann Ney. 2021.       Tight integrated end-to-          coding for speech-to-text translation. Proceedings
  end training for cascaded speech translation. In           of the Thirty-Fifth AAAI Conference on Artificial In-
  2021 IEEE Spoken Language Technology Workshop              telligence.
  (SLT), pages 950–957. IEEE.
                                                           Xinwei Geng, Xiaocheng Feng, Bing Qin, and Ting
Dzmitry Bahdanau, Kyunghyun Cho, and Yoshua Ben-             Liu. 2018. Adaptive multi-pass decoder for neural
  gio. 2015. Neural machine translation by jointly           machine translation. In Proceedings of the 2018
  learning to align and translate.   In 3rd Inter-           Conference on Empirical Methods in Natural Lan-
  national Conference on Learning Representations,           guage Processing, pages 523–532, Brussels, Bel-
  ICLR 2015.                                                 gium. Association for Computational Linguistics.
Satanjeev Banerjee and Alon Lavie. 2005. METEOR:
  An automatic metric for MT evaluation with im-           Pengcheng Guo, Florian Boyer, Xuankai Chang,
  proved correlation with human judgments. In Pro-           Tomoki Hayashi, Yosuke Higuchi, Hirofumi In-
  ceedings of the ACL Workshop on Intrinsic and Ex-          aguma, Naoyuki Kamo, Chenda Li, Daniel Garcia-
  trinsic Evaluation Measures for Machine Transla-           Romero, Jiatong Shi, et al. 2021. Recent develop-
  tion and/or Summarization, pages 65–72, Ann Ar-            ments on ESPnet toolkit boosted by conformer. In
  bor, Michigan. Association for Computational Lin-          2021 IEEE international conference on acoustics,
  guistics.                                                  speech and signal processing (ICASSP). IEEE.

Daniel Beck, Trevor Cohn, and Gholamreza Haffari.          Parisa Haghani, Arun Narayanan, Michiel Bacchiani,
  2019. Neural speech translation using lattice trans-       Galen Chuang, Neeraj Gaur, Pedro Moreno, Rohit
  formations and graph networks. In Proceedings of           Prabhavalkar, Zhongdi Qu, and Austin Waters. 2018.
  the Thirteenth Workshop on Graph-Based Methods             From audio to semantics: Approaches to end-to-end
  for Natural Language Processing (TextGraphs-13),           spoken language understanding. In 2018 IEEE Spo-
  pages 26–31, Hong Kong. Association for Computa-           ken Language Technology Workshop (SLT), pages
  tional Linguistics.                                        720–726. IEEE.
Awni Hannun, Vineel Pratap, Jacob Kahn, and Wei-             Constantin, and Evan Herbst. 2007. Moses: Open
  Ning Hsu. 2020. Differentiable weighted finite-state       source toolkit for statistical machine translation. In
  transducers. arXiv preprint arXiv:2010.01003.              Proceedings of the 45th Annual Meeting of the As-
                                                             sociation for Computational Linguistics Companion
Jindřich Helcl, Jindřich Libovický, and Dušan Variš.       Volume Proceedings of the Demo and Poster Ses-
   2018. CUNI system for the WMT18 multimodal                sions, pages 177–180, Prague, Czech Republic. As-
   translation task. In Proceedings of the Third Confer-     sociation for Computational Linguistics.
   ence on Machine Translation: Shared Task Papers,
   pages 616–623, Belgium, Brussels. Association for       Philipp Koehn and Josh Schroeder. 2007. Experiments
   Computational Linguistics.                                in domain adaptation for statistical machine transla-
                                                             tion. In Proceedings of the second workshop on sta-
Geoffrey Hinton, Li Deng, Dong Yu, George E Dahl,            tistical machine translation, pages 224–227.
  Abdel-rahman Mohamed, Navdeep Jaitly, Andrew
  Senior, Vincent Vanhoucke, Patrick Nguyen, Tara N        Taku Kudo and John Richardson. 2018. SentencePiece:
  Sainath, et al. 2012. Deep neural networks for             A simple and language independent subword tok-
  acoustic modeling in speech recognition: The shared        enizer and detokenizer for neural text processing. In
  views of four research groups. IEEE Signal process-        Proceedings of the 2018 Conference on Empirical
  ing magazine, 29(6):82–97.                                 Methods in Natural Language Processing: System
                                                             Demonstrations, pages 66–71.
Takaaki Hori, Shinji Watanabe, Yu Zhang, and William
  Chan. 2017. Advances in joint CTC-attention based        Gaurav Kumar, Matt Post, Daniel Povey, and Sanjeev
  end-to-end speech recognition with a deep CNN en-          Khudanpur. 2014. Some insights from translating
  coder and RNN-LM. In Proc. Interspeech 2017,               conversational telephone speech. In IEEE Interna-
  pages 949–953.                                             tional Conference on Acoustics, Speech and Signal
                                                             Processing, ICASSP 2014, Florence, Italy, May 4-9,
Ke Hu, Tara N Sainath, Ruoming Pang, and Rohit Prab-         2014, pages 3231–3235. IEEE.
  havalkar. 2020. Deliberation model based two-pass
  end-to-end speech recognition. In ICASSP 2020-           Shankar Kumar, Yonggang Deng, and William Byrne.
  2020 IEEE International Conference on Acoustics,           2006. A weighted finite state transducer translation
  Speech and Signal Processing (ICASSP), pages               template model for statistical machine translation.
  7799–7803. IEEE.                                           Natural Language Engineering, 12(1):35–76.
                                                           Chih-Hua Kuo. 1995. Cohesion and coherence in aca-
Liang Huang and David Chiang. 2007. Forest Rescor-
                                                             demic writing: From lexical choice to organization.
  ing: Faster decoding with integrated language mod-
                                                             RELC Journal, 26(1):47–62.
  els. In Proceedings of the 45th Annual Meeting of
  the Association of Computational Linguistics, pages      Brenden M Lake. 2019. Compositional generalization
  144–151, Prague, Czech Republic. Association for           through meta sequence-to-sequence learning. In Ad-
  Computational Linguistics.                                 vances in Neural Information Processing Systems,
                                                             pages 9791–9801.
Hirofumi Inaguma, Shun Kiyono, Kevin Duh, Shigeki
  Karita, Nelson Yalta, Tomoki Hayashi, and Shinji         Brenden M. Lake and Marco Baroni. 2018. General-
  Watanabe. 2020. ESPnet-ST: All-in-one speech               ization without systematicity: On the compositional
  translation toolkit. In Proceedings of the 58th An-        skills of sequence-to-sequence recurrent networks.
  nual Meeting of the Association for Computational          In Proceedings of the 35th International Conference
  Linguistics: System Demonstrations, pages 302–311.         on Machine Learning, ICML 2018, pages 2879–
  Association for Computational Linguistics.                 2888. PMLR.
Patricia Johnson. 1992. Cohesion and coherence in          Hang Le, Juan Pino, Changhan Wang, Jiatao Gu, Di-
  compositions in Malay and English. RELC Journal,           dier Schwab, and Laurent Besacier. 2020. Dual-
  23(2):1–17.                                                decoder transformer for joint automatic speech
                                                             recognition and multilingual speech translation.
Suyoun Kim, Takaaki Hori, and Shinji Watanabe. 2017.         Proceedings of the 28th International Conference on
  Joint CTC-attention based end-to-end speech recog-         Computational Linguistics.
  nition using multi-task learning. In 2017 IEEE inter-
  national conference on acoustics, speech and signal      Alexander H. Levis, Neville Moray, and Baosheng Hu.
  processing (ICASSP), pages 4835–4839. IEEE.                1994. Task decomposition and allocation problems
                                                             and discrete event systems. Automatica, 30(2):203 –
Diederik P. Kingma and Jimmy Ba. 2015. Adam: A               216.
  method for stochastic optimization. In 3rd Inter-
  national Conference on Learning Representations,         Yuchen Liu, Jiajun Zhang, Hao Xiong, Long Zhou,
  ICLR 2015.                                                 Zhongjun He, Hua Wu, Haifeng Wang, and
                                                             Chengqing Zong. 2020a.          Synchronous speech
Philipp Koehn, Hieu Hoang, Alexandra Birch, Chris            recognition and speech-to-text translation with inter-
  Callison-Burch, Marcello Federico, Nicola Bertoldi,        active decoding. In The Thirty-Fourth AAAI Con-
  Brooke Cowan, Wade Shen, Christine Moran,                  ference on Artificial Intelligence, AAAI 2020, pages
  Richard Zens, Chris Dyer, Ondřej Bojar, Alexandra         8417–8424.
Zhiyuan Liu, Yankai Lin, and Maosong Sun. 2020b.         N. Pham, Thai-Son Nguyen, Thanh-Le Ha, J. Hussain,
  Compositional Semantics, pages 43–57. Springer           Felix Schneider, J. Niehues, Sebastian Stüker, and
  Singapore, Singapore.                                    A. Waibel. 2019. The IWSLT 2019 KIT speech
                                                           translation system. In International Workshop on
Clara Meister, Tim Vieira, and Ryan Cotterell. 2020.       Spoken Language Translation (IWSLT).
  Best-first beam search. Transactions of the Associa-
  tion for Computational Linguistics, 8:795–809.         Matt Post. 2018. A call for clarity in reporting BLEU
                                                          scores. In Proceedings of the Third Conference on
Bernd T Meyer, Sri Harish Mallidi, Angel Mario Cas-       Machine Translation: Research Papers, pages 186–
  tro Martinez, Guillermo Payá-Vayá, Hendrik Kayser,      191, Belgium, Brussels. Association for Computa-
  and Hynek Hermansky. 2016. Performance monitor-         tional Linguistics.
  ing for automatic speech recognition in noisy multi-
  channel environments. In 2016 IEEE Spoken Lan-         Matt Post, Gaurav Kumar, Adam Lopez, Damianos
  guage Technology Workshop (SLT), pages 50–56.           Karakos, Chris Callison-Burch, and Sanjeev Khu-
                                                          danpur. 2013. Improved speech-to-text transla-
Mehryar Mohri, Fernando Pereira, and Michael Ri-          tion with the Fisher and Callhome Spanish–English
 ley. 2002. Weighted finite-state transducers in          speech translation corpus. In International Work-
 speech recognition. Computer Speech & Language,          shop on Spoken Language Translation (IWSLT
 16(1):69–88.                                             2013).
Mathias Müller, Annette Rios, and Rico Sennrich.         Vikas Raunak, Vaibhav Kumar, and Florian Metze.
 2020. Domain robustness in neural machine trans-          2019. On compositionality in neural machine trans-
 lation. In Proceedings of the 14th Conference of          lation. NeurIPS Workshop, Context and Composi-
 the Association for Machine Translation in the Amer-      tionality in Biological and Artificial Neural Systems.
 icas, pages 151–164, Virtual. Association for Ma-
 chine Translation in the Americas.                      Raj Reddy. 1988. Foundations and grand challenges
                                                           of artificial intelligence: AAAI presidential address.
Kenton Murray and David Chiang. 2018. Correcting           AI Mag., 9(4):9–21.
  length bias in neural machine translation. In Pro-
                                                         Shruti Rijhwani, Antonios Anastasopoulos, and Gra-
  ceedings of the Third Conference on Machine Trans-
                                                           ham Neubig. 2020. OCR post correction for endan-
  lation: Research Papers, pages 212–223, Brussels,
                                                           gered language texts. In Proceedings of the 2020
  Belgium.
                                                           Conference on Empirical Methods in Natural Lan-
Nicholas A. Nystrom, Michael J. Levine, Ralph Z.           guage Processing (EMNLP), pages 5931–5942, On-
  Roskies, and J. Ray Scott. 2015. Bridges: A              line. Association for Computational Linguistics.
  uniquely flexible HPC resource for new commu-          Tara N Sainath, Ruoming Pang, David Rybach,
  nities and data analytics. In Proceedings of the         Yanzhang He, Rohit Prabhavalkar, Wei Li, Mirkó Vi-
  2015 XSEDE Conference: Scientific Advancements           sontai, Qiao Liang, Trevor Strohman, Yonghui Wu,
  Enabled by Enhanced Cyberinfrastructure. Associa-        et al. 2019. Two-pass end-to-end speech recognition.
  tion for Computing Machinery.                            Proc. Interspeech 2019, pages 2773–2777.
Franz Josef Och and Hermann Ney. 2002. Discrimina-       Elizabeth Salesky, Matthias Sperber, and Alan W
  tive training and maximum entropy models for sta-         Black. 2019. Exploring Phoneme-Level Speech
  tistical machine translation. In Proceedings of the       Representations for End-to-End Speech Translation.
  40th Annual meeting of the Association for Compu-         In Proceedings of the 57th Annual Meeting of the
  tational Linguistics, pages 295–302.                      Association for Computational Linguistics, pages
                                                           1835–1841, Florence, Italy. Association for Compu-
Kishore Papineni, Salim Roukos, Todd Ward, and Wei-
                                                            tational Linguistics.
  Jing Zhu. 2002. BLEU: a method for automatic eval-
  uation of machine translation. In Proceedings of       Lahiru Samarakoon, Brian Mak, and Albert YS
  the 40th Annual Meeting of the Association for Com-      Lam. 2018.     Domain adaptation of end-to-end
  putational Linguistics, pages 311–318, Philadelphia,     speech recognition in low-resource settings. In
  Pennsylvania, USA.                                       2018 IEEE Spoken Language Technology Workshop
                                                           (SLT), pages 382–388. IEEE.
Daniel S Park, William Chan, Yu Zhang, Chung-Cheng
  Chiu, Barret Zoph, Ekin D Cubuk, and Quoc V Le.        Hiroshi Seki, Takaaki Hori, Shinji Watanabe, Niko
  2019. SpecAugment: A simple data augmentation            Moritz, and Jonathan Le Roux. 2019. Vectorized
  method for automatic speech recognition. Proc. In-       beam search for CTC-attention-based speech recog-
  terspeech 2019, pages 2613–2617.                         nition. In Proc. Interspeech 2019, pages 3825–
                                                           3829.
Vijayaditya Peddinti, Guoguo Chen, Vimal Manohar,
  Tom Ko, Daniel Povey, and Sanjeev Khudanpur.           Matthew Snover, Bonnie Dorr, Richard Schwartz, Lin-
   2015. JHU ASpIRE system: Robust LVCSR with             nea Micciulla, and John Makhoul. 2006. A study of
  TDNNS, iVector adaptation and RNN-LMS. In               translation edit rate with targeted human annotation.
  2015 IEEE Workshop on Automatic Speech Recog-           In Proceedings of Association for Machine Transla-
   nition and Understanding (ASRU), pages 539–546.        tion in the Americas.
Pavel Sountsov and Sunita Sarawagi. 2016. Length            In Proceedings of the Thirty-Second AAAI Confer-
  bias in encoder decoder models and a case for global      ence on Artificial Intelligence, (AAAI-18), pages
  conditioning. In Proceedings of the 2016 Confer-          7371–7379. AAAI Press.
  ence on Empirical Methods in Natural Language
  Processing, pages 1516–1525, Austin, Texas. Asso-       Changhan Wang, Yun Tang, Xutai Ma, Anne Wu,
  ciation for Computational Linguistics.                    Dmytro Okhonko, and Juan Pino. 2020a. Fairseq
                                                            S2T: Fast speech-to-text modeling with fairseq. In
Matthias Sperber, Graham Neubig, Jan Niehues, and           Proceedings of the 1st Conference of the Asia-Pacific
 Alex Waibel. 2019. Attention-passing models for ro-        Chapter of the Association for Computational Lin-
 bust and data-efficient end-to-end speech translation.     guistics (AACL): System Demonstrations, pages 33–
 Transactions of the Association for Computational          39. Association for Computational Linguistics.
 Linguistics, 7:313–325.
                                                          Changhan Wang, Anne Wu, and Juan Pino. 2020b.
Matthias Sperber and Matthias Paulik. 2020. Speech          CoVoST 2: A massively multilingual speech-
 translation and the end-to-end promise: Taking stock       to-text translation corpus. arXiv preprint
 of where we are. Proceedings of the 58th Annual            arXiv:2007.10310.
 Meeting of the Association for Computational Lin-
 guistics.                                                Chengyi Wang, Yu Wu, Shujie Liu, Zhenglu Yang, and
                                                            Ming Zhou. 2020c. Bridging the gap between pre-
Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014.         training and fine-tuning for end-to-end speech trans-
   Sequence to sequence learning with neural networks.      lation. Proceedings of the AAAI Conference on Arti-
   In Advances in neural information processing sys-        ficial Intelligence, 34(05):9161–9168.
   tems, pages 3104–3112.
                                                          Shinji Watanabe, Takaaki Hori, Shigeki Karita, Tomoki
Christoph Tillmann and Hermann Ney. 2003. Word re-          Hayashi, Jiro Nishitoba, Yuya Unno, Nelson En-
  ordering and a dynamic programming beam search            rique Yalta Soplin, Jahn Heymann, Matthew Wies-
  algorithm for statistical machine translation. Com-       ner, Nanxin Chen, Adithya Renduchintala, and
  putational linguistics, 29(1):97–133.                     Tsubasa Ochiai. 2018. ESPnet: End-to-end speech
                                                            processing toolkit. In Proc. Interspeech 2018, pages
Shubham Toshniwal, Hao Tang, Liang Lu, and Karen            2207–2211.
  Livescu. 2017. Multitask learning with low-level
  auxiliary tasks for encoder-decoder based speech        Shinji Watanabe, Takaaki Hori, Suyoun Kim, John R.
  recognition. In Proc. Interspeech 2017, pages 3532–       Hershey, and Tomoki Hayashi. 2017.          Hybrid
  3536.                                                     CTC/attention architecture for end-to-end speech
                                                            recognition. IEEE Journal of Selected Topics in Sig-
John Towns, Timothy Cockerill, Maytal Dahan, Ian            nal Processing, 11(8):1240–1253.
  Foster, Kelly Gaither, Andrew Grimshaw, Victor Ha-
  zlewood, Scott Lathrop, Dave Lifka, Gregory D Pe-       Ron J. Weiss, Jan Chorowski, Navdeep Jaitly, Yonghui
  terson, et al. 2014. XSEDE: accelerating scientific       Wu, and Zhifeng Chen. 2017.          Sequence-to-
  discovery. Computing in science & engineering,            sequence models can directly translate foreign
  16(5):62–74.                                              speech. In Proc. Interspeech 2017, pages 2625–
                                                            2629.
Zhaopeng Tu, Yang Liu, Lifeng Shang, Xiaohua Liu,
                                                          Sam Wiseman and Alexander M. Rush. 2016.
  and Hang Li. 2017. Neural machine translation with
                                                            Sequence-to-sequence learning as beam-search op-
  reconstruction. In Proceedings of the Thirty-First
                                                            timization. In Proceedings of the 2016 Conference
  AAAI Conference on Artificial Intelligence, 2017,
                                                            on Empirical Methods in Natural Language Process-
  pages 3097–3103. AAAI Press.
                                                            ing, pages 1296–1306, Austin, Texas. Association
Evelyne Tzoukermann and Corey Miller. 2018. Evalu-          for Computational Linguistics.
  ating automatic speech recognition in translation. In
                                                          Yilin Yang, Liang Huang, and Mingbo Ma. 2018.
  Proceedings of the 13th Conference of the Associa-
                                                             Breaking the beam search curse: A study of (re-
  tion for Machine Translation in the Americas (Vol-
                                                             )scoring methods and stopping criteria for neural
  ume 2: User Track), pages 294–302, Boston, MA.
                                                             machine translation. In Proceedings of the 2018
  Association for Machine Translation in the Ameri-
                                                            Conference on Empirical Methods in Natural Lan-
  cas.
                                                             guage Processing, pages 3054–3059, Brussels, Bel-
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob             gium. Association for Computational Linguistics.
  Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz           Chengqi Zhao, Mingxuan Wang, and Lei Li. 2020.
  Kaiser, and Illia Polosukhin. 2017. Attention is all      NeurST: Neural speech translation toolkit. arXiv
  you need. In Advances in neural information pro-          preprint arXiv:2012.10018.
  cessing systems, pages 5998–6008.
                                                          Renjie Zheng, Junkun Chen, Mingbo Ma, and Liang
Ashwin K. Vijayakumar, Michael Cogswell, Ram-               Huang. 2021. Fused acoustic and text encoding for
  prasaath R. Selvaraju, Qing Sun, Stefan Lee, David J.     multimodal bilingual pretraining and speech transla-
  Crandall, and Dhruv Batra. 2018. Diverse beam             tion. arXiv preprint arXiv:2102.05766.
  search for improved description of complex scenes.
You can also read