{ "paper_id": "2022", "header": { "generated_with": "S2ORC 1.0.0", "date_generated": "2023-01-19T11:41:38.117026Z" }, "title": "Finding Sub-task Structure with Natural Language Instruction", "authors": [ { "first": "Ryokan", "middle": [], "last": "Ri", "suffix": "", "affiliation": { "laboratory": "", "institution": "The University of Tokyo", "location": {} }, "email": "" } ], "year": "", "venue": null, "identifiers": {}, "abstract": "When mapping a natural language instruction to a sequence of actions, it is often useful to identify sub-tasks in the instruction. Such subtask segmentation, however, is not necessarily provided in the training data. We present the A2LCTC (Action-to-Language Connectionist Temporal Classification) algorithm to automatically discover a sub-task segmentation of an action sequence. A2LCTC does not require annotations of correct sub-task segments and learns to find them from pairs of instruction and action sequence in a weakly-supervised manner. We experiment with the ALFRED dataset and show that A2LCTC accurately finds the sub-task structures. With the discovered subtasks segments, we also train agents that work on the downstream task and empirically show that our algorithm improves the performance.", "pdf_parse": { "paper_id": "2022", "_pdf_hash": "", "abstract": [ { "text": "When mapping a natural language instruction to a sequence of actions, it is often useful to identify sub-tasks in the instruction. Such subtask segmentation, however, is not necessarily provided in the training data. We present the A2LCTC (Action-to-Language Connectionist Temporal Classification) algorithm to automatically discover a sub-task segmentation of an action sequence. A2LCTC does not require annotations of correct sub-task segments and learns to find them from pairs of instruction and action sequence in a weakly-supervised manner. We experiment with the ALFRED dataset and show that A2LCTC accurately finds the sub-task structures. With the discovered subtasks segments, we also train agents that work on the downstream task and empirically show that our algorithm improves the performance.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Abstract", "sec_num": null } ], "body_text": [ { "text": "Building computational agents that execute actions given natural language instruction has a great deal of potential in real-world applications. One common approach is to cast the problem as mapping from instruction text into action sequence, and train an agent with supervised learning (Chen and Mooney, 2011; Mei et al., 2016) . A challenge on effective machine learning stems from a long horizon of the tasks. Typical navigation tasks often involve more than a paragraph of instructions (Chen and Mooney, 2011; Misra et al., 2017; Shridhar et al., 2020) . In such cases, many existing approaches exploit the task hierarchy, e.g., decompose one episode of the task into sub-tasks and treat them as separate training instances (Mei et al., 2016) , incorporate the hierarchical information into the model (Zhu et al., 2020) , or aid the learning process with progress monitoring (Ma et al., 2019) .", "cite_spans": [ { "start": 286, "end": 309, "text": "(Chen and Mooney, 2011;", "ref_id": "BIBREF1" }, { "start": 310, "end": 327, "text": "Mei et al., 2016)", "ref_id": "BIBREF9" }, { "start": 489, "end": 512, "text": "(Chen and Mooney, 2011;", "ref_id": "BIBREF1" }, { "start": 513, "end": 532, "text": "Misra et al., 2017;", "ref_id": "BIBREF11" }, { "start": 533, "end": 555, "text": "Shridhar et al., 2020)", "ref_id": "BIBREF14" }, { "start": 727, "end": 745, "text": "(Mei et al., 2016)", "ref_id": "BIBREF9" }, { "start": 804, "end": 822, "text": "(Zhu et al., 2020)", "ref_id": "BIBREF17" }, { "start": 878, "end": 895, "text": "(Ma et al., 2019)", "ref_id": "BIBREF8" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "However, annotations for such a decomposition are not necessarily available in training data. In * Work done as an intern at IBM Research. this case, previous work has attempted to perform sub-task segmentation by augmenting data through crowdsourcing (Hong et al., 2020) , or developing a heuristic algorithm (Zhu et al., 2020) .", "cite_spans": [ { "start": 252, "end": 271, "text": "(Hong et al., 2020)", "ref_id": "BIBREF6" }, { "start": 310, "end": 328, "text": "(Zhu et al., 2020)", "ref_id": "BIBREF17" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "In this paper, we present A2LCTC (Action-to-Language Connectionist Temporal Classification), an unsupervised algorithm that automatically discovers a sub-task segmentation of an action sequence. Given pairs of a natural language instruction and action sequence, A2LCTC maps each action to a fine-grained instruction ( Figure 1) .", "cite_spans": [], "ref_spans": [ { "start": 318, "end": 327, "text": "Figure 1)", "ref_id": "FIGREF0" } ], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "We formulate the problem of sub-task segmentation as classification of each action into a finegrained instruction. Inspired by the connectionist temporal classification algorithm (Graves et al., 2006) , we consider an objective function that maximizes the log-likelihood of the coarsely aligned data. This formulation allows us to learn the classification in a weakly-supervised manner without any need of the ground truth mapping.", "cite_spans": [ { "start": 179, "end": 200, "text": "(Graves et al., 2006)", "ref_id": "BIBREF4" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "We experiment with the ALFRED dataset which involves navigating a robot to perform household task (Shridhar et al., 2020) . A2LCTC successfully discovers the sub-task information in the unsegmented training data ( \u00a73), which are shown to be useful for the downstream task ( \u00a74).", "cite_spans": [ { "start": 98, "end": 121, "text": "(Shridhar et al., 2020)", "ref_id": "BIBREF14" } ], "ref_spans": [], "eq_spans": [], "section": "Introduction", "sec_num": "1" }, { "text": "Given a training instance of a navigation task which consists of language instruction X and action sequence A = [a 1 , ..., a T ], we aim at finding a decomposition of the instance (X", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "2 Learning Sub-task Segmentation", "sec_num": "1" }, { "text": ", A) = [(X 1 , A 1 ), ..., (X L , A L )]", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "2 Learning Sub-task Segmentation", "sec_num": "1" }, { "text": ", which is semantically plausible and useful for learning the task.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "2 Learning Sub-task Segmentation", "sec_num": "1" }, { "text": "Our approach starts with dividing the instruction into fine-grained segments. In the experiment, the segment X i corresponds to a verbal phrase (e.g., \"go to the desk\") extracted from the instruction using a simple rule-based algorithm (Appendix A.1).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "2 Learning Sub-task Segmentation", "sec_num": "1" }, { "text": "We formulate the decomposition task as classification of actions into one of the fine-grained instructions: our algorithm predicts a mapping", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "\u03c0 = [\u03c0 1 , ..., \u03c0 T ] where \u03c0 t \u2208 [1, L].", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "We assume that the alignment is monotonic, i.e., the actions and instructions are both arranged in a chronological order, which typically holds for stepby-step instruction text 1 . This formulation allows us to develop an unsupervised learning method to effectively solve the task without ground-truth label mappings. We introduce A2LCTC (Action-to-Language Connectionist Temporal Classification), which is based on the CTC algorithm originally used for speech recognition (Graves et al., 2006) or action labeling in video (Huang et al., 2016) .", "cite_spans": [ { "start": 473, "end": 494, "text": "(Graves et al., 2006)", "ref_id": "BIBREF4" }, { "start": 523, "end": 543, "text": "(Huang et al., 2016)", "ref_id": "BIBREF7" } ], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "Our model attempts to maximize the following likelihood for each training instance:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "{\u03c0|B(\u03c0)=[1,2,..,L]} P (\u03c0|X, A)", "eq_num": "(1)" } ], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "where B is the operator to remove repeated labels, e.g.,,", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "B([1, 1, 2, 2, 2, 3]) = [1, 2, 3].", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "That is, the objective is the sum of the probabilities over all the possible assignments under the monotonic constraint. Under the conditional independence assumption, we further decompose the likelihood into P (\u03c0|X, A) = T t P (\u03c0 t |X, A). With this decomposition, we employ the forward algorithm (Stratonovich, 1965) to efficiently calculate the sum over all possible paths.", "cite_spans": [ { "start": 298, "end": 318, "text": "(Stratonovich, 1965)", "ref_id": "BIBREF16" } ], "ref_spans": [], "eq_spans": [], "section": "Formulating the Task as Temporal Classification", "sec_num": "2.1" }, { "text": "We model the probability computation using neural networks. In our approach, each fine-grained instruction and action are represented as feature vectors. We then define the probability P (\u03c0 t |X, A) as the softmax of the dot product of the feature vectors of the action a t and the instruction [x 1 , ..., x L ]:", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Modeling with Neural Network", "sec_num": "2.2" }, { "text": "P (\u03c0 t = i|X, A) = exp(a t \u2022 x i ) L j=1 exp(a t \u2022 x j )", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Modeling with Neural Network", "sec_num": "2.2" }, { "text": ". 2In our implementation, a fine-grained instruction X i is tokenized into words and x i is computed as the average of the word embeddings followed by a linear layer. The action feature vectors are computed by feeding action embeddings [\u00e2 1 , ...,\u00e2 T ] into a one-layer bidirectional LSTM (Hochreiter and Schmidhuber, 1997) to capture the semantics of actions in the context.", "cite_spans": [ { "start": 289, "end": 323, "text": "(Hochreiter and Schmidhuber, 1997)", "ref_id": "BIBREF5" } ], "ref_spans": [], "eq_spans": [], "section": "Modeling with Neural Network", "sec_num": "2.2" }, { "text": "At inference time, the model induces the best mapping \u03c0 that maximize P (\u03c0|X, A) with the viterbi algorithm to form decomposed instances (X i , A i ).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Modeling with Neural Network", "sec_num": "2.2" }, { "text": "Optimizing the aforementioned objective function with neural networks turns out to be highly unstable. Therefore, we further employ the following techniques to better guide the training process. Length-based curriculum learning. The training objective (1) is defined as the sum of the probabilities over all possible assignments, whose size grows at the speed of O(L T ). When L and T are large, too many degenerated assignments take up a large part of the probability mass, which hinders the training especially at the beginning of training.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Stabilization of Unsupervised Learning", "sec_num": "2.3" }, { "text": "To address this problem, we adopt the curriculum learning framework (Bengio et al., 2009) . At the beginning of training, we restrict the training instances with a shorter action sequence and gradually expose the model to longer instances (Spitkovsky et al., 2010) . Details are shown in Appendix A.4. Initializing with pre-trained word embeddings. We initialize the word embeddings w with the Glove embeddings (Pennington et al., 2014) to inject a model with the prior knowledge of words. Initializing with pre-trained action embeddings. We also initialize the action embeddings\u00e2 with pretrained embeddings. We obtain action embeddings from the action sequences in the training data with Skip-gram (Mikolov et al., 2013) .", "cite_spans": [ { "start": 68, "end": 89, "text": "(Bengio et al., 2009)", "ref_id": "BIBREF0" }, { "start": 239, "end": 264, "text": "(Spitkovsky et al., 2010)", "ref_id": "BIBREF15" }, { "start": 411, "end": 436, "text": "(Pennington et al., 2014)", "ref_id": "BIBREF12" }, { "start": 699, "end": 721, "text": "(Mikolov et al., 2013)", "ref_id": "BIBREF10" } ], "ref_spans": [], "eq_spans": [], "section": "Stabilization of Unsupervised Learning", "sec_num": "2.3" }, { "text": "We test A2LCTC with the ALFRED dataset (Shridhar et al., 2020) 2 , where the agent performs household tasks in a simulated indoor environment given English instruction. We choose this dataset because (1) the data involves relatively long sequence of actions (over 100); (2) it offers ground-truth sub-task segmentation data which allows the evaluation of our algorithm. Although the ALFRED task itself involves understanding visual inputs, which is currently beyond the scope of A2LCTC, the action set defined in the dataset is semantically rich enough for A2LCTC to solve the segmentation problem.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experiments", "sec_num": "3" }, { "text": "Training and Evaluation Data. The dataset contains expert demonstrations, which are split into the training set and two validation sets: valid_seen and valid_unseen. In valid_seen, the ALFRED tasks are given in the same set of environments as the training data, while unseen environments are used for valid_unseen. We use valid_seen as the validation set during training and report the results evaluated with both of them.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setups", "sec_num": "3.1" }, { "text": "The task defines 12 types of actions: five for navigation (e.g., LookUp, MoveForward, RotateRight) and seven for interaction (e.g., Pickup, Slice). To simplify training and evaluation, we preprocess data by merging the same consecutive navigation actions into a single action (see Appendix B for detailed statistics on the data). Evaluation Metrics. Each training instance in the ALFRED dataset contains ground-truth sub-goal segments for instructions and actions. We use them to evaluate the learned sub-task segments.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setups", "sec_num": "3.1" }, { "text": "Note that our algorithm operates on finer-grained instructions segmented into verbal phrases, while the ground-truth segments are coarser; some sentences contain a couple of verbal phrases (e.g., \"Rinse the sponge out in the sink, and pick it up again\"). In our evaluation, we first merge the finegrained instructions into their corresponding subgoal instruction, together with the mapped actions, and then compare the overlap with the gold subgoal segments of actions.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setups", "sec_num": "3.1" }, { "text": "We report the sub-goal exact match (EM) score defined as the percentage of the sub-goal segments perfectly reconstructed. We also report the subgoal F1 score defined as the macro average of the 2 https://askforalfred.com/, MIT License F1 scores calculated by taking the overlap between the predicted and ground-truth segment.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Experimental Setups", "sec_num": "3.1" }, { "text": "Under our task formulation, A2LCTC offers an advantage that it can take into account the temporal constraint and the semantics of instructions. However, many existing unsupervised sequence alignment algorithms (e.g., IBM models) operate only between discrete symbols and are not directly applicable in this situation where we need to align actions to text (or a bag of words in our model). Thus, we compare A2LCTC with two baselines that do not consider the textual information. Uniform. The uniform baseline assigns an equal number of actions to each of the fine-grained instructions. Byte-pair Encoding. The byte-pair encoding (BPE) baseline is based on a data compression algorithm that finds repeated patterns in the data and merges them into chunks (Gage, 1994; Sennrich et al., 2016) . Concretely, we repeat the following two steps until each action sequence is split into the number of the corresponding fine-grained instructions: (1) count bigrams in the action sequences;", "cite_spans": [ { "start": 754, "end": 766, "text": "(Gage, 1994;", "ref_id": "BIBREF2" }, { "start": 767, "end": 789, "text": "Sennrich et al., 2016)", "ref_id": "BIBREF13" } ], "ref_spans": [], "eq_spans": [], "section": "Baselines", "sec_num": "3.2" }, { "text": "(2) merge the most frequent bigrams. Table 1 shows that A2LCTC significantly outperforms the baselines, which indicates that our model successfully leverages textual information and learns meaningful alignments between the finegrained instructions and actions.", "cite_spans": [], "ref_spans": [ { "start": 37, "end": 44, "text": "Table 1", "ref_id": null } ], "eq_spans": [], "section": "Baselines", "sec_num": "3.2" }, { "text": "Although the BPE baseline does not use textual information, it exhibits reasonable F1 scores (61.9 points in valid_unseen and much higher EM scores than UNIFORM (27.1 vs. 9.3 points). This reflects the characteristics of the ALFRED dataset. As each episode in the dataset is generated from specific templates, the actions follow specific patterns, which enable the BPE to learn correct segmentation to some extent.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Results", "sec_num": "3.3" }, { "text": "In A2LCTC, we utilize several techniques to stabilize the training process. Ablating pre-training language or action embeddings still obtains mean values comparable to the full A2LCTC but yields much larger standard deviations. This indicates that these two stabilization methods are also beneficial for A2LCTC.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Ablation Study", "sec_num": "3.4" }, { "text": "We evaluate the effectiveness of sub-task segments induced by A2LCTC on the downstream ALFRED task. Models. Our baseline agent (BASELINE) is based on the CNN-LSTM sequence-to-sequence architecture in Shridhar et al. (2020) , which takes the whole instruction and current state as input and then predicts an action at each time step. Unless specified, we use the same hyperparameters as the original implementation.", "cite_spans": [ { "start": 200, "end": 222, "text": "Shridhar et al. (2020)", "ref_id": "BIBREF14" } ], "ref_spans": [], "eq_spans": [], "section": "Evaluation with the Downstream Task", "sec_num": "4" }, { "text": "To incorporate the sub-task information, we extend the baseline with the progress monitoring module (Ma et al., 2019) . We use two progress monitoring schemes from Shridhar et al. (2020) , which estimate the current time step and the completed sub-tasks. Specifically, the modules are trained to predict the proportion of elapsed steps or completed sub-tasks to the total numbers.", "cite_spans": [ { "start": 100, "end": 117, "text": "(Ma et al., 2019)", "ref_id": "BIBREF8" }, { "start": 164, "end": 186, "text": "Shridhar et al. (2020)", "ref_id": "BIBREF14" } ], "ref_spans": [], "eq_spans": [], "section": "Evaluation with the Downstream Task", "sec_num": "4" }, { "text": "We evaluate the segmentation of A2LCTC as well as the two baseline methods (UNIFORM and BPE). Those algorithms are applied on the training data and the agents are trained with the progress monitoring according to the segmentation. Metric. We evaluate the agents with the subgoal sequence accuracy. The agent predicts the next valid_seen valid_unseen BASELINE 57.6 \u00b1 2.0 29.6 \u00b1 1.5 UNIFORM 63.6 \u00b1 2.5 38.8 \u00b1 1.4 BPE 66.0 \u00b1 2.0 39.1 \u00b1 0.4 A2LCTC 69.7 \u00b1 1.4 41.4 \u00b1 0.8 Table 3 : Performance on the ALFRED task measured by the subgoal sequence accuracy. The values show the mean and standard deviation of 5 runs. action given the history from an expert trajectory. The metric measures how many subgoal chunks of actions, which is defined by the ground-truth segmentation of the dataset, are successfully predicted in the evaluation data. Note that this metric simplifies the original task in that it ignores the object interactions and focuses on action prediction 3 . Results. Table 3 summarizes the result. On both valid_seen and valid_unseen splits, the agents trained with fine-grained instruction (UNIFORM, BPE and A2LCTC) significantly outperform BASELINE. The fact that UNIFORM achieves improvement indicates that keeping track of detailed progress is helpful even if it inaccurately performs the fine-grained task segmentation (see Section 3.3). A2LCTC performs best because of the better accuracy of the segmentation than the others. This demonstrates that A2LCTC successfully provides more informative instructions for the agent in solving the downstream task.", "cite_spans": [], "ref_spans": [ { "start": 466, "end": 473, "text": "Table 3", "ref_id": null }, { "start": 974, "end": 981, "text": "Table 3", "ref_id": null } ], "eq_spans": [], "section": "Evaluation with the Downstream Task", "sec_num": "4" }, { "text": "We presented A2LCTC, which finds a hierarchical structure of an action sequence by mapping each action to fine-grained natural language instructions without ground-truth mapping data. We demonstrated that A2LCTC successfully learns meaningful segments and training the ALFRED agents with these segments leads to improved performance.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion", "sec_num": "5" }, { "text": "A2LCTC currently relies only on semantic correspondence between actions and text. Applying A2LCTC to the tasks with low-level actions is an important extension, e.g., actions specifying the direction to move. Furthermore, the instruction may often describe the visual input whose information is not encoded in the actions. Another important future direction is to incorporate visual or additional information to tackle a broader range of domains.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Conclusion", "sec_num": "5" }, { "text": "We split a whole instruction into fine-grained instructions each of which is a verbal phrase. Our current implementation employs a simple rule-based algorithm, which segments a whole instruction at periods, commas, phrases such as and or then. For example, \"Turn around and go back to the table\" will be segmented into [\"Turn around\", \"go back to the table\"], while we also write few rules to handle erroneous splits such as conjunctions between nouns ([\"take the apple\", \"(and) banana\"]), commas before a prepositional phrase ([\"Put the bowl\", \"(,) on the coffee table to the left of the statue\"]), or verbal phrases that cannot be instruction by itself ([\"go to the left\", \"(and) face the bathtub\"]).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A.1 Instruction Decomposition", "sec_num": null }, { "text": "The feature vector of a fine-grained instruction X l is simply modeled by taking the average of the word embeddings [w 1 , ..., w N ] followed by a linear layer (the results from different encoding strategies are shown in Appendix C).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A.2 Neural Network Architecture Instruction Feature Vectors", "sec_num": null }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "x = tanh(Linear( 1 N N i w i )).", "eq_num": "(3)" } ], "section": "A.2 Neural Network Architecture Instruction Feature Vectors", "sec_num": null }, { "text": "In our experiment, the dimension of word embeddings, the input and output size of the linear layer is all set to 50. The total number of parameters of A2LCTC is about 44K. The training takes approximately one hour with a single GPU.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A.2 Neural Network Architecture Instruction Feature Vectors", "sec_num": null }, { "text": "The action feature vectors are computed through feeding embeddings for primitive actions [\u00e2 1 , ...,\u00e2 T ] to a one-layer bidirectional LSTM.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Action Feature Vectors", "sec_num": null }, { "text": "EQUATION", "cite_spans": [], "ref_spans": [], "eq_spans": [ { "start": 0, "end": 8, "text": "EQUATION", "ref_id": "EQREF", "raw_str": "a 1 , ..., a T = LSTM(\u00e2 1 , ...,\u00e2 T )", "eq_num": "(4)" } ], "section": "Action Feature Vectors", "sec_num": null }, { "text": "where [\u00e2 1 , ...,\u00e2 T ] are embeddings for primitive actions. The size of the action embeddings and the hidden size of the LSTM are set to 50. The outputs of LSTM in the forward and backward directions are summed to merged into one single feature vector.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "Action Feature Vectors", "sec_num": null }, { "text": "optimizer Adam learning rate 0.001 batch size 128 validation metric the sub-goal F1 score patience 5", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A.3 Hyperparameters for training", "sec_num": null }, { "text": "Length-based curriculum learning", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A.4 Stabilization of Unsupervised Learning", "sec_num": null }, { "text": "We set the maximum length of training instances for each training epoch according to a schedule. In our experiment, the maximum length starts at 20, and linearly increases to 60 with 30 steps.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "A.4 Stabilization of Unsupervised Learning", "sec_num": null }, { "text": "We use the gensim library 4 to train action embeddings. We use the Skip-gram algorithm and the hyperparameters are shown in Table 4 embedding size 50 window size 1 # of iterations 15 # of negative samples 5 ", "cite_spans": [], "ref_spans": [ { "start": 124, "end": 131, "text": "Table 4", "ref_id": "TABREF2" } ], "eq_spans": [], "section": "Initializing with pre-trained action embeddings", "sec_num": null }, { "text": "Our experiments are based on the expert demonstration data in the ALFRED dataset (Table 5) .", "cite_spans": [], "ref_spans": [ { "start": 81, "end": 90, "text": "(Table 5)", "ref_id": null } ], "eq_spans": [], "section": "B Data Statistics", "sec_num": null }, { "text": "training data valid_seen valid_unseen 21,023 820 821 Table 5 : The number of expert demonstrations in the ALFRED dataset.", "cite_spans": [], "ref_spans": [ { "start": 53, "end": 60, "text": "Table 5", "ref_id": null } ], "eq_spans": [], "section": "B Data Statistics", "sec_num": null }, { "text": "After the instruction decomposition, the instructions contain 10 fine-grained instructions on average. The entire distribution is shown in Figure2. To simplify training and evaluation, we preprocess data by merging the same consecutive actions into one single action. This results in the average sequence length of 25. The entire distribution before and after the merge preprocessing is shown in Figure3 and 4. ", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "B Data Statistics", "sec_num": null }, { "text": "Here we compare different strategies for encoding instruction segments. Besides the mean pooling of word embeddings followed by a linear layer in A2LCTC (MEAN), we also tried the summation of word embeddings (SUM), the mean pooling of one-layer Bi-LSTM outputs LSTM. The hidden size of LSTM is set to 50, which is the same as the word embeddings, and the vectors of the forward and backward directions were summed to form the output vectors. The result is shown in Table 6 : Performance for sub-task segmentation. The value of A2LCTC is the best among 10 runs with different random seeds.", "cite_spans": [], "ref_spans": [ { "start": 465, "end": 472, "text": "Table 6", "ref_id": "TABREF3" } ], "eq_spans": [], "section": "C Additional Results for the Segmentation Task", "sec_num": null }, { "text": "We find that MEAN gives the most stable result. LSTM exhibits the worst performance, indicating that it is hard to optimize the unsupervised objective in A2LCTC with an overly complex architecture.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "C Additional Results for the Segmentation Task", "sec_num": null }, { "text": "In case that the monotonic assumption does not hold, we could reorder instructions with some sentence-ordering algorithm such as(Ghosal et al., 2021) as preprocessing.", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "We find the original navigation task is too difficult for the baseline model: the success rate is very low with high variance, which prevents meaningful comparison among the variants of the model (Appendix D).", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null }, { "text": "https://radimrehurek.com/gensim/", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "", "sec_num": null } ], "back_matter": [ { "text": "In section 4, we compared the models trained with automatically generated fine-grained sub-task segments. Here we provide the results from the model trained with the ground-truth sub-goal segments (SUBGOAL) in Table 7 . Note that the granularity of SUBGOAL is coarser than the other models.valid_seen valid_unseen BASELINE 57.6 \u00b1 2.0 29.6 \u00b1 1.5 SUBGOAL 59.5 \u00b1 2.1 31.4 \u00b1 1.4 UNIFORM 63.6 \u00b1 2.5 38.8 \u00b1 1.4 BPE 66.0 \u00b1 2.0 39.1 \u00b1 0.4 A2LCTC 69.7 \u00b1 1.4 41.4 \u00b1 0.8 Table 7 : Performance on the ALFRED task measured by the subgoal sequence accuracy. The values show the mean and standard deviation of 5 runs.SUBGOAL provides better results than BASELINE, which demonstrates the benefit of the ground-truth sub-gold segmentation in the dataset. However, the improvement is limited compared to the UNIFORM segmentation, which segments an action sequence into the chunks of the same size. The model benefits from inaccurate but finer-grained segmentation more than accurate but coarse segmentation.We hypothesize that this reflects the characteristics of the dataset. The ALFRED dataset is created by generating expert trajectories from task templates. As a result, the type of actions are somewhat correlated with the time step within an episode. For example, navigation actions such as MoveForward are more likely to be executed at the beginning of the episode, whereas interactive actions such as PutObject are at the end. Adding finer-grained progress monitoring supervision at training time can help the agent learn the correlation between time steps and actions better than coarser progress monitoring.", "cite_spans": [], "ref_spans": [ { "start": 210, "end": 217, "text": "Table 7", "ref_id": null }, { "start": 460, "end": 467, "text": "Table 7", "ref_id": null } ], "eq_spans": [], "section": "annex", "sec_num": null }, { "text": "In our preliminary experiments, we find the original navigation task is too difficult for the baseline model: the success rate is very low with high variance, which prevents meaningful comparison among the variants of the model. The success rate (SC) and goal condition success rate (GC) are provided on Table 8 . With multiple runs, we did not observe any significant difference (p > 0.05 in the Welch's t-test) among the models.", "cite_spans": [], "ref_spans": [ { "start": 304, "end": 311, "text": "Table 8", "ref_id": null } ], "eq_spans": [], "section": "D.2 Success Rates of the Downstream Task", "sec_num": null }, { "text": "GC valid_seen valid_unseen valid_seen valid_unseen", "cite_spans": [], "ref_spans": [], "eq_spans": [], "section": "SC", "sec_num": null }, { "text": "2.4 \u00b1 0.9 0.0 \u00b1 0.0 9.5 \u00b1 0.5 6.7 \u00b1 0.3 SUBGOAL 2.6 \u00b1 0.5 0.0 \u00b1 0.0 8.9 \u00b1 0.8 6.6 \u00b1 0.4 A2LCTC 2.3 \u00b1 0.7 0.0 \u00b1 0.0 9.4 \u00b1 0.8 6.7 \u00b1 0.3 Table 8 : Performance on the ALFRED task measured by the task success rate (SC) and goal condition success rate (GC). The values show the mean and standard deviation of 5 runs.", "cite_spans": [], "ref_spans": [ { "start": 135, "end": 142, "text": "Table 8", "ref_id": null } ], "eq_spans": [], "section": "Baseline", "sec_num": null } ], "bib_entries": { "BIBREF0": { "ref_id": "b0", "title": "Curriculum learning", "authors": [ { "first": "Yoshua", "middle": [], "last": "Bengio", "suffix": "" }, { "first": "J\u00e9r\u00f4me", "middle": [], "last": "Louradour", "suffix": "" }, { "first": "Ronan", "middle": [], "last": "Collobert", "suffix": "" }, { "first": "Jason", "middle": [], "last": "Weston", "suffix": "" } ], "year": 2009, "venue": "ACM International Conference Proceeding Series", "volume": "382", "issue": "", "pages": "41--48", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yoshua Bengio, J\u00e9r\u00f4me Louradour, Ronan Collobert, and Jason Weston. 2009. Curriculum learning. In ICML, volume 382 of ACM International Conference Proceeding Series, pages 41-48. ACM.", "links": null }, "BIBREF1": { "ref_id": "b1", "title": "Learning to interpret natural language navigation instructions from observations", "authors": [ { "first": "L", "middle": [], "last": "David", "suffix": "" }, { "first": "Raymond", "middle": [ "J" ], "last": "Chen", "suffix": "" }, { "first": "", "middle": [], "last": "Mooney", "suffix": "" } ], "year": 2011, "venue": "AAAI", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "David L. Chen and Raymond J. Mooney. 2011. Learn- ing to interpret natural language navigation instruc- tions from observations. In AAAI.", "links": null }, "BIBREF2": { "ref_id": "b2", "title": "A new algorithm for data compression", "authors": [ { "first": "P", "middle": [], "last": "Gage", "suffix": "" } ], "year": 1994, "venue": "The C Users Journal archive", "volume": "12", "issue": "", "pages": "23--38", "other_ids": {}, "num": null, "urls": [], "raw_text": "P. Gage. 1994. A new algorithm for data compression. The C Users Journal archive, 12:23-38.", "links": null }, "BIBREF3": { "ref_id": "b3", "title": "Stack: Sentence ordering with temporal commonsense knowledge", "authors": [ { "first": "Deepanway", "middle": [], "last": "Ghosal", "suffix": "" }, { "first": "Navonil", "middle": [], "last": "Majumder", "suffix": "" }, { "first": "Rada", "middle": [], "last": "Mihalcea", "suffix": "" }, { "first": "Soujanya", "middle": [], "last": "Poria", "suffix": "" } ], "year": 2021, "venue": "Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Deepanway Ghosal, Navonil Majumder, Rada Mihalcea, and Soujanya Poria. 2021. Stack: Sentence ordering with temporal commonsense knowledge. In Proceed- ings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP).", "links": null }, "BIBREF4": { "ref_id": "b4", "title": "Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks", "authors": [ { "first": "Alex", "middle": [], "last": "Graves", "suffix": "" }, { "first": "Santiago", "middle": [], "last": "Fern\u00e1ndez", "suffix": "" }, { "first": "Faustino", "middle": [ "J" ], "last": "Gomez", "suffix": "" }, { "first": "J\u00fcrgen", "middle": [], "last": "Schmidhuber", "suffix": "" } ], "year": 2006, "venue": "ACM International Conference Proceeding Series", "volume": "", "issue": "", "pages": "369--376", "other_ids": {}, "num": null, "urls": [], "raw_text": "Alex Graves, Santiago Fern\u00e1ndez, Faustino J. Gomez, and J\u00fcrgen Schmidhuber. 2006. Connectionist tem- poral classification: labelling unsegmented sequence data with recurrent neural networks. In ICML, vol- ume 148 of ACM International Conference Proceed- ing Series, pages 369-376. ACM.", "links": null }, "BIBREF5": { "ref_id": "b5", "title": "Long short-term memory", "authors": [ { "first": "Sepp", "middle": [], "last": "Hochreiter", "suffix": "" }, { "first": "J\u00fcrgen", "middle": [], "last": "Schmidhuber", "suffix": "" } ], "year": 1997, "venue": "Neural Computation", "volume": "9", "issue": "8", "pages": "1735--1780", "other_ids": {}, "num": null, "urls": [], "raw_text": "Sepp Hochreiter and J\u00fcrgen Schmidhuber. 1997. Long short-term memory. Neural Computation, 9(8):1735- 1780.", "links": null }, "BIBREF6": { "ref_id": "b6", "title": "Sub-instruction aware vision-andlanguage navigation", "authors": [ { "first": "Yicong", "middle": [], "last": "Hong", "suffix": "" }, { "first": "Cristian", "middle": [], "last": "Rodriguez", "suffix": "" }, { "first": "Qi", "middle": [], "last": "Wu", "suffix": "" }, { "first": "Stephen", "middle": [], "last": "Gould", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", "volume": "", "issue": "", "pages": "3360--3376", "other_ids": {}, "num": null, "urls": [], "raw_text": "Yicong Hong, Cristian Rodriguez, Qi Wu, and Stephen Gould. 2020. Sub-instruction aware vision-and- language navigation. In Proceedings of the 2020 Conference on Empirical Methods in Natural Lan- guage Processing (EMNLP), pages 3360-3376, On- line. Association for Computational Linguistics.", "links": null }, "BIBREF7": { "ref_id": "b7", "title": "Connectionist temporal modeling for weakly supervised action labeling", "authors": [ { "first": "De-An", "middle": [], "last": "Huang", "suffix": "" }, { "first": "Li", "middle": [], "last": "Fei-Fei", "suffix": "" }, { "first": "Juan", "middle": [ "Carlos" ], "last": "Niebles", "suffix": "" } ], "year": 2016, "venue": "Computer Vision -ECCV 2016 -14th European Conference", "volume": "9908", "issue": "", "pages": "137--153", "other_ids": { "DOI": [ "10.1007/978-3-319-46493-0_9" ] }, "num": null, "urls": [], "raw_text": "De-An Huang, Li Fei-Fei, and Juan Carlos Niebles. 2016. Connectionist temporal modeling for weakly supervised action labeling. In Computer Vision - ECCV 2016 -14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part IV, volume 9908 of Lecture Notes in Computer Science, pages 137-153. Springer.", "links": null }, "BIBREF8": { "ref_id": "b8", "title": "Selfmonitoring navigation agent via auxiliary progress estimation", "authors": [ { "first": "Chih-Yao", "middle": [], "last": "Ma", "suffix": "" }, { "first": "Jiasen", "middle": [], "last": "Lu", "suffix": "" }, { "first": "Zuxuan", "middle": [], "last": "Wu", "suffix": "" }, { "first": "G", "middle": [], "last": "Al-Regib", "suffix": "" }, { "first": "Z", "middle": [], "last": "Kira", "suffix": "" }, { "first": "R", "middle": [], "last": "Socher", "suffix": "" }, { "first": "Caiming", "middle": [], "last": "Xiong", "suffix": "" } ], "year": 2019, "venue": "ArXiv", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Chih-Yao Ma, Jiasen Lu, Zuxuan Wu, G. Al-Regib, Z. Kira, R. Socher, and Caiming Xiong. 2019. Self- monitoring navigation agent via auxiliary progress estimation. ArXiv, abs/1901.03035.", "links": null }, "BIBREF9": { "ref_id": "b9", "title": "Listen, attend, and walk: Neural mapping of navigational instructions to action sequences", "authors": [ { "first": "Hongyuan", "middle": [], "last": "Mei", "suffix": "" }, { "first": "Mohit", "middle": [], "last": "Bansal", "suffix": "" }, { "first": "Matthew", "middle": [ "R" ], "last": "Walter", "suffix": "" } ], "year": 2016, "venue": "AAAI", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Hongyuan Mei, Mohit Bansal, and Matthew R. Walter. 2016. Listen, attend, and walk: Neural mapping of navigational instructions to action sequences. In AAAI.", "links": null }, "BIBREF10": { "ref_id": "b10", "title": "Efficient estimation of word representations in vector space", "authors": [ { "first": "Tomas", "middle": [], "last": "Mikolov", "suffix": "" }, { "first": "Kai", "middle": [], "last": "Chen", "suffix": "" }, { "first": "Greg", "middle": [], "last": "Corrado", "suffix": "" }, { "first": "Jeffrey", "middle": [], "last": "Dean", "suffix": "" } ], "year": 2013, "venue": "CoRR", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Tomas Mikolov, Kai Chen, Greg Corrado, and Jeffrey Dean. 2013. Efficient estimation of word representa- tions in vector space. CoRR, abs/1301.3781.", "links": null }, "BIBREF11": { "ref_id": "b11", "title": "Mapping instructions and visual observations to actions with reinforcement learning", "authors": [ { "first": "Dipendra", "middle": [], "last": "Misra", "suffix": "" }, { "first": "John", "middle": [], "last": "Langford", "suffix": "" }, { "first": "Yoav", "middle": [], "last": "Artzi", "suffix": "" } ], "year": 2017, "venue": "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing", "volume": "", "issue": "", "pages": "1004--1015", "other_ids": { "DOI": [ "10.18653/v1/D17-1106" ] }, "num": null, "urls": [], "raw_text": "Dipendra Misra, John Langford, and Yoav Artzi. 2017. Mapping instructions and visual observations to ac- tions with reinforcement learning. In Proceedings of the 2017 Conference on Empirical Methods in Natu- ral Language Processing, pages 1004-1015, Copen- hagen, Denmark. Association for Computational Lin- guistics.", "links": null }, "BIBREF12": { "ref_id": "b12", "title": "GloVe: Global vectors for word representation", "authors": [ { "first": "Jeffrey", "middle": [], "last": "Pennington", "suffix": "" }, { "first": "Richard", "middle": [], "last": "Socher", "suffix": "" }, { "first": "Christopher", "middle": [], "last": "Manning", "suffix": "" } ], "year": 2014, "venue": "Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)", "volume": "", "issue": "", "pages": "1532--1543", "other_ids": { "DOI": [ "10.3115/v1/D14-1162" ] }, "num": null, "urls": [], "raw_text": "Jeffrey Pennington, Richard Socher, and Christopher Manning. 2014. GloVe: Global vectors for word representation. In Proceedings of the 2014 Confer- ence on Empirical Methods in Natural Language Pro- cessing (EMNLP), pages 1532-1543, Doha, Qatar. Association for Computational Linguistics.", "links": null }, "BIBREF13": { "ref_id": "b13", "title": "Neural machine translation of rare words with subword units", "authors": [ { "first": "Rico", "middle": [], "last": "Sennrich", "suffix": "" }, { "first": "Barry", "middle": [], "last": "Haddow", "suffix": "" }, { "first": "Alexandra", "middle": [], "last": "Birch", "suffix": "" } ], "year": 2016, "venue": "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics", "volume": "1", "issue": "", "pages": "1715--1725", "other_ids": { "DOI": [ "10.18653/v1/P16-1162" ] }, "num": null, "urls": [], "raw_text": "Rico Sennrich, Barry Haddow, and Alexandra Birch. 2016. Neural machine translation of rare words with subword units. In Proceedings of the 54th Annual Meeting of the Association for Computational Lin- guistics (Volume 1: Long Papers), pages 1715-1725, Berlin, Germany. Association for Computational Lin- guistics.", "links": null }, "BIBREF14": { "ref_id": "b14", "title": "Alfred: A benchmark for interpreting grounded instructions for everyday tasks", "authors": [ { "first": "Mohit", "middle": [], "last": "Shridhar", "suffix": "" }, { "first": "Jesse", "middle": [], "last": "Thomason", "suffix": "" }, { "first": "Daniel", "middle": [], "last": "Gordon", "suffix": "" }, { "first": "Yonatan", "middle": [], "last": "Bisk", "suffix": "" }, { "first": "Winson", "middle": [], "last": "Han", "suffix": "" }, { "first": "Roozbeh", "middle": [], "last": "Mottaghi", "suffix": "" }, { "first": "Luke", "middle": [], "last": "Zettlemoyer", "suffix": "" }, { "first": "Dieter", "middle": [], "last": "Fox", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)", "volume": "", "issue": "", "pages": "", "other_ids": {}, "num": null, "urls": [], "raw_text": "Mohit Shridhar, Jesse Thomason, Daniel Gordon, Yonatan Bisk, Winson Han, Roozbeh Mottaghi, Luke Zettlemoyer, and Dieter Fox. 2020. Alfred: A bench- mark for interpreting grounded instructions for ev- eryday tasks. In Proceedings of the IEEE/CVF Con- ference on Computer Vision and Pattern Recognition (CVPR).", "links": null }, "BIBREF15": { "ref_id": "b15", "title": "From baby steps to leapfrog: How \"less is more\" in unsupervised dependency parsing", "authors": [ { "first": "I", "middle": [], "last": "Valentin", "suffix": "" }, { "first": "Hiyan", "middle": [], "last": "Spitkovsky", "suffix": "" }, { "first": "Daniel", "middle": [], "last": "Alshawi", "suffix": "" }, { "first": "", "middle": [], "last": "Jurafsky", "suffix": "" } ], "year": 2010, "venue": "Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "751--759", "other_ids": {}, "num": null, "urls": [], "raw_text": "Valentin I. Spitkovsky, Hiyan Alshawi, and Daniel Ju- rafsky. 2010. From baby steps to leapfrog: How \"less is more\" in unsupervised dependency parsing. In Human Language Technologies: The 2010 An- nual Conference of the North American Chapter of the Association for Computational Linguistics, pages 751-759, Los Angeles, California. Association for Computational Linguistics.", "links": null }, "BIBREF16": { "ref_id": "b16", "title": "Conditional markov processes", "authors": [ { "first": "Ruslan", "middle": [], "last": "Leont", "suffix": "" }, { "first": "'", "middle": [], "last": "Stratonovich", "suffix": "" } ], "year": 1965, "venue": "Non-linear transformations of stochastic processes", "volume": "", "issue": "", "pages": "427--453", "other_ids": {}, "num": null, "urls": [], "raw_text": "Ruslan Leont'evich Stratonovich. 1965. Conditional markov processes. In Non-linear transformations of stochastic processes, pages 427-453. Elsevier.", "links": null }, "BIBREF17": { "ref_id": "b17", "title": "BabyWalk: Going farther in vision-and-language navigation by taking baby steps", "authors": [ { "first": "Wang", "middle": [], "last": "Zhu", "suffix": "" }, { "first": "Hexiang", "middle": [], "last": "Hu", "suffix": "" }, { "first": "Jiacheng", "middle": [], "last": "Chen", "suffix": "" }, { "first": "Zhiwei", "middle": [], "last": "Deng", "suffix": "" }, { "first": "Vihan", "middle": [], "last": "Jain", "suffix": "" }, { "first": "Eugene", "middle": [], "last": "Ie", "suffix": "" }, { "first": "Fei", "middle": [], "last": "Sha", "suffix": "" } ], "year": 2020, "venue": "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics", "volume": "", "issue": "", "pages": "2539--2556", "other_ids": { "DOI": [ "10.18653/v1/2020.acl-main.229" ] }, "num": null, "urls": [], "raw_text": "Wang Zhu, Hexiang Hu, Jiacheng Chen, Zhiwei Deng, Vihan Jain, Eugene Ie, and Fei Sha. 2020. BabyWalk: Going farther in vision-and-language navigation by taking baby steps. In Proceedings of the 58th Annual Meeting of the Association for Computational Lin- guistics, pages 2539-2556, Online. Association for Computational Linguistics.", "links": null } }, "ref_entries": { "FIGREF0": { "uris": null, "num": null, "type_str": "figure", "text": "Example illustrating a mapping from each action to a corresponding fine-grained instruction" }, "FIGREF1": { "uris": null, "num": null, "type_str": "figure", "text": "The distribution of the number of fine-grained instructions in the training split" }, "FIGREF2": { "uris": null, "num": null, "type_str": "figure", "text": "The distribution of action sequence length in the training split." }, "FIGREF3": { "uris": null, "num": null, "type_str": "figure", "text": "The distribution of action sequence length in the training split after the merge preprocessing." }, "TABREF0": { "html": null, "num": null, "content": "
shows the effects
", "type_str": "table", "text": "" }, "TABREF1": { "html": null, "num": null, "content": "", "type_str": "table", "text": "Ablation model performance(valid_seen). The values show the mean and standard deviation of 10 runs." }, "TABREF2": { "html": null, "num": null, "content": "
", "type_str": "table", "text": "The hyperparameters for training action embeddings" }, "TABREF3": { "html": null, "num": null, "content": "
valid_seenvalid_unseen
EMF1EMF1
MEAN 41.2
", "type_str": "table", "text": "\u00b1 10.2 78.2 \u00b1 10.8 44.5 \u00b1 12.1 79.0 \u00b1 11.5 SUM 34.5 \u00b1 13.8 70.7 \u00b1 15.2 36.8 \u00b1 8.4 73.3 \u00b1 12.1 LSTM 18.9 \u00b1 10.8 51.8 \u00b1 20.3 33.8 \u00b1 18.2 65.1 \u00b1 20.6" } } } }