Learning High-Level Planning from Text
Branavan, S.R.K. and Kushman, Nate and Lei, Tao and Barzilay, Regina

Article Structure

Abstract

Comprehending action preconditions and effects is an essential step in modeling the dynamics of the world.

Introduction

Understanding action preconditions and effects is a basic step in modeling the dynamics of the world.

Related Work

Extracting Event Semantics from Text The task of extracting preconditions and effects has previously been addressed in the context of lexical semantics (Sil et al., 2010; Sil and Yates, 2011).

Problem Formulation

Our task is twofold.

Model

The key idea behind our model is to leverage textual descriptions of preconditions and effects to guide the construction of high level plans.

In our dataset only 11% of Candidate Relations are valid.

Input: A document d, Set of planning tasks G, Set of candidate precondition relations Call, Reward function 7“(), Number of iterations T

Experimental Setup

Datasets As the text description of our virtual world, we use documents from the Minecraft Wiki,7 the most popular information source about the game.

Results

Relation Extraction Figure 5 shows the performance of our method on identifying preconditions in text.

Conclusions

In this paper, we presented a novel technique for inducing precondition relations from text by grounding them in the semantics of planning operations.

Topics

relation extraction

Appears in 9 sentences as: Relation Extraction (1) relation extraction (6) relations extracted (3)
In Learning High-Level Planning from Text
  1. In this paper, we express the semantics of precondition relations extracted from text in terms of planning operations.
    Page 1, “Abstract”
  2. When applied to a complex virtual world and text describing that world, our relation extraction technique performs on par with a supervised baseline, yielding an F-measure of 66% compared to the baseline’s 65%.
    Page 1, “Abstract”
  3. The central idea of our work is to express the semantics of precondition relations extracted from text in terms of planning operations.
    Page 2, “Introduction”
  4. We build on the intuition that the validity of precondition relations extracted from text can be informed by the execution of a low-level planner.3 This feedback can enable us to learn these relations without annotations.
    Page 2, “Introduction”
  5. Our results demonstrate the strength of our relation extraction technique — while using planning feedback as its only source of supervision, it achieves a precondition relation extraction accuracy on par with that of a supervised SVM baseline.
    Page 2, “Introduction”
  6. Evaluation Metrics We use our manual annotations to evaluate the type-level accuracy of relation extraction .
    Page 6, “Experimental Setup”
  7. Baselines To evaluate the performance of our relation extraction , we compare against an SVM classifier8 trained on the Gold Relations.
    Page 6, “Experimental Setup”
  8. Relation Extraction Figure 5 shows the performance of our method on identifying preconditions in text.
    Page 7, “Results”
  9. While using planning feedback as its only source of supervision, our method for relation extraction achieves a performance on par with that of a supervised baseline.
    Page 8, “Conclusions”

See all papers in Proc. ACL 2012 that mention relation extraction.

See all papers in Proc. ACL that mention relation extraction.

Back to top.

SVM

Appears in 8 sentences as: SVM (8)
In Learning High-Level Planning from Text
  1. Our results demonstrate the strength of our relation extraction technique — while using planning feedback as its only source of supervision, it achieves a precondition relation extraction accuracy on par with that of a supervised SVM baseline.
    Page 2, “Introduction”
  2. Baselines To evaluate the performance of our relation extraction, we compare against an SVM classifier8 trained on the Gold Relations.
    Page 6, “Experimental Setup”
  3. We test the SVM baseline in a leave-one-out fashion.
    Page 6, “Experimental Setup”
  4. Model F-score 0.4 _ ---- -- SVM F-score ---------- -- All-text F-score
    Page 7, “Experimental Setup”
  5. Figure 5: The performance of our model and a supervised SVM baseline on the precondition prediction task.
    Page 7, “Experimental Setup”
  6. We also show the performance of the supervised SVM baseline.
    Page 7, “Results”
  7. Feature Analysis Figure 7 shows the top five positive features for our model and the SVM baseline.
    Page 8, “Results”
  8. Figure 7: The top five positive features on words and dependency types learned by our model (above) and by SVM (below) for precondition prediction.
    Page 8, “Results”

See all papers in Proc. ACL 2012 that mention SVM.

See all papers in Proc. ACL that mention SVM.

Back to top.

model parameters

Appears in 5 sentences as: Model parameters (1) model parameters (4)
In Learning High-Level Planning from Text
  1. Update the model parameters , using the low-level planner’s success or failure as the source of supervision.
    Page 4, “Model”
  2. where 6C is the vector of model parameters .
    Page 4, “Model”
  3. Initialization: Model parameters 0,, = 0 and 00 = 0.
    Page 4, “In our dataset only 11% of Candidate Relations are valid.”
  4. As before, 633 is the vector of model parameters , and gbx is the feature function.
    Page 5, “In our dataset only 11% of Candidate Relations are valid.”
  5. Therefore, during learning, we need to find the model parameters that maximize expected future reward (Sutton and Barto, 1998).
    Page 5, “In our dataset only 11% of Candidate Relations are valid.”

See all papers in Proc. ACL 2012 that mention model parameters.

See all papers in Proc. ACL that mention model parameters.

Back to top.

dependency parse

Appears in 3 sentences as: dependency parse (2) dependency parses (1)
In Learning High-Level Planning from Text
  1. grounding, the sentence 13k, and a given dependency parse qk of the sentence.
    Page 4, “In our dataset only 11% of Candidate Relations are valid.”
  2. The text component features gbc are computed over sentences and their dependency parses .
    Page 6, “In our dataset only 11% of Candidate Relations are valid.”
  3. The Stanford parser (de Marneffe et al., 2006) was used to generate the dependency parse information for each sentence.
    Page 6, “In our dataset only 11% of Candidate Relations are valid.”

See all papers in Proc. ACL 2012 that mention dependency parse.

See all papers in Proc. ACL that mention dependency parse.

Back to top.

F-score

Appears in 3 sentences as: F-score (5)
In Learning High-Level Planning from Text
  1. Specifically, it yields an F-score of 66% compared to the 65% of the baseline.
    Page 2, “Introduction”
  2. Model F-score 0.4 _ ---- -- SVM F-score ---------- -- All-text F-score
    Page 7, “Experimental Setup”
  3. Precondition prediction F-score
    Page 7, “Experimental Setup”

See all papers in Proc. ACL 2012 that mention F-score.

See all papers in Proc. ACL that mention F-score.

Back to top.

manual annotations

Appears in 3 sentences as: manual annotations (2) manually annotated (1)
In Learning High-Level Planning from Text
  1. We also manually annotated the relations expressed in the text, identifying 94 of the Candidate Relations as valid.
    Page 6, “Experimental Setup”
  2. Evaluation Metrics We use our manual annotations to evaluate the type-level accuracy of relation extraction.
    Page 6, “Experimental Setup”
  3. The first, Manual Text, is a variant of our model which directly uses the links derived from manual annotations of preconditions in text.
    Page 7, “Experimental Setup”

See all papers in Proc. ACL 2012 that mention manual annotations.

See all papers in Proc. ACL that mention manual annotations.

Back to top.

significantly outperforms

Appears in 3 sentences as: significantly outperforms (3)
In Learning High-Level Planning from Text
  1. Additionally, we show that a high—level planner utilizing these extracted relations significantly outperforms a strong, text unaware baseline — successfully completing 80% of planning tasks as compared to 69% for the baseline.1
    Page 1, “Abstract”
  2. Our results show that our text-driven high-level planner significantly outperforms all baselines in terms of completed planning tasks — it successfully solves 80% as compared to 41% for the Metric-FF planner and 69% for the text unaware variant of our model.
    Page 2, “Introduction”
  3. We show that building high-level plans in this manner significantly outperforms traditional techniques in terms of task completion.
    Page 8, “Conclusions”

See all papers in Proc. ACL 2012 that mention significantly outperforms.

See all papers in Proc. ACL that mention significantly outperforms.

Back to top.