Pattern Learning

From Knowitall
Revision as of 20:52, 3 January 2012 by Schmmd (talk | contribs) (Slots)

Jump to: navigation, search

Further Work

Bootrapping

Patterns

  • reduce postag restriction (collapse VB, VBZ, VBN, etc.)

Building the boostrapping data

Determining target relations

  1. Restrict high quality set of ClueWeb extractions to have proper noun arguments
  2. Choose the most frequent relations from this set

Determining target extractions (seeds)

  1. Start with the clean, chunked dataset of ReVerb extractions from ClueWeb.
  2. Apply Jonathan Berant's relation string normalization.
  3. Filter relations so each relation's normalized relation string matches a target relation and the arguments only contain DT, IN, NNP, and NNPS.
  4. Filter extractions
    1. Remove extractions that occur a single time.
    2. Remove extractions with single or double letter arguments, optionally ending with a period.
  5. Filter arguments
    1. Remove inc, ltd, vehicle, turn, page, site
    2. Remove arguments that are 2 or fewer characters
  6. Measure the occurrence of the arguments.
  7. Keep extractions from the target relations that have arguments that occur commonly (20 times).
  8. Remove target relations which have fewer than 15 seeds

Lemma grep

  1. Search corpus for all sentences that contain the lemmas in a target extraction.
  2. Remove duplicate sentences (sentence*extraction pairs must be unique).
  3. For each sentence*extraction pair, search for a pattern that connects the lemmas.
    1. Pattern must start with the arg1

Reducing the patterned results

  1. Don't allow patterns that contain punct edges or edges with non-word ([^\w]) characters
  2. Remove patterns that occur less than 10 times.
  3. Remove extractions that have an (extraction, pattern) pairs that occurs anomalously frequently.
    1. There was a single one: (hotel reservation, be make, online) ocurred 32k times, the next one ocurred 8k times

Extracting

Collapsing Dependencies

Collapsing noun edges

We want only the nn edge between "Barack Obama" to be collapsed in the following.

  1. US president Barack Obama declared victory yesterday.
  2. US president Barack Obama likes to drink beer.
  3. Barack Obama, the president of the US, has a wife.

Executing the Extractor

Generalized Extractor

  1. Apply pattern to sentence
  2. Remove matches with an adjacent `neg` edge
  3. Convert the match into an extraction

Specific Extractor

  1. Run the generalized extractor with the pattern from a (pattern, relation) pair
  2. Keep any extractions where the relations match

LDA Extractor

  1. Run the generalized extractor with a pattern
  2. Remove extractions with "relation strings" that don't match any target relation
  3. Keep best associated target relation by maximizing P(p | r)

Extractions

Slots

prepc

  1. After winning the Superbowl, the Saints are top dogs of the NFL.
  2. He purchased it without paying a premium.
  3. After winning the lottery, James becomes an Epicurean.
  4. Two months after joining the European Union , Bulgaria began attracting increasing interest towards local real estates.

partmod/advcl

  1. Having won the lottery, James becomes an Epicurean.