Conference PaperPDF Available

An Analysis of Open Information Extraction based on Semantic Role Labeling

Authors:

Abstract and Figures

Open Information Extraction extracts relations from text without requiring a pre-specified domain or vocabulary. While existing techniques have used only shallow syntactic features, we investigate the use of semantic role labeling techniques for the task of Open IE. Semantic role labeling (SRL) and Open IE, although developed mostly in isolation, are quite related. We compare SRL-based open extractors, which perform computationally expensive, deep syntactic analysis, with TextRunner, an open extractor, which uses shallow syntactic analysis but is able to analyze many more sentences in a fixed amount of time and thus exploit corpus-level statistics. Our evaluation answers questions regarding these systems, including, can SRL extractors, which are trained on PropBank, cope with heterogeneous text found on the Web? Which extractor attains better precision, recall, f-measure, or running time? How does extractor performance vary for binary, n-ary and nested relations? How much do we gain by running multiple extractors? How do we select the optimal extractor given amount of data, available time, types of extractions desired?
Content may be subject to copyright.
An Analysis of Open Information Extraction based on
Semantic Role Labeling
Janara Christensen ,Mausam ,Stephen Soderland and Oren Etzioni
Turing Center
University of Washington
Seattle, WA 98195, USA
{janara,mausam,soderlan,etzioni}@cs.washington.edu
ABSTRACT
Open Information Extraction extracts relations from text
without requiring a pre-specified domain or vocabulary.
While existing techniques have used only shallow syn-
tactic features, we investigate the use of semantic role
labeling techniques for the task of Open IE. Semantic
role labeling (SRL) and Open IE, although developed
mostly in isolation, are quite related. We compare SRL-
based open extractors, which perform computationally
expensive, deep syntactic analysis, with TextRunner,
an open extractor, which uses shallow syntactic analysis
but is able to analyze many more sentences in a fixed
amount of time and thus exploit corpus-level statistics.
Our evaluation answers questions regarding these sys-
tems, including, can SRL extractors, which are trained
on PropBank, cope with heterogeneous text found on
the Web? Which extractor attains better precision, re-
call, f-measure, or running time? How does extractor
performance vary for binary, n-ary and nested relations?
How much do we gain by running multiple extractors?
How do we select the optimal extractor given amount
of data, available time, types of extractions desired?
1. INTRODUCTION
The challenge of Machine Reading and Knowledge Ex-
traction at Web scale [10] requires a scalable system
for extracting diverse information from large, hetero-
geneous corpora. The traditional approaches to infor-
mation extraction, e.g., [17, 1], seek to learn individual
extractors for each relation of interest and hence, cannot
scale to the millions of relations found on the Web. In
response, the Open Information Extraction paradigm [5]
attempts to overcome this knowledge acquisition bottle-
neck by extracting relational tuples without any restric-
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. To copy otherwise, or
republish, to post on servers or to redistribute to lists, requires prior specific
permission and/or a fee.
Copyright ACM ...$10.00
tions of a pre-specified vocabulary, domain or ontology.
TextRunner[6], a state-of-the-art open extractor, is a
relation classifier based primarily on shallow syntactic
features. In this paper, we study the applicability of
semantic role labeling (SRL) for the task of Open IE.
Our first observation is that SRL and Open IE, although
developed in isolation, are related tasks: semantically
labeled arguments correspond to the arguments in Open
IE extractions, and verbs often match up with Open IE
relations. We construct a scheme for Open IE based on
SRL, and create novel extractors based on two state-of-
the-art SRL systems, one developed at UIUC and the
other at Lund [15, 12]. These systems represent the
top ranking systems at CoNLL 2005 and 2008, respec-
tively. We study the trade-offs between TextRunner
and SRL-based extractors across a broad range of met-
rics and experimental conditions, both qualitative and
quantitative.
Given the distinct perspectives from which Open IE
and SRL have been developed, we expect TextRun-
ner and SRL extractors to be quite different. For ex-
ample, we expect the SRL extractors to have lower re-
calls on Web text due to out-of-vocabulary verbs and
diverse writing styles, since they are trained on a more
homogeneous corpus (PropBank). On the other hand,
their deep processing in comparison to TextRunner’s
shallow syntactic features may result in a much higher
precision. We also believe a priori that TextRunner
will be faster, but cannot quantify the difference, as no
previous work has studied this. This paper reports that,
contrary to our beliefs, SRL is robust to noisy Web text,
and achieves a much larger recall; whereas TextRun-
ner obtains a much higher precision than SRL extrac-
tors, at lower recalls. Finally, TextRunner is 20-700
times faster than SRL systems we have tested (depen-
dent on use of dependency parser versus constituency).
While previous studies assume a small data set and am-
ple time available for processing, to our knowledge, we
are the first to study the alternative experimental con-
ditions in which the data set is large and the processing
time is limited. This is especially important for most
universities and companies, which do not have access
to Microsoft or Google sized cluster of machines. Even
cloud computing resources are expensive and hence lim-
ited. In this paper, we examine both of these conditions,
SmallCorpus and LargeCorpus respectively. For
LargeCorpus, we devise a scheme to make use of the
available time carefully. We first run TextRunner,
which is enormously faster than the SRL-extractors we
tested. For the remaining available time, we run other
extractors over an intelligently chosen subset of the cor-
pus. This hybrid scheme obtains the best value for time
compared to the individual extractors.
We first describe the Open IE paradigm in more de-
tail and then explain how SRL systems can be used to
produce Open IE output. Next, we discuss qualitative
differences between the output produced by traditional
Open IE systems and the output produced by SRL-
based systems. Finally, we describe two experiments
under different conditions and end with related work
and conclusions.
2. BACKGROUND
Open Information Extraction [5] is a paradigm
where the system makes a single (or constant number
of) pass(es) over its corpus and extracts a large set of
relational tuples without requiring any relation-specific
training data. These tuples attempt to capture the
salient relationships expressed in each sentence. For in-
stance, given the sentence, “McCain fought hard against
Obama, but finally lost the election,” an Open IE sys-
tem would extract two tuples <McCain, fought against,
Obama>, and <McCain, lost, the election>. These tu-
ples can be binary or n-ary, where the relationship is ex-
pressed between more than two entities, or nested (e.g.,
<Microsoft, announced,<they, acquired, Komoku>>).
TextRunner is a state-of-the-art Open IE system that
performs extraction in two key steps. (1) A self-supervised
learner outputs a CRF-based classifier that uses unlex-
icalized features (it models closed class words but not
function words) for extracting relationships. The self-
supervised nature alleviates the need for hand-labeled
training data and unlexicalized features help scale to the
multitudes of relations found on the Web. (2) A single
pass extractor, which uses shallow syntactic techniques
like POS tagging and NP chunking, applies the CRF
classifier to extract an unbounded number of relation-
ships expressed in text. The use of shallow features
makes TextRunner highly efficient.
The extractions from TextRunner are ranked using
redundancy, an assessor that assigns higher confidence
to tuples occurring multiple times based on a proba-
bilistic model [9].
Semantic Role Labeling consists of detecting se-
Tuples Binary, N-ary, Nested
Metrics Precision, Recall, F-measure
Settings SmallCorpus,LargeCorpus
Systems TextRunner,SRL-IE-UIUC,SRL-IE-Lund
Table 1: A table outlining the various experi-
mental conditions in this paper.
mantic arguments associated with a verb in a sentence
and their roles (such as Agent, Patient, Instrument,
etc.). Given the sentence “The pearls I left to my son
are fake” an SRL system would conclude that for the
verb ‘leave’, ‘I’ is the agent, ‘pearls’ is the patient and
‘son’ is the benefactor. Because not all roles feature in
each verb, the roles are commonly divided into meta-
roles (A0-A7) and additional common classes such as
location, time, etc. Each Aican represent a different
role based on the verb, though A0 and A1 most often
refer to agents and patients respectively.
Availability of lexical resources such as PropBank [13]
and FrameNet [3], both of which annotate text with
roles for each argument, has enabled significant progress
in SRL systems over the last few years [18, 12, 8, 14].
We use UIUC-SRL [15] and Lund-SRL [12] as our
base SRL systems. We choose these systems, as they
represent the state-of-the-art for systems based on con-
stituency and dependency parsing – they are winners of
the CoNLL shared tasks 2005 and 2008 respectively.
Both these SRL systems apply a pipeline of parsing,
argument identification, and classification trained over
PropBank. UIUC-SRL operates in four key steps: prun-
ing, argument identification, argument classification and
inference. Pruning involves using a full parse tree and
heuristic rules to eliminate constituents that are un-
likely to be arguments. Argument identification uses a
classifier to identify constituents that are potential ar-
guments. In argument classification, a classifier assigns
role labels to the candidates identified in the previous
stage. Argument information is incorporated across ar-
guments in the inference stage, which uses an integer
linear program to make global role predictions.
Lund-SRL has a similar process. It first applies a de-
pendency parser and then uses a pipeline of classifiers to
identify predicates and identify and classify arguments.
Next it applies a set of linguistic constraints and uses
a predicate-argument reranker to rank the candidates.
Lastly, it uses a syntactic-semantic reranker to score the
joint syntactic-semantic models.
3. SRL-BASED OPEN IE
Our first observation is that verbs and their semanti-
cally labeled arguments almost always correspond to
Open IE relations and arguments respectively. SRL
computes more information than Open IE requires. There-
fore, for the purpose of a comparison with Open IE
systems, we convert SRL output into extractions. We
illustrate this conversion process via an example.
For example, given the sentence, “Eli Whitney created
the cotton gin in 1793,” TextRunner extracts two
tuples, one binary and one n-ary:
arg0 Eli Whitney arg0 Eli Whitney
rel created rel created (arg1) in
arg1 the cotton gin arg1 the cotton gin
arg2 1793
binary tuple n-ary tuple
The SRL systems label constituents of a sentence with
the role they play in regards to the verb in the sentence.
An SRL system will identify the following semantic roles
for the verb ‘create’:
A0 Eli Whitney
verb created
A1 the cotton gin
temporal in 1793
It is easy to see that the two formats are very related.
For fair comparisons we convert SRL output to equiva-
lent number of Open IE tuples. Our method first assigns
the verb along with its modifiers, following preposition,
and negation, if present, to be the relation. It then as-
signs all constituents labeled Aifor that verb, as well
as any that are marked Direction,Location, or Tempo-
ral to be the arguments of the relation. We order the
arguments in the same order as they are in the sentence
and with regard to the relation (except for direction, lo-
cation and temporal, which cannot be arg0 of an Open
IE extraction and are placed at the end of argument
list). As we are interested in relationships between en-
tities, we consider only the verbs that have at least two
arguments.
The generation of nested relations happens similarly.
The key difference is in identifying whether a seman-
tic tuple is a nested extraction. SRL-IE identifies such
cases by noticing that an argument to one verb is long
and contains a full semantic tuple with a different verb.
This is easy to operationalize since an SRL system al-
ways reports all the semantic tuples found in the sen-
tence.
In our experiments, we ignore part of the semantic in-
formation (such as distinctions between various Ai’s)
that UIUC-SRL and Lund-SRL provide. An IE sys-
tem built using SRL may retain this information, if the
downstream process (such as question answering en-
gine) can make use of this information. Notice that,
in our conversion, an SRL extraction that was correct
in the original format is never changed to an incorrect
Open IE extraction. However, an incorrectly labeled
SRL extraction could convert to a correct Open IE ex-
traction, if the arguments were correctly identified but
assigned incorrect semantic roles.
4. QUALITATIVE COMPARISON OF EX-
TRACTORS
Because SRL and Open IE are developed from differ-
ent perspectives, we first study their differences quali-
tatively.
Argument boundaries: The SRL systems are le-
nient in deciding what constitutes an argument and
tend to err on the side of including too much rather than
too little; TextRunner is more conservative, some-
times to the extent of omitting crucial information, par-
ticularly post-modifying clauses and PPs. For example,
TextRunner extracts <Bunsen, invented, a device>
from the sentence “Bunsen invented a device called the
Spectroscope”. SRL extractors include the entire phrase
“a device called the Spectroscope” as the second argu-
ment. Generally, the longer arguments in SRL-IE-
UIUC and SRL-IE-Lund are more informative, but
TextRunner’s succinct arguments normalize better
leading to an effective use of redundancy in ranking.
Out-of-vocabulary verbs: While we expected Tex-
tRunner to handle unknown verbs with little difficulty
due to its unlexicalized nature, the SRL-based systems
could have had severe trouble leading to a limited appli-
cability in the context of Web text. However, contrary
to our expectations, both SRL systems gracefully han-
dle new verbs (i.e., verbs not in their PropBank train-
ing) by only attempting to identify A0 (the agent) and
A1 (the patient). In practice, this is very effective –
both SRL extractors recognize the verb and its two ar-
guments correctly in “Larry Page googled his name and
launched a revolution.” In practice, out-of-vocabulary
verbs are rare and appeared in only 5% of our data.
Part-of-speech ambiguity: All systems have dif-
ficulty in cases where the part of speech of a word is
ambiguous or difficult to tag automatically. For exam-
ple, the word ‘write’ when used as a noun causes trouble
for both systems. In the sentence, “Be sure the file has
write permission.”, all three extractors extract <the
file, write, permission>. Part-of-speech ambiguity af-
fected about 20% of sentences.
Complex sentences: Because TextRunner relies
on shallow syntactic features, it performs more poorly
on complex sentences. SRL-based systems, due to their
deeper processing, can better handle complex syntax
and long-range dependencies.
N-ary and nested relations All extractors suffer sig-
nificant quality loss in complex extractions compared
to binary. For example, given the sentence “Google an-
nounced it will acquire YouTube,” TextRunner mis-
takenly extracts <Google, announced, it>. N-ary and
nested relations were present in 40% and 35% of our
data respectively.
5. EXPERIMENTAL RESULTS
In our quantitative evaluation we examine the strengths
and weaknesses of these extractors under two exper-
imental conditions: (1) SmallCorpus (Section 5.1),
in which we have a small corpus and ample compu-
tation time available, and (2) LargeCorpus (Section
5.2), in which the corpus is large and all systems cannot
complete the processing. We evaluate the quality of bi-
nary, n-ary, and nested extractions in all these settings
(see Table 1). In all experiments, we threshold Tex-
tRunner’s CRF-confidence at 0.4, which maximizes
F-measure on a development set.
Data sets: Because of the SRL-based systems’ rela-
tively slow processing time, we required our test sets be
of manageable size. Moreover, Open IE on the Web has
benefited from redundancy, and so the data sets needed
to mimic the redundancy found on the Web. We created
two test sets, one for binary and n-ary, and the other for
nested extractions. The first set focused on five target
relations – invent, graduate, study, write, and develop,
and the second used two relations with common nested
extractions – say and announce. The first set of rela-
tions is similar to that used in [5]. To our knowledge,
we are the first to investigate nested extractions, hence
our dataset to study that is unique.
A key challenge was dataset construction that was of
manageable size but still mimicked the redundancy found
on the Web. We could not use prior datasets, since ei-
ther they were not Web text, or they were too large.
To obtain redundant data commonly found on the Web,
we first queried a corpus of 500M Web documents for
a sample of sentences with these verbs (or their in-
flected forms, e.g., invents). We chose 200 typical agents
per verb (e.g., Edison (for invent), Microsoft (for an-
nounce)) and searched for sentences with both the verb
as well as these agents. This resulted in a test set of
29,842 sentences for binary and n-ary relations, and
a second set of 16,777 sentences for nested relations.
These test sets have the desired properties and enable
us to study the performance of different extractors on
Web text.
To compute precision on these test sets, we tagged a
random sample of over 4,800 extractions. A tuple is cor-
rect if the arguments have correct boundaries and the
relation accurately expresses the relationship between
all of the arguments, even if the relation is uninforma-
tive. For example, for the sentence “Bunsen invented a
device called the Spectroscope”, both second arguments,
‘a device’ and ‘a device called the Spectroscope’ would
be marked as correct.
Determining the absolute recall is precluded by the amount
of hand labeling necessary. Instead, we compute pseudo-
recall by taking the union of correct tuples from all
methods as denominator.1
5.1 SmallCorpus Setting
Table 2 reports the performance of the three extractors
on our data sets for this traditional NLP setting. Over-
all, SRL-IE-Lund achieves the highest precision, and
SRL-IE-UIUC achieves the highest recall and the high-
est F1 score. TextRunner’s performance on nested
extractions is especially poor, but that is expected, since
it only extracts relations between noun phrases and
nested extractions, by definition, have a full extraction
as an argument. On the other hand, both SRL-based
systems run far slower than TextRunner.TextRun-
ner on average processes a sentence in under 0.02 secs
whereas SRL-IE-Lund and SRL-IE-UIUC are over
20x and 500x slower respectively. SRL-IE-UIUC ran
much slower because of its use of constituency parsing
(which is slower compared to dependency parsing) and
its use of an integer linear program for global inference.
By taking a union of the SRL-based systems’ output
and the highest precision subset of TextRunner’s ex-
tractions, we achieve the highest recall and F-measure
(Table 2). We identify the highest precision subset of
TextRunner’s extractions by our novel locality rank-
ing (see Figure 2).2This shows the benefit of using
multiple systems for extraction – they extract different
tuples. We call this the smart union of all systems,
and this is the method of choice for the SmallCorpus
setting.
Although SRL-IE-Lund has higher overall precision,
there are some conditions under which the shallow pro-
cessing of TextRunner can obtain superior precision!
We analyze the performance of these systems under two
different rankings – redundancy, which has been exam-
ined before for TextRunner[4], and locality, a novel
measure.
Redundancy: Redundancy is the number of times a
relation has been extracted from unique sentences. In-
tuitively, we have more confidence in a relation that is
extracted many times than a relation that is extracted
only a small number of times. We compute redundancy
over normalized extractions, ignoring noun modifiers,
adverbs, and verb inflection. Figure 1(a) displays the
results for binary extractions, ranked by redundancy.
We use a log scale on the x-axis, since high redundancy
extractions account for less than 1% of the recall. For
binary extractions, redundancy improved TextRun-
ner’s precision significantly, but at a dramatic loss in
recall – it achieved 0.82 precision at 0.009 recall. For a
highly redundant corpus, TextRunner would be the
1Tuples from two systems are considered equivalent if for
the relation and each argument, the extracted phrases are
equal or if one phrase is contained within the phrase of the
other.
2discussed in more detail next.
TextRunner SRL-IE-Lund SRL-IE-UIUC Smart Union
P R F1 P R F1 P R F1 P R F1
Binary .55 .26 .35 .70 .51 .59 .63 .75 .68 .67 .95 .77
N-ary .42 27 .32 .61 .27 .37 .53 .57 .55 .56 .78 .66
Cpu Time: Binary, N-ary 6 minutes 155 minutes 3126 minutes 3287 minutes
Nested .09 .02 .03 .63 .44 .59 .52 .84 .64 .57 1.0 .72
Cpu Time: Nested 3 minutes 60 minutes 2016 minutes 2078 minutes
Table 2: In SmallCorpus, SRL-IE-Lund has the highest precision. Taking the union of the SRL systems and the
higher precision results from TextRunner achieves the highest recall and F-measure. Both SRL-based systems require
over an order of magnitude more processing time. The bold values indicate the highest values for the metric and
relation-type.
1e−04 1e−03 1e−02 1e−01 1e+00
0.0 0.2 0.4 0.6 0.8 1.0
Recall
Precision
TextRunner
SRL−IE−Lund
SRL−IE−UIUC
1e−04 1e−03 1e−02 1e−01 1e+00
0.0 0.2 0.4 0.6 0.8 1.0
Recall
Precision
TextRunner
SRL−IE−Lund
SRL−IE−UIUC
Figure 1: SmallCorpus: redundancy ranking for binary and n-ary relations. (Note the log scale) (a) TextRunner
has highest precision at highest redundancy, but at a very low recall (0.01). The arguments in SRL-based extractors
do not normalize well; the high redundancy region has a large fraction of systematic extraction errors. (b) For n-ary
extractions, redundancy is not very effective.
0.0 0.2 0.4 0.6 0.8 1.0
0.0 0.2 0.4 0.6 0.8 1.0
Recall
Precision
TextRunner
SRL−IE−Lund
SRL−IE−UIUC
0.0 0.2 0.4 0.6 0.8 1.0
0.0 0.2 0.4 0.6 0.8 1.0
Recall
Precision
TextRunner
SRL−IE−Lund
SRL−IE−UIUC
Figure 2: SmallCorpus: locality ranking for binary, and n-ary relations. (a) Locality ranking gives a large boost
to TextRunner’s precision for binary relations, and at a recall of 0.2, much higher than that achieved by redundancy
ranking. (b) For n-ary extractions, locality helps all systems.
algorithm of choice, however, this experiment clearly
shows that highly redundant extractions are usually
very limited, even in Web-style text.
For n-ary and nested relations, and binary relations for
SRL-based extractors (Figure 1(a,b)), redundancy ac-
tually hurts precision (nested relation graphs omitted).
These extractions tend to be so specific that genuine
redundancy is rare, and the highest frequency extrac-
tions are often systematic errors. E.g., the most fre-
quent SRL-IE-UIUC extraction was <nothing, write,
home>, from sentences with the phrase “nothing to
write home about”.
Locality: Our experiments with TextRunner led us
to discover a new validation scheme for the extractions
locality. We define locality as the number of tokens in
between the first and the last arguments in the sentence.
We observed that TextRunner’s shallow features can
identify relations more reliably when the arguments are
closer to each other in the sentence. Figure 2 reports
the results from ranking extractions by locality.
We find a clear correlation between locality and pre-
cision of TextRunner, with precision 0.81 at recall
0.17, where the locality is 3 tokens or less for binary
extractions. This result is very surprising because SRL
systems perform deep syntactic and additional seman-
tic analysis, still TextRunner’s shallow syntactic pro-
cessing with simple locality assessment is able to obtain
significantly higher precision (though at reasonable, but
lower recall of 0.17). Comparing this result with preci-
sion 0.82 at recall 0.01 for redundancy-based assessing,
we find that locality-based ranking is dramatically more
useful. SRL extractors do not benefit from locality for
binary extractions. For n-ary relations, all systems can
improve precision by varying locality. Locality has little
effect on nested relations.
This new assessor allows us to construct a high preci-
sion subset of TextRunner, which we use in the smart
union (see the experiment above). For binary extrac-
tions we filter all extractions where locality >5, i.e.,
there are more than five tokens between the arguments.
An Ablation Study: SRL systems perform two key
types of processing in addition to TextRunner’s shal-
low syntactic techniques: (1) they use a full syntac-
tic parser, and (2) they perform argument identifica-
tion and classification. To tease apart the benefits, we
perform an additional experiment in which we create
extractions directly from the output of Lund-SRL’s
parser. These extractions achieve a precision of 0.62
at recall of 0.41 for binary extractions. This is much
lower than SRL-IE-Lund’s results (precision 0.7 and
recall 0.51) illustrating the gain due to a complete SRL
system.
0 50 100 150
0.0 0.2 0.4 0.6 0.8
Time (minutes)
F−measure
TextRunner
SRL−IE−Lund
Hybrid
Figure 3: LargeCorpus: F-measure achieved in a given
amount of computation time. The hybrid extractor ob-
tains the best F-measure for binary extractions.
In summary, we find that SRL extractors perform better
overall, however, TextRunner, under locality rank-
ing, achieves superior precision at lower recalls. Nei-
ther redundancy nor locality benefits SRL extractors
much (except for n-ary). SRL is orders of magnitude
slower, which becomes a bottleneck in the next experi-
ment, when the available time is limited.
5.2 LargeCorpus Setting
To determine scaling to Web-scale knowledge extrac-
tion, we study an experimental condition, often not
considered in NLP and IE communities – the setting in
which the data is large enough that all processing can-
not be achieved. Because experimenting with massive
corpora makes it very hard to estimate (pseudo)recall
and thus the F-measure, we simulate this setting by
using our current data set, and varying the amount of
available time.
In the extreme setup when the time is so limited that
no extractor can complete processing, TextRunner is
the extractor of choice, because the recalls of the slower
extractors will be so low that they will far outweigh
the benefits of higher precision. TextRunner, on the
other hand, will be able to generate a large number of
extractions at reasonable precision.
In the more interesting and likely more realistic case,
where additional time is available after TextRunner
completes its processing, we have the opportunity to
combine the different extractors. We present a hybrid
system that combines the strengths of TextRunner
(fast processing time and high precision on a subset of
sentences) with the strengths of SRL-IE-Lund (higher
recall and better handling of long-range dependencies).
The focus is on using the remaining time efficiently. We
illustrate the binary setting, though results on n-ary are
similar, and don’t consider SRL-IE-UIUC owing to its
very slow speed, though it is straightforward to include.
We first run TextRunner over all the sentences and
then use the remaining time to run SRL-IE-Lund and
take the union of all extractions. We add to this idea by
using a filter policy and an intelligent order of sentences
for extraction to improve the precision.
TextRunner’s precision is low when the redundancy
of the extraction is low, and when the arguments are
far apart. Thus, redundancy, and locality form the key
factors for our filter policy: if both of these factors are
below a given threshold, discard the tuple. The thresh-
olds were determined by a parameter search over a small
development set.
A good ordering policy would apply SRL-IE-Lund first
to the sentences in which TextRunner extractions
have been filtered by the filter policy. We could rank
a sentence S according to the average distance between
pairs of arguments from all tuples extracted by Tex-
tRunner from S. While this ranking system would or-
der sentences according to their likelihood of yielding
maximum new information, it would miss the cost of
computation. To account for computation time, we ad-
ditionally estimate the amount of time SRL-IE-Lund
will take to process each sentence using a linear model
trained on the sentence length. We then choose the
sentence that maximizes information gain divided by
its estimated computation time.
If the available time is minimal then our hybrid ex-
tractor reduces to TextRunner. If it is very large,
then the hybrid is similar to the smart union of two
systems (see SmallCorpus). In intermediate cases,
hybrids make effective use of available time. Overall,
the hybrid extractor run the best algorithm given the
available computation time.
In summary, we find that all extractors are useful and,
for best results, they should all be employed to varying
degrees, which are based on the available time and their
efficiency.
Evaluation: Figure 3 reports F-measure for binary
extractions measured against available computation time.
Hybrid has substantially better F1 scores than both
TextRunner’s and SRL-IE-Lund’s demonstrating the
power of combining extractors.
6. RELATED WORK
Open information extraction is a relatively recent paradigm
and hence, has been studied by only a small number of
researchers. The most salient is TextRunner, which
also introduced the model [5, 6].
A recent Open IE system, WOE [21], uses dependency
features (WOEparse) and training data generated using
Wikipedia infoboxes to learn a series of open extractors
(WOEpos). Our ablation study in Section 5.1 suggests
the quality of the parser-based extractor to be between
TextRunner and complete SRL systems; we expect
WOEparse, their better performing system, to be sim-
ilar. Moreover, WOE does not output n-ary or nested
extractions.
A paradigm related to Open IE is Preemptive IE [16].
While one goal of Preemptive IE is to avoid relation-
specificity, Preemptive IE does not emphasize Web scal-
ability, which is essential to Open IE.
A version of Knext uses heuristic rules and syntactic
parses to convert a sentence into an unscoped logical
form [19]. This work is more suitable for extracting
common sense knowledge as opposed to factual infor-
mation.
Another related system is Wanderlust [2]. After an-
notating 10,000 sentences parsed with LinkGrammar,
it learns 46 general linkpaths as patterns for relation
extraction. In contrast to our approaches, this requires
a large set of hand-labeled examples.
We are the first to use SRL for Open IE, but its use
for traditional IE is investigated by Harabagiu et al.
[11]. They used a lexico-semantic feedback loop in a
question-answering system for a set of pre-defined rela-
tions.
7. CONCLUSIONS
This paper investigates the use of Semantic Role La-
beling for the task of Open Information Extraction.
Although the two tasks were developed in isolation,
they are quite related. We describe SRL-IE-UIUC
and SRL-IE-Lund, the first SRL-based Open IE sys-
tems. We empirically study the trade-offs between these
systems and TextRunner, a state-of-the-art Open IE
system under several settings: SmallCorpus with un-
bounded computation time, and LargeCorpus with
limited amount of time; using different metrics: preci-
sion, recall, F1 score and running time; and for different
kinds of extractions: binary, n-ary and nested.
We find that in the traditional NLP setting (SmallCorpus),
the deeper analysis of SRL-based systems overall out-
performs TextRunner. However, TextRunner out-
put can be ranked using our novel measure, locality,
leading to superior precision at non-trivial recalls. A
smart union of the three approaches performs best.
TextRunner is over an order of magnitude faster,
making it the algorithm of choice when time is ex-
tremely limited. These complimentary strengths lead
us to design a hybrid extractor that intelligently chooses
sentences to extract from, and thus, efficiently uses the
remaining computation time. Our hybrid extractor achieves
better performance than either system if an intermedi-
ate amount of time is available for processing. Overall,
we provide evidence that, contrary to belief in the Open
IE literature [6], deep syntactic approaches have a lot
to offer for the task of Open IE.
8. REFERENCES
[1] Eugene Agichtein and Luis Gravano. Snowball:
Extracting relations from large plain-text
collections. In Proceedings of the Fifth ACM
International Conference on Digital Libraries,
2000.
[2] Alan Akbik and J¨ugen Broß. Wanderlust:
Extracting semantic relations from natural
language text using dependency grammar
patterns. In Proceedings of the WWW 2009
Workshop on Semantic Search, 2009.
[3] Collin F. Baker, Charles J. Fillmore, and John B.
Lowe. The berkeley framenet project. In COLING
’98: Proceedings of the 17th international
conference on Computational linguistics, pages
86–90, 1998.
[4] Michele Banko. Open Information Extraction for
the Web. PhD thesis, University of Washington,
2009.
[5] Michele Banko, Michael J. Cafarella, Stephen
Soderland, Matt Broadhead, and Oren Etzioni.
Open information extraction from the web. In
IJCAI’07: Proceedings of the 20th international
joint conference on Artifical intelligence, pages
2670–2676, 2007.
[6] Michele Banko and Oren Etzioni. The tradeoffs
between open and traditional relation extraction.
In ACL ’08: Proceedings of the 46th Annual
Meeting of the Association for Computational
Linguistics, pages 28–36, 2008.
[7] Andrew Carlson, Justin Betteridge, Richard C.
Wang, Estevam R. Hruschka Jr., and Tom M.
Mitchell. Coupled semi-supervised learning for
information extraction. In WSDM ’10:
Proceedings of the Third ACM International
Conference on Web Search and Data Mining,
2010.
[8] Bonaventura Coppola, Alessandro Moschitti, and
Giuseppe Riccardi. Shallow semantic parsing for
spoken language understanding. In NAACL ’09:
Proceedings of Human Language Technologies:
The Annual Conference of the North American
Chapter of the Association for Computational
Linguistics, pages 85–88, 2009.
[9] Doug Downey, Oren Etzioni, and Stephen
Soderland. A probabilistic model of redundancy in
information extraction. In IJCAI ’05: Proceedings
of the 20th international joint conference on
Artifical intelligence, pages 1034–1041, 2005.
[10] Oren Etzioni, Michele Banko, and Michael J.
Cafarella. Machine reading. In AAAI’06:
Proceedings of the 21st national conference on
Artificial intelligence, pages 1517–1519, 2006.
[11] Sanda Harabagiu, Dan Moldovan, Marius Psca,
Rada Mihalcea, Mihai Surdeanu, Rˇazvan
Bunescu, Roxana Gˆırju, Vasile Rus, and Paul
Morˇarescu. The role of lexico-semantic feedback
in open-domain textual question-answering. In
ACL ’01: Proceedings of the 39th Annual Meeting
on Association for Computational Linguistics,
pages 282–289, 2001.
[12] Richard Johansson and Pierre Nugues. The effect
of syntactic representation on semantic role
labeling. In COLING ’08: Proceedings of the 22nd
International Conference on Computational
Linguistics, pages 393–400, 2008.
[13] Paul Kingsbury Martha and Martha Palmer.
From treebank to propbank. In LREC ’02:
Proceedings of the Third International Conference
on Language Resources and Evaluation, 2002.
[14] Alessandro Moschitti, Daniele Pighin, and
Roberto Basili. Tree kernels for semantic role
labeling. Computational Linguistics,
34(2):193–224, 2008.
[15] V. Punyakanok, D. Roth, and W. Yih. The
importance of syntactic parsing and inference in
semantic role labeling. Computational Linguistics,
34(2), 2008.
[16] Yusuke Shinyama and Satoshi Sekine. Preemptive
information extraction using unrestricted relation
discovery. In NAACL ’06: Proceedings of Human
Language Technologies: The Annual Conference
of the North American Chapter of the Association
for Computational Linguistics, pages 304–311,
2006.
[17] Stephen Soderland. Learning information
extraction rules for semi-structured and free text.
Machine Learning, 34(1-3):233–272, 1999.
[18] Kristina Toutanova, Aria Haghighi, and
Christopher D. Manning. A global joint model for
semantic role labeling. Computational Linguistics,
34(2):161–191, 2008.
[19] Benjamin Van Durme and Lenhart Schubert.
Open knowledge extraction through
compositional language processing. In STEP ’08:
Proceedings of the 2008 Conference on Semantics
in Text Processing, pages 239–254, 2008.
[20] Daniel S. Weld, Raphael Hoffmann, and Fei Wu.
Using wikipedia to bootstrap open information
extraction. SIGMOD Rec., 37(4):62–68, 2008.
[21] Fei Wu and Daniel S. Weld. Open information
extraction using wikipedia. In ACL ’10:
Proceedings of the 48th Annual Meeting on
Association for Computational Linguistics, 2010.
... Figure 2 shows a result of the SRL for this example, where meaning of the labels will be described in detail in Section IV. SRL has been used in many natural language processing (NLP) applications such as question answering [1], machine translation [2], document summarization [3] and information extraction [4]. Therefore, SRL is an important task in NLP. ...
... In addition to the ME classifier, we also try the Support Vector Machine (SVM) classifier, which usually gives good accuracy in a wide variety of classification problems. 4 Table IV shows the F 1 scores of different labelling strategies. We see that the SVM classifier outperforms ME the classifier by a large margin. ...
... We use the logistic regression classifier with L 2 regularization provided by the scikit-learn software package. The regularization term is fixed at 1.4 We use a linear SVM provided in the scikit-learn software package with default parameter values. ...
Preprint
Semantic role labelling (SRL) is a task in natural language processing which detects and classifies the semantic arguments associated with the predicates of a sentence. It is an important step towards understanding the meaning of a natural language. There exists SRL systems for well-studied languages like English, Chinese or Japanese but there is not any such system for the Vietnamese language. In this paper, we present the first SRL system for Vietnamese with encouraging accuracy. We first demonstrate that a simple application of SRL techniques developed for English could not give a good accuracy for Vietnamese. We then introduce a new algorithm for extracting candidate syntactic constituents, which is much more accurate than the common node-mapping algorithm usually used in the identification step. Finally, in the classification step, in addition to the common linguistic features, we propose novel and useful features for use in SRL. Our SRL system achieves an F1F_1 score of 73.53\% on the Vietnamese PropBank corpus. This system, including software and corpus, is available as an open source project and we believe that it is a good baseline for the development of future Vietnamese SRL systems.
... Based on these linguistic relations, they obtain a significantly higher performance over previous systems in the first generation. Several Open IE systems have been proposed after TextRunner and WOE, including ReVerb, OLLIE, Christensen et al., [5], ClausIE, Vo & Bagheri [21] with two extraction paradigms, namely verb-based relation extraction and clause-based relation extraction. ...
Preprint
Open Information Extraction (Open IE) systems aim to obtain relation tuples with highly scalable extraction in portable across domain by identifying a variety of relation phrases and their arguments in arbitrary sentences. The first generation of Open IE learns linear chain models based on unlexicalized features such as Part-of-Speech (POS) or shallow tags to label the intermediate words between pair of potential arguments for identifying extractable relations. Open IE currently is developed in the second generation that is able to extract instances of the most frequently observed relation types such as Verb, Noun and Prep, Verb and Prep, and Infinitive with deep linguistic analysis. They expose simple yet principled ways in which verbs express relationships in linguistics such as verb phrase-based extraction or clause-based extraction. They obtain a significantly higher performance over previous systems in the first generation. In this paper, we describe an overview of two Open IE generations including strengths, weaknesses and application areas.
... It utilizes a set of patterns in order to obtain propositions but does not capture the 'context' of each clause for effective extraction. A follow-up study relies on semantic features (semantic roles) for the OIE task, demonstrating that Semantic role labeling (SRL) can be used to increase the precision and recall of OIE [8]. Separately, a greedy parser, which relies on a classifier to predict the correct transition based on a small number of dense features, is treated for speedy parsing [6]. ...
Preprint
Full-text available
Large Language Models (LLMs) have received considerable interest in wide applications lately. During pre-training via massive datasets, such a model implicitly memorizes the factual knowledge of trained datasets in its hidden parameters. However, knowledge held implicitly in parameters often makes its use by downstream applications ineffective due to the lack of common-sense reasoning. In this article, we introduce a general framework that permits to build knowledge bases with an aid of LLMs, tailored for processing Web news. The framework applies a rule-based News Information Extractor (NewsIE) to news items for extracting their relational tuples, referred to as knowledge bases, which are then graph-convoluted with the implicit knowledge facts of news items obtained by LLMs, for their classification. It involves two lightweight components: 1) NewsIE: for extracting the structural information of every news item, in the form of relational tuples; 2) BERTGraph: for graph convoluting the implicit knowledge facts with relational tuples extracted by NewsIE. We have evaluated our framework under different news-related datasets for news category classification, with promising experimental results.
... The benefit of this method is that it may be used with texts from any domain. Some OpenIE systems that can extract information from the free text include: KnowItAll [52], TEX-TRUNNER [53], REVERB [54], SRL-IE [55], OLLIE [56], and RELNOUN [57]. ...
Thesis
Full-text available
Information Extraction (IE) refers to the process of automatically extracting structured data from unstructured sources to enable the utilization of such data by other applications. Extracting relations from textual sources, which seeks to detect the semantic relation represented between entities ref-erenced in the texts, is a common sub-problem. The objective of the RE task is to develop automatic extractors that can identify and extract structured, relational information from unstructured sources like natural language text. Assigning a relationship label to a pair of entities may be considered a classification problem. As a result, supervised machine learning methodologies can be employed. It is essential to pre-process the data using methods from natural language processing to organize the textual contents into meaningful data structures before extracting relations from the unprocessed text. In addition , as relations are represented between entities, it is necessary to locate the entities using an entity extraction technique, which is another information extraction sub-problem. Relation extraction methods that use entity-annotated text are called pipeline approaches. Relations can be represented between two or more than two entities which are known as binary relations and N-ary relations, respectively. This thesis limits our research to binary relations using pipeline approaches.
... This broadly comprises of two steps -identifying entities signifying specific information of interest (known as named entity recognition or NER), and inferring the relationships between identified entities (known as relation extraction or RE). Approaches underlying these tasks utilize grammatical structures to extract semantic frames [96,97], utilizing knowledge bases for relation inference [98], syntactic dependency parsing-based information extraction [99], semantic role labeling [100], coreference resolution [101], and so on. [110,90], process flowsheets [92,84], and many more applications of graph-based data mining as presented in [111]. ...
Thesis
Full-text available
Process systems engineering (PSE) involves a systems-level approach to solving problems in chemical engineering related to process modeling, design, control, and optimization and involves modeling interactions between various systems (and subsystems) governing the process. This requires using a combination of mathematical methods, physical intuition, and recently machine learning techniques. Recently, language models have seen tremendous advances due to new and more efficient model architectures (such as transformers), computing power, and large volumes of training data. Many of these language models could be appropriately adapted to solve several PSE-related problems. However, language models are inherently complex and are often characterized by several million parameters, which could only be trained efficiently in data-rich areas, unlike PSE. Moreover, PSE is characterized by decades of rich process knowledge that must be utilized during model training to avoid mismatch between process knowledge and data-driven language models. This thesis presents a framework for building domain-informed language models for several central problems in PSE spanning multiple scales. Specifically, the frameworks presented include molecular property prediction, forward and retrosynthesis reaction outcome prediction, chemical flowsheet representation and generation, pharmaceutical information extraction, and reaction classification. Domain knowledge is integrated with language models using custom model architectures, standard and custom-built ontologies, linguistics-inspired chemistry and process flowsheet grammar, adapted problem formulations, graph theory techniques, and so on. This thesis is intended to provide a path for future developments of domain-informed language models in process systems engineering that respect domain knowledge, but leverage their computational advantages.
... VBSRL and open IE. The adoption of SRL in machine understanding is not a novelty in the Open IE panorama; in fact, its potential has been already recognized and exploited in several Open IE system implementations -see, for example, Exemplar (de Sá Mesquita et al., 2013) and SRL-IE (Christensen et al., 2010). This constitutes an important legacy that guarantees some important benefits to the Cnosso method devised in the next section, and addresses some key points to improve the results of Open IE techniques -as pointed out in Etzioni et al. (2011). ...
Preprint
Open Information Extraction (OpenIE) methods extract (noun phrase, relation phrase, noun phrase) triples from text, resulting in the construction of large Open Knowledge Bases (Open KBs). The noun phrases (NPs) and relation phrases in such Open KBs are not canonicalized, leading to the storage of redundant and ambiguous facts. Recent research has posed canonicalization of Open KBs as clustering over manuallydefined feature spaces. Manual feature engineering is expensive and often sub-optimal. In order to overcome this challenge, we propose Canonicalization using Embeddings and Side Information (CESI) - a novel approach which performs canonicalization over learned embeddings of Open KBs. CESI extends recent advances in KB embedding by incorporating relevant NP and relation phrase side information in a principled manner. Through extensive experiments on multiple real-world datasets, we demonstrate CESI's effectiveness.
Article
Full-text available
Semantics means the process of extracting precise meaning from text or sentences. Semantic Role Labelling (SRL) plays a crucial role in Natural Language Processing by providing insights into the underlying meaning of sentences. SRL involves assigning generic labels or roles to words within a sentence, indicating their respective semantic roles. It helps in tasks such as information extraction, question answering, educational systems, sentiment analysis, and machine translation by identifying the roles of different words in a sentence and their relationships with each other. This process enables computers to better understand and process human language, leading to more accurate and effective language understanding. SRL also powers AI-driven educational tools like intelligent tutoring systems, personalized learning, and automated assessments, making education more adaptive and effective. In this paper, we are giving a brief review of SRL system developed for different languages. This literature review focuses on key aspects of SRL like techniques used, datasets used, languages for which SRL is developed, and accuracy. Our future work is inclined towards SRL for Hindi, so our review focuses on SRL developed for Hindi, along with other Indian Languages like Urdu, Malayalam and Tamil.
Chapter
Full-text available
Frame semantics (Fillmore 1976) is a framework for the analysis of natural language semantics in terms of conceptual structures which can provide a semantic basis to constructions. Its cross-lingual applicability is subject to debate. Our study leverages recent developments in computational linguistics to assess the cross-lingual applicability of frames as defined in the FrameNet framework. Using a pre-trained multilingual model fine-tuned on frame identification, we test if frames have comparable representations in different languages and if we can identify frames cross-lingually using only English annotated examples. We show, using data from German and Korean, that frames are generally cross-lingually stable and identifiable; cross-lingual differences in lexicalization patterns are the biggest obstacle to cross-lingual application.
Patent
Full-text available
To implement open information extraction, a new extraction paradigm has been developed in which a system makes a single data-driven pass over a corpus of text, extracting a large set of relational tuples without requiring any human input. Using training data, a Self-Supervised Learner employs a parser and heuristics to determine criteria that will be used by an extraction classifier (or other ranking model) for evaluating the trustworthiness of candidate tuples that have been extracted from the corpus of text, by applying heuristics to the corpus of text. The classifier retains tuples with a sufficiently high probability of being trustworthy. A redundancy-based assessor assigns a probability to each retained tuple to indicate a likelihood that the retained tuple is an actual instance of a relationship between a plurality of objects comprising the retained tuple. The retained tuples comprise an extraction graph that can be queried for information.
Article
Full-text available
A great share of applications in modern information tech-nology can benefit from large coverage, machine accessible knowledge bases. However, the bigger part of todays knowl-edge is provided in the form of unstructured data, mostly plain text. As an initial step to exploit such data, we present Wanderlust, an algorithm that automatically extracts se-mantic relations from natural language text. The procedure uses deep linguistic patterns that are defined over the depen-dency grammar of sentences. Due to its linguistic nature, the method performs in an unsupervised fashion and is not restricted to any specific type of semantic relation. The ap-plicability of the proposed approach is examined in a case study, in which it is put to the task of generating a seman-tic wiki from the English Wikipedia corpus. We present an exhaustive discussion about the insights obtained from this particular case study including considerations about the gen-erality of the approach.
Article
Full-text available
We present results for a system designed to perform Open Knowledge Extraction, based on a tradition of compositional lan-guage processing, as applied to a large collection of text derived from the Web. Evaluation through manual assessment shows that well-formed propositions of reasonable quality, representing general world knowledge, given in a logical form potentially useable for inference, may be extracted in high volume from arbitrary input sentences. We compare these re-sults with those obtained in recent work on Open Information Extraction, indicat-ing with some examples the quite differ-ent kinds of output obtained by the two approaches. Finally, we observe that por-tions of the extracted knowledge are com-parable to results of recent work on class attribute extraction.
Conference Paper
Full-text available
We consider the problem of semi-supervised learning to extract categories (e.g., academic fields, athletes) and relations (e.g., PlaysSport(athlete, sport)) from web pages, starting with a handful of labeled training examples of each category or relation, plus hundreds of millions of unlabeled web documents. Semi-supervised training using only a few labeled examples is typically unreliable because the learning task is underconstrained. This paper pursues the thesis that much greater accuracy can be achieved by further constraining the learning task, by coupling the semi-supervised training of many extractors for different categories and relations. We characterize several ways in which the training of category and relation extractors can be coupled, and present experimental results demonstrating significantly improved accuracy as a result.
Conference Paper
Full-text available
It is often too expensive to compute and materialize a complete high-dimensional data cube. Computing an iceberg cube, which contains only aggregates above certain thresholds, is an effective way to derive nontrivial multi-dimensional aggregations for ...
Conference Paper
Full-text available
Almost all automatic semantic role label- ing (SRL) systems rely on a preliminary parsing step that derives a syntactic struc- ture from the sentence being analyzed. This makes the choice of syntactic repre- sentation an essential design decision. In this paper, we study the influence of syn- tactic representation on the performance of SRL systems. Specifically, we com- pare constituent-based and dependency- based representations for SRL of English in the FrameNet paradigm. Contrary to previous claims, our results demonstrate that the systems based on de- pendencies perform roughly as well as those based on constituents: For the ar- gument classification task, dependency- based systems perform slightly higher on average, while the opposite holds for the argument identification task. This is re- markable because dependency parsers are still in their infancy while constituent pars- ing is more mature. Furthermore, the re- sults show that dependency-based seman- tic role classifiers rely less on lexicalized features, which makes them more robust to domain changes and makes them learn more efficiently with respect to the amount of training data.
Conference Paper
Full-text available
Information-extraction (IE) systems seek to distill semantic relations from natural-language text, but most systems use supervised learning of relation-specific examples and are thus limited by the availability of training data. Open IE systems such as TextRunner, on the other hand, aim to handle the unbounded number of relations found on the Web. But how well can these open systems perform? This paper presents WOE, an open IE system which improves dramatically on TextRunner's precision and recall. The key to WOE's performance is a novel form of self-supervised learning for open extractors -- using heuristic matches between Wikipedia infobox attribute values and corresponding sentences to construct training data. Like TextRunner, WOE's extractor eschews lexicalized features and handles an unbounded set of semantic relations. WOE can operate in two modes: when restricted to POS tag features, it runs as quickly as TextRunner, but when set to use dependency-parse features its precision and recall rise even higher.
Article
A wealth of on-line text information can be made available to automatic processing by information extraction (IE) systems. Each IE application needs a separate set of rules tuned to the domain and writing style. WHISK helps to overcome this knowledge-engineering bottleneck by learning text extraction rules automatically.WHISK is designed to handle text styles ranging from highly structured to free text, including text that is neither rigidly formatted nor composed of grammatical sentences. Such semi-structured text has largely been beyond the scope of previous systems. When used in conjunction with a syntactic analyzer and semantic tagging, WHISK can also handle extraction from free text such as news stories.
Conference Paper
Traditional Information Extraction (IE) takes a relation name and hand-tagged examples of that relation as input. Open IE is a relation- independent extraction paradigm that is tai- lored to massive and heterogeneous corpora such as the Web. An Open IE system extracts a diverse set of relational tuples from text with- out any relation-specific input. How is Open IE possible? We analyze a sample of English sentences to demonstrate that numerous rela- tionships are expressed using a compact set of relation-independent lexico-syntactic pat- terns, which can be learned by an Open IE sys- tem. What are the tradeoffs between Open IE and traditional IE? We consider this question in the context of two tasks. First, when the number of relations is massive, and the rela- tions themselves are not pre-specified, we ar- gue that Open IE is necessary. We then present a new model for Open IE called O-CRF and show that it achieves increased precision and nearly double the recall than the model em- ployed by TEXTRUNNER, the previous state- of-the-art Open IE system. Second, when the number of target relations is small, and their names are known in advance, we show that O-CRF is able to match the precision of a tra- ditional extraction system, though at substan- tially lower recall. Finally, we show how to combine the two types of systems into a hy- brid that achieves higher precision than a tra- ditional extractor, with comparable recall.