You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is an experimental model intended for research and evaluation.
Factors
The training data is based on text extracts from the Wall Street Journal, an American business-focused, English-language international daily newspaper based in New York City. It may not perform as well with text from other domains.
The Coreference Resolver was trained and evaluated using equivalent partitions of the CoNLL-2011 Shared Task dataset which covers a subset of the data in PARC 3.0.
The performance of this model is reasonably good but there is a significant error rate. Extracted quotes should always be checked against the original text to confirm the accuracy of the text spans and correctness of the attribution.
Quantitative Analysis
Overall Performance
The overall performance of Citron using this model was measured using the Citron Evaluate script.
Cue Span
Exact Metric
Score
Precision
93.3%
Recall
73.6%
F1
82.3%
Overlap Metric
Score
Precision
96.5%
Recall
63.9%
F1
76.8%
Source Spans
Exact Metric
Score
Precision
92.9%
Recall
73.3%
F1
82.0%
Overlap Metric
Score
Precision
97.4%
Recall
75.8%
F1
85.3%
Content Spans
Exact Metric
Score
Precision
67.3%
Recall
53.1%
F1
59.3%
Overlap Metric
Score
Precision
93.2%
Recall
75.3%
F1
83.3%
All Quote Spans
Exact Metric
Score
Precision
64.3%
Recall
50.7%
F1
56.7%
Overlap Metric
Score
Precision
74.7%
Recall
93.8%
F1
83.2%
Performance of the Individual Components
The performance of the individual components of Citron using this model was measured using the build scripts.