For the Hidden Markov Model (HMM) tagger, the parameters used to tune the tagger include
the Lidstone, MLE, ELE, and LaPlace probability distribution estimators. Using these various
probability distribution estimators can produce a more robust HMM tagger, as the performance
of the probability distribution estimators is dependent on the training dataset. By comparing
the performance of these estimators, we can determine the best estimator to choose for
calculating the accuracy. The parameters that yield the best performance for the HMM tagger
are the Lidstone probability distribution estimates, with a gamma value of 0.00001.
For the Brill tagger, the parameters used to tune the tagger include three template collections.
These are composed of the nltkdemo18, fntbl37, and brill24 template collections. This
influenced the accuracy of the tagger because different templates capture different linguistic
nuances. The parameters that yield the best performance for the Brill tagger is the brill24
template collection, with a max rule count of 200.
Tagger | Input (dataset used) | Accuracy of the Tagger |
---|---|---|
HMM | in-domain test set | 81.05 |
HMM | out-of-domain test set | 78.68 |
Brill | in-domain test set | 72.45 |
Brill | out-of-domain test set | 67.71 |
HMM Misclassifications | Brill Misclassifications |
---|---|
1. Misclassifies “‘s” as POS instead of VBZ. The possessive marker is highly context-dependent, and having contractions in datasets can be hard for taggers to discern. 2. Misclassifies many out-of-vocabulary (OOV) words as NN. This is most likely due to the tagger recognizing nthat the NN tag appears many times throughout the dataset, assigning it a higher transition probability. Because of the higher transition probability, OOV defaults to a NN tag. This can be a form of bias the HMM tagger has learned. |
1. Misclassifies verbs as nouns (NN) in sentences where the tag of the verb is highly context-dependent, or the verb is at the start of a sentence. Other words are also misclassified as NN if the rules are not in the baseline rules. This led to overclassification of NN in the data. - Example: Lines 1, 15, and 102 in test_brill.txt are VB missclassified as NN. 2. Misclassifies “‘s” as POS instead of VBZ. The possessive marker is highly context-dependent, and having contractions in datasets can be hard for taggers to discern. 3. All NNP are misclassified as NN. This is most likely due to NNP not being an initial rule, leading to misclassification. |
Similarities Between HMM Tagger and Brill Tagger Misclassifications | Differences Between HMM |
---|---|
● Both taggers misclassify the “‘s” possessive marker as POS instead of VBZ. Because there is much contextual ambiguity surrounding possessive markers/contractions interms of tagging, it is expected that both the HMM and Brill tagger would struggle to correctly classify these. ● Both taggers misclassify many out-of-vocabulary (OOV) words as NN, leading to overclassification of NN in all datasets. |
● The difference between misclassifications for in-domain vs. out-of-domain data was higher for Brill tagging in comparison to HMM tagging. Brill tagging seems to struggle more with out-of-domain data, which can be due to HMM being able to make more contextual inferences based on the sentence a word is located in. This is most likely why more NN misclassification is present in the out-of-domain Brill-tagged dataset. |