-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
assessment of the quality of ERG analyses #32
Comments
The mismatches can be grouped and quantified. Note that most cases are expected if we consider ERG semantics representation.
|
From all the MRS that I obtained from the sentences that were parsed by ERG, 3748 did no pass in the validations from https://pydelphin.readthedocs.io/en/latest/api/delphin.mrs.html#module-functions
Some cases below @danflick: Example of not connected:
but the 3th reading is No intrinsic variable property and not well-formed (probably the error is e2 or x4):
Using a recall procedure, I obtained a valid MRS for the majority of the 3748 cases above and inspect the next readings. In 77 cases, the valid MRS is the 6th reading. In 1686 cases, the second reading is valid etc.
|
In cases where a derivational suffix (such as re, co, un) shifts the meaning, we need a verb with the suffix: revisit (third-person singular simple present revisits, present participle revisiting, simple past and past participle revisited)
|
Executive summary:
Considering the spans that group predications and tokens for each sentence. In total, we have 1842193 such groups. In only 49793 of them, I found apparent POS inconsistency between ERG and the sense annotation.
49793/1842193 = 0.02
Note that I only consider the tokens that were sense tagged. If we count per sentence, 38883 sentences contain at least one error from a total of 159614 sentences. If we ignore the mismatches a/r (adverbs as adjectives) and q/n (someone), we have 28358 sentences with at least one error. If we also ignore mismatches caused by verb/adjective we have 17401 sentences:
38883/159614 = 0.24
28358/159614 = 0.17
17401/159614 = 0.11
The dataset contains 165994 sentences, but not all of them got a parse from ERG.
Details:
For all sentences, I join the tokens with the MRS predicates using the spans.
Below I found no conflict between ERG and the annotation. For instance, affect%2 means it was annotated as a verb, and ERG made it the predicate _affect_v_1. For hydrarthrosis, it was annotated as a noun, and ERG preprocessing instantiated a generic token from NNS pos tagger.
Next, excess was annotated as an adjective (%5) but analysed as NOUN by ERG. See the line starting with “D>"
ERG annotated adverbs and adjectives as adjoins, so another common mismatch is a vs r. The fragment after the first semi-colon should be an example "equally balanced”?
Adjective vs verb:
Someone vs person+some_q. (1829 cases), I need to improve my check to remove this from the suspicious cases.
What is especially below? Tagged as an adverb, in the ERG analysis, it is X?
The text was updated successfully, but these errors were encountered: