Evaluation scenarios¶
The eHealth-KD proposes 3 evaluation scenarios to better understand the performance of different sytems: a main scenario where both subtasks are performed, and two additional scenarios for evaluating each subtask independently.
Scenario 1 is more complex than solving each optional scenario separately, since errors in subtask A will necessary translate to errors in subtask B. For this reason it is considered the main evaluation metric. Additionally, this scenario also provides the possibility for integrated end-to-end solutions that solve both subtask simultaneously.
Main Evaluation (Scenario 1)¶
This scenario evaluates all of the subtasks together as a pipeline. The input consists only of a plain text, and the expected output is a BRAT .ann
file with all the corresponding entities and relations found.
The measures will be precision, recall and F1 as follows:
$$Rec_{AB} = \frac{C_A + C_B + \frac{1}{2} P_A}{C_A + I_A + C_B + P_A + M_A + M_B} $$
$$Prec_{AB} = \frac{C_A + C_B + \frac{1}{2} P_A}{C_A + I_A + C_B + P_A + S_A + S_B} $$
$$F_{1AB} = 2 \cdot \frac{Prec_{AB} \cdot Rec_{AB}}{Prec_{AB} + Rec_{AB}} $$
F1 will determine the ranking of Scenario 1 and consequently of the eHealthKD challenge.
The exact definition of Correct, Missing, Spurious, Partial and Incorrect is presented in the following sections for each subtask.
Optional Subtask A (Scenario 2)¶
This scenario only evaluates Subtask A. The input is a plain text with several sentences and the output is a BRAT .ann
file with only entity annotations in it (relation annotations are ignored if present).
To compute the scores we define correct, partial, missing, incorrect and spurious matches. The expected and actual output files do not need to agree on the ID for each entity, nor on their order. The evaluator matches are based on the start and end of text spans and the corresponding type. A brief description about the metrics follows:
- Correct matches are reported when a text in the dev file matches exactly with a corresponding text span in the gold file in START and END values, and also the entity type. Only one correct match per entry in the gold file can be matched. Hence, duplicated entries will count as Spurious.
- Incorrect matches are reported when START and END values match, but not the type.
- Partial matches are reported when two intervals [START, END] have a non-empty intersection, such as the case of “vías respiratorias” and “respiratorias” in the previous example (and matching LABEL). Notice that a partial phrase will only be matched against a single correct phrase. For example, “tipo de cáncer” could be a partial match for both “tipo” and “cáncer”, but it is only counted once as a partial match with the word “tipo”. The word “cancer” is counted then as Missing. This aims to discourage a few large text spans that cover most of the document from getting a very high score.
- Missing matches are those that appear in the gold file but not in the dev file.
- Spurious matches are those that appear in the dev file but not in the gold file.
From these definitions, we compute precision, recall, and a standard F1 measure as follows:
$$Rec_{A} = \frac{C_A + \frac{1}{2} P_A}{C_A + I_A + P_A + M_A} $$
$$Prec_{A} = \frac{C_A + \frac{1}{2} P_A}{C_A + I_A + P_A + S_A} $$
$$F_{1A} = 2 \cdot \frac{Prec_{A} \cdot Rec_{A}}{Prec_{A} + Rec_{A}} $$
A higher precision means that the number of spurious identifications is smaller compared to the number of missing identifications, and a higher recall means the opposite. Partial matches are given half the score of correct matches, while missing and spurious identifications are given no score.
F1 will determine the ranking of Scenario 2.
Optional Subtask B (Scenario 3)¶
This scenario only evaluates Subtask B. The input is plain text and a corresponding .ann
file with the correct entities annotated.
The expected output is a .ann
file with both the entities and relations. Feel free to simply copy the entity annotations from the provided .ann
file and append the relation annotations.
To compute the scores we define correct, missing, and spurious matches. The expected and actual output files do not need to agree on the ID for each relation (which is ignored) nor on their order. The evaluator matches are based on the start and end of text spans and the corresponding type. A brief description about the metrics follows:
- Correct: relationships that matched exactly to the gold file, including the type and the corresponding IDs for each of the participants.
- Missing: relationships that are in the gold file but not in the dev file, either because the type is wrong, or because one of the IDs didn’t match.
- Spurious: relationships that are in the dev file but not in the gold file, either because the type is wrong, or because one of the IDs didn’t match.
We define standard precision, recall and F1 metrics as follows:
$$Rec_{B} = \frac{C_B}{C_B + M_B} $$
$$Prec_{B} = \frac{C_B}{C_B + S_B} $$
$$F_{1B} = 2 \cdot \frac{Prec_{B} \cdot Rec_{B}}{Prec_{B} + Rec_{B}} $$
F1 will determine the ranking of Scenario 3.