Difference between revisions of "RTE6 - Ablation Tests"

From ACL Wiki
Jump to navigation Jump to search
Line 116: Line 116:
 
| style="text-align: center;"| -0.23
 
| style="text-align: center;"| -0.23
 
| style="text-align: center;"| No WordNet similarity for the comparison.
 
| style="text-align: center;"| No WordNet similarity for the comparison.
 +
 +
|- bgcolor="#FFFFF" "align="left"
 +
| Coreference resolver
 +
| DFKI1_Main
 +
| style="text-align: center;"| -1.54
 +
| style="text-align: center;"| Coreference resolution used for the comparison.
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| WordNet
 +
| DirRelCond23_abl-1
 +
| style="text-align: center;"| 8.43
 +
| style="text-align: center;"| WordNet removed. Only basic word comparison used instead of word relations.
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Wikipedia
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -23.91
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from Wikipedia
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Wikipedia
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -3.58
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from Wikipedia with probability above 0.7
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Proximity similarity dictionary of Dekang Lin
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -7.79
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from proximity similarity dictionary of Dekang Lin
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| WordNet
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -3.21
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from WordNet
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| WordNet
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -2.08
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from WordNet with probability above 0.7
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| VerbOcean
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -4
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from Verbocean
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Dependency similarity dictionary of Dekang Lin
 +
| FBK_irst3_Main
 +
| style="text-align: center;"| -13.56
 +
| style="text-align: center;"| This run is produced by the system configuration for run3 and uses rules extracted from dependency similarity dictionary of Dekang Lin
 +
 +
 +
 
|}
 
|}
 
<br>
 
<br>

Revision as of 04:40, 3 February 2011

The following table lists the results of the ablation tests (a mandatory track since the RTE5 campaign), submitted by participants to RTE6 .


Participants are kindly invited to check if all the inserted information is correct and complete.


Ablated Component Ablation Run[1] Resource impact - F1 Resource Usage Description
WordNet BIU1_abl-1 0.9 No Word-Net. On Dev set: 39.18% (compared to 40.73% when WN is used)
CatVar BIU1_abl-2 0.63 No CatVar. On Dev set achieved about 40.20% (compared to 40.73% when CatVar is used)
Coreference resolver BIU1_abl-3 -0.88 No coreference resolver

On Dev set 41.62% (Compared to 40.73% when Coreference resolver is used). This ablation test is an unusual ablation test, since it shows that the co-reference resolution component has a negative impact.

DIRT Boeing1_abl-1 3.97 DIRT removed
WordNet Boeing1_abl-2 4.42 No WordNet
Name Normalization budapestcad2_abl-2 0.65 no name normalization was performed (e.g. George W. Bush -> Bush).
Named Entities Recognition budapestcad2_abl-3 -1.23 no NER
WordNet budapestcad2_abl-4 -1.11 No WordNet. (In the original run, WordNet was used to find the synonyms of words in the triplets, and additional triplets were generated from all possible combinations.)
WordNet deb_iitb1_abl-1 8.68 Wordnet is albated in this test.No change of code required only wordnet module is removed while matching.
VerbOcean deb_iitb1_abl-2 1.87 VerbOcean is albated in this test.No change of code required only VerbOcean module is removed while matching.
WordNet deb_iitb2_abl-1 7.9 Wordnet is albated in this test.No change of code required only wordnet module is removed while matching.
VerbOcean deb_iitb2_abl-2 0.94 VerbOcean is albated in this test.No change of code required only VerbOcean module is removed while matching.
WordNet deb_iitb3_abl-1 11.43 Wordnet is albated in this test. No change of code required only wordnet module is removed while matching.
WordNet deb_iitb3_abl-2 2.54 VerbOcean is albated in this test.No change of code required only VerbOcean module is removed while matching.
POS-Tagger DFKI1_abl-4 4.99 No wordform/POS-tags included for the comparison.
POS-Tagger DFKI1_abl-6 2.22 No named entity recognition for the comparison.
WordNet DFKI1_abl-7 -0.23 No WordNet similarity for the comparison.
Coreference resolver DFKI1_Main -1.54 Coreference resolution used for the comparison.
WordNet DirRelCond23_abl-1 8.43 WordNet removed. Only basic word comparison used instead of word relations.
Wikipedia FBK_irst3_Main -23.91 This run is produced by the system configuration for run3 and uses rules extracted from Wikipedia
Wikipedia FBK_irst3_Main -3.58 This run is produced by the system configuration for run3 and uses rules extracted from Wikipedia with probability above 0.7
Proximity similarity dictionary of Dekang Lin FBK_irst3_Main -7.79 This run is produced by the system configuration for run3 and uses rules extracted from proximity similarity dictionary of Dekang Lin
WordNet FBK_irst3_Main -3.21 This run is produced by the system configuration for run3 and uses rules extracted from WordNet
WordNet FBK_irst3_Main -2.08 This run is produced by the system configuration for run3 and uses rules extracted from WordNet with probability above 0.7
VerbOcean FBK_irst3_Main -4 This run is produced by the system configuration for run3 and uses rules extracted from Verbocean
Dependency similarity dictionary of Dekang Lin FBK_irst3_Main -13.56 This run is produced by the system configuration for run3 and uses rules extracted from dependency similarity dictionary of Dekang Lin



Footnotes

  1. For further information about participants, click here: RTE Challenges - Data about participants


   Return to RTE Knowledge Resources