Difference between revisions of "RTE6 - Ablation Tests"

From ACL Wiki
Jump to navigation Jump to search
Line 281: Line 281:
 
| style="text-align: center;"| The resource that has been ablated is VerbOcean.
 
| style="text-align: center;"| The resource that has been ablated is VerbOcean.
  
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Wikipedia
 +
| UAIC20101_abl-1
 +
| style="text-align: center;"| 1.93
 +
| style="text-align: center;"| Run 1 with no BK
  
 +
|- bgcolor="#ECECEC" "align="left"
 +
| DIRT
 +
| UAIC20101_abl-2
 +
| style="text-align: center;"| -1.56
 +
| style="text-align: center;"| Run 1 without DIRT.
  
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Wikipedia
 +
| UAIC20102_abl-1
 +
| style="text-align: center;"| 1.08
 +
| style="text-align: center;"| Run 2 with no BK
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| DIRT
 +
| UAIC20102_abl-2
 +
| style="text-align: center;"| -0.72
 +
| style="text-align: center;"| Run 2 without DIRT.
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| Wikipedia
 +
| UAIC20103_abl-1
 +
| style="text-align: center;"| 1.3
 +
| style="text-align: center;"| Run 3 with no BK
 +
 +
|- bgcolor="#ECECEC" "align="left"
 +
| DIRT
 +
| UAIC20103_abl-2
 +
| style="text-align: center;"| -0.98
 +
| style="text-align: center;"| Run 3 without DIRT.
 
|}
 
|}
 
<br>
 
<br>

Revision as of 06:25, 3 February 2011

The following table lists the results of the ablation tests (a mandatory track since the RTE5 campaign), submitted by participants to RTE6 .


Participants are kindly invited to check if all the inserted information is correct and complete.


Ablated Component Ablation Run[1] Resource impact - F1 Resource Usage Description
WordNet BIU1_abl-1 0.9 No Word-Net. On Dev set: 39.18% (compared to 40.73% when WN is used)
CatVar BIU1_abl-2 0.63 No CatVar. On Dev set achieved about 40.20% (compared to 40.73% when CatVar is used)
Coreference resolver BIU1_abl-3 -0.88 No coreference resolver

On Dev set 41.62% (Compared to 40.73% when Coreference resolver is used). This ablation test is an unusual ablation test, since it shows that the co-reference resolution component has a negative impact.

DIRT Boeing1_abl-1 3.97 DIRT removed
WordNet Boeing1_abl-2 4.42 No WordNet
Name Normalization budapestcad2_abl-2 0.65 no name normalization was performed (e.g. George W. Bush -> Bush).
Named Entities Recognition budapestcad2_abl-3 -1.23 no NER
WordNet budapestcad2_abl-4 -1.11 No WordNet. (In the original run, WordNet was used to find the synonyms of words in the triplets, and additional triplets were generated from all possible combinations.)
WordNet deb_iitb1_abl-1 8.68 Wordnet is albated in this test.No change of code required only wordnet module is removed while matching.
VerbOcean deb_iitb1_abl-2 1.87 VerbOcean is albated in this test.No change of code required only VerbOcean module is removed while matching.
WordNet deb_iitb2_abl-1 7.9 Wordnet is albated in this test.No change of code required only wordnet module is removed while matching.
VerbOcean deb_iitb2_abl-2 0.94 VerbOcean is albated in this test.No change of code required only VerbOcean module is removed while matching.
WordNet deb_iitb3_abl-1 11.43 Wordnet is albated in this test. No change of code required only wordnet module is removed while matching.
WordNet deb_iitb3_abl-2 2.54 VerbOcean is albated in this test.No change of code required only VerbOcean module is removed while matching.
POS-Tagger DFKI1_abl-4 4.99 No wordform/POS-tags included for the comparison.
POS-Tagger DFKI1_abl-6 2.22 No named entity recognition for the comparison.
WordNet DFKI1_abl-7 -0.23 No WordNet similarity for the comparison.
Coreference resolver DFKI1_Main -1.54 Coreference resolution used for the comparison.
WordNet DirRelCond23_abl-1 8.43 WordNet removed. Only basic word comparison used instead of word relations.
Wikipedia FBK_irst3_Main -23.91 This run is produced by the system configuration for run3 and uses rules extracted from Wikipedia
Wikipedia FBK_irst3_Main -3.58 This run is produced by the system configuration for run3 and uses rules extracted from Wikipedia with probability above 0.7
Proximity similarity dictionary of Dekang Lin FBK_irst3_Main -7.79 This run is produced by the system configuration for run3 and uses rules extracted from proximity similarity dictionary of Dekang Lin
WordNet FBK_irst3_Main -3.21 This run is produced by the system configuration for run3 and uses rules extracted from WordNet
WordNet FBK_irst3_Main -2.08 This run is produced by the system configuration for run3 and uses rules extracted from WordNet with probability above 0.7
VerbOcean FBK_irst3_Main -4 This run is produced by the system configuration for run3 and uses rules extracted from Verbocean
Dependency similarity dictionary of Dekang Lin FBK_irst3_Main -13.56 This run is produced by the system configuration for run3 and uses rules extracted from dependency similarity dictionary of Dekang Lin
Dictionary of Named Entities Acronyms and Synonyms IKOMA2_abl-3 -0.76 Remove synonym dictionaries: as acronym dictionary constructed automatically from the corpus and a synonym dictionary that contains geographical terms.
WordNet JU_CSE_TAC1_abl-1 13.29 The Run-1 is based on the composition of lexical based RTE methods and Syntactic RTE Method. The lexical based RTE methods are: WordNet based unigram match, bigram match, longest common sub-sequence, skip-gram and stemming. Here we ablated the WordNet based unigram match only.
WordNet JU_CSE_TAC2_abl-1 10.19 The Run-2 is based on the composition of lexical based RTE methods, Syntactic RTE Method, Chunk and Named Entity Methods. The lexical based RTE methods are: WordNet based unigram match, bigram match, longest common sub-sequence, skip-gram and stemming. Here we ablated the WordNet based unigram match only.
WordNet JU_CSE_TAC3_abl-1 3.86 The Run-3 is based on the Support Vector Machine that uses twenty five features for lexical similarity, the output tag from a rule based syntactic two-way TE system as feature, and output from a rule based Chunk Module and Named Entity Module. The important lexical features that are used in the present system are: WordNet based unigram match, bigram match, longest common sub-sequence, skip-gram, stemming and lexical distance (17 features). Here we ablated the WordNet based unigram match only.
LingPipe co-reference PKUTM2_abl-1 0.17 Lingpipe co-reference are removed, the experiment was based on named-entity, wordnet, verbocean
VerbOcean PKUTM2_abl-2 1.02 Verbocean are removed, the experiment was based on named-entity, wordnet, co-reference
LingPipe Named Entities PKUTM2_abl-3 13.84 Lingpipe named-entity are removed, the experiment was based on wordnet, co-reference, verbocean
WordNet saicnlp1_abl-1 -0.02 Ablation run, with WordNet stubbed.
Shalmanesar Parser Sangyan1_abl-1 -1.76 This ablation run was executed with the Shalmanesar parser ablated from our system. Our system uses the Shalmanesar Parser for Framenet frame assignment.

Our system required minor tweaking of the source code in order ablate Shalmanesar Parser.

WordNet Sangyan1_abl-2 -0.39 This ablation run was executed with the WordNet ablated from our system. Our system uses the Wordnet for Word mathcing.

Our system required minor tweaking of the source code in order ablate Wordnet.

VerbOcean Sangyan1_abl-3 0.07 This ablation run was executed with the Verb Ocean ablated from our system. Our system uses the Verb Ocean resource for detection of antonyms. Our system required minor tweaking of the source code in order ablate Verb Ocean resource.
Named Entities Recognition SINAI1_abl-1 7.62 Only PPVs on WN 1.7
Wordnet SINAI1_abl-2 15.87 Only NEs on WN 1.7
Named Entities Recognition SINAI2_abl-1 20.25 Only PPVs on WN 3.0
Wordnet SINAI2_abl-2 18.28 Only NEs on WN 3.0
Wordnet SJTU_CIT3_abl-1 0.03 The resource that has been ablated is WordNet.
Wikipedia SJTU_CIT3_abl-2 4.7 The resource that has been ablated is Wikipedia.
VerbOcean SJTU_CIT3_abl-3 -1.15 The resource that has been ablated is VerbOcean.
Wikipedia UAIC20101_abl-1 1.93 Run 1 with no BK
DIRT UAIC20101_abl-2 -1.56 Run 1 without DIRT.
Wikipedia UAIC20102_abl-1 1.08 Run 2 with no BK
DIRT UAIC20102_abl-2 -0.72 Run 2 without DIRT.
Wikipedia UAIC20103_abl-1 1.3 Run 3 with no BK
DIRT UAIC20103_abl-2 -0.98 Run 3 without DIRT.


Footnotes

  1. For further information about participants, click here: RTE Challenges - Data about participants


   Return to RTE Knowledge Resources