July 18, 2023 | BY opitz
Contact:
Daniel Deutsch
Rotem Dror
Juri Opitz
Dear Colleagues,
the 4th Workshop on Evaluation and Comparison for NLP systems (Eval4NLP), co-located at the 2023 Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics (AACL 2023), invites the submission of long and short papers, with a theoretical or experimental nature, describing recent advances in system evaluation and comparison in NLP.
** Important Dates **
All deadlines are 11.59 pm UTC -12h (“Anywhere on Earth”).
May 25, 2023 | BY Anya Belz
ReproNLP 2023: First Call for Participation
Background
May 16, 2023 | BY Eleftherios Avramidis
Contact:
Ondrej Bojar
Eleftherios Avramidis
Dear all,
the “Test suites” sub-task will be included for the sixth time in the General MT Shared Task of the Conference on Machine Translation (WMT23).
*OVERVIEW*
Test suites are custom extensions to the test sets of the General MT Shared Task, constructed so that they can focus on particular aspects of the MT output. They cοnsist of a source-side test-set and a customized evaluation service. As opposed to the standard evaluation process which produces generic quality scores, test suites often produce separate fine-grained results for each phenomenon.
February 13, 2023 | BY b.ross
August 08, 2022 | BY opitz
Contact:
Daniel Deutsch
Can Udomcharoenchaikit
Juri Opitz
The 3rd Workshop on Evaluation and Comparison for NLP systems (Eval4NLP, https://eval4nlp.github.io/2022), co-located at the 2022 Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics (AACL 2022, https://www.aacl2022.org/), invites the submission of long and short papers, with a theoretical or experimental nature, describing recent advances in system evaluation and comparison in NLP.
May 30, 2022 | BY opitz
Contact:
Daniel Deutsch
Juri Opitz
Can Udomcharoenchaikit
Fair evaluations and comparisons are of fundamental importance to the NLP community to properly track progress, especially within the current deep learning revolution, with new state-of-the-art results reported in ever shorter intervals. This concerns the creation of benchmark datasets that cover typical use cases and blind spots of existing systems, the designing of metrics for evaluating the performance of NLP systems on different dimensions, and the reporting of evaluation results in an unbiased manner.
February 11, 2022 | BY b.ross
Location:
Co-located with ICWSM 2022
Contact:
Björn Ross
Roberto Navigli
Agostina Calabrese
The automatic or semiautomatic analysis of textual data is a key approach to analyse the massive amounts of user-generated content online, from the identification of sentiment in text and topic classification to the detection of abusive language, misinformation or propaganda. However, the development of such systems faces a crucial challenge.
July 17, 2021 | BY plkumjorn
Event Dates:
10 Nov 2021 to 11 Nov 2021
Location:
in conjunction with EMNLP 2021
--------------------------------------------
Second Call for Papers
--------------------------------------------
The 2nd Workshop on Evaluation and Comparison for NLP systems (Eval4NLP), co-located at the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP2021), invites the submission of long and short papers, with a theoretical or experimental nature, describing recent advances in system evaluation and comparison in NLP.
July 05, 2021 | BY smille
Contact:
Kaustubh Dole
Sebastian Gehrmann
The GEM (Generation, Evaluation, Metrics) workshop at ACL, 2021 is inviting transformation submissions to NL-Augmenter.
June 14, 2021 | BY plkumjorn
Event Dates:
10 Nov 2021 to 11 Nov 2021
Location:
in conjunction with EMNLP 2021
--------------------------------------------
First Call for Papers
--------------------------------------------
The 2nd Workshop on Evaluation and Comparison for NLP systems (Eval4NLP), co-located at the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP2021), invites the submission of long and short papers, with a theoretical or experimental nature, describing recent advances in system evaluation and comparison in NLP.
Pages