Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/conll/reference-coreference-scorers
This is the reference implementation of commonly used coreference metrics.
https://github.com/conll/reference-coreference-scorers
Last synced: 4 months ago
JSON representation
This is the reference implementation of commonly used coreference metrics.
- Host: GitHub
- URL: https://github.com/conll/reference-coreference-scorers
- Owner: conll
- License: other
- Created: 2015-03-11T21:11:58.000Z (almost 10 years ago)
- Default Branch: master
- Last Pushed: 2018-04-13T08:35:05.000Z (almost 7 years ago)
- Last Synced: 2024-08-02T16:54:38.628Z (7 months ago)
- Language: Perl
- Homepage: http://conll.github.io/reference-coreference-scorers
- Size: 167 KB
- Stars: 73
- Watchers: 5
- Forks: 23
- Open Issues: 10
-
Metadata Files:
- Readme: README.rst
- License: LICENSE
Awesome Lists containing this project
README
Reference Coreference Scorer
============================DESCRIPTION
-----------This is the official implementation of the revised coreference scorer
used for CoNLL-2011/2012 shared tasks on coreference resolution. It
addresses issues that prevented the consistent scoring of predicted
mentions in the past.VERSION
-------The current stable (official) version for scoring predicted mentions is **v8.01**
CITATION
--------We would appreciate if you cite the paper when you use this scorer as
some of us are academics or wanting to be academics, and citations
matter.::
@InProceedings{pradhan-EtAl:2014:P14-2,
author = {Pradhan, Sameer and Luo, Xiaoqiang and Recasens, Marta and Hovy, Eduard and Ng, Vincent and Strube, Michael},
title = {Scoring Coreference Partitions of Predicted Mentions: A Reference Implementation},
booktitle = {Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},
month = {June},
year = {2014},
address = {Baltimore, Maryland},
publisher = {Association for Computational Linguistics},
pages = {30--35},
url = {http://www.aclweb.org/anthology/P14-2006}
}USAGE
-----::
perl scorer.pl []
: the metric desired to score the results. one of the following values:
muc: MUCScorer (Vilain et al, 1995)
bcub: B-Cubed (Bagga and Baldwin, 1998)
ceafm: CEAF (Luo et al., 2005) using mention-based similarity
ceafe: CEAF (Luo et al., 2005) using entity-based similarity
blanc: BLANC (Luo et al., 2014) BLANC metric for gold and predicted mentions
all: uses all the metrics to score: file with expected coreference chains in CoNLL-2011/2012 format
: file with output of coreference system (CoNLL-2011/2012 format)
: optional. The name of the document to score. If name is not
given, all the documents in the dataset will be scored. If given
name is "none" then all the documents are scored but only total
results are shown.OUTPUT
------The score subroutine returns an array with four values in this order:
Coreference Score
~~~~~~~~~~~~~~~~~::
Recall = recall_numerator / recall_denominator
Precision = precision_numerator / precision_denominator
F1 = 2 * Recall * Precision / (Recall + Precision)These values are to standard output when variable ``$VERBOSE`` is not null.
Identification of Mentions
~~~~~~~~~~~~~~~~~~~~~~~~~~A score for identification of mentions (recall, precision and F1) is
also included. Mentions from system response are compared with key
mentions. This version performs strict mention matching as was used in
the CoNLL-2011 and 2012 shared tasks.AUTHORS
-------* Emili Sapena, Universitat Politècnica de Catalunya, http://www.lsi.upc.edu/~esapena, esapena lsi.upc.edu
* Sameer Pradhan, http://cemantix.org, pradhan cemantix.org
* Sebastian Martschat, sebastian.martschat h-its.org
* Xiaoqiang Luo, xql google.comCOPYRIGHT
---------::
2009-2011, Emili Sapena esapena lsi.upc.edu
2011- Sameer Pradhan pradhan cemantix.org