Locked History Actions

Diff for "PolEval"

Differences between revisions 2 and 6 (spanning 4 versions)
Revision 2 as of 2017-09-25 10:22:17
Size: 1266
Comment:
Revision 6 as of 2018-06-04 16:47:03
Size: 1327
Comment:
Deletions are marked like this. Additions are marked like this.
Line 2: Line 2:
[[http://www.poleval.pl|PolEval] is a !SemEval-inspired evaluation campaign for natural language processing tools for Polish. Submitted tools compete against one another within certain tasks selected by organisers, using available data and are evaluated according to pre-established procedures. [[http://www.poleval.pl|PolEval]] is a !SemEval-inspired evaluation campaign for natural language processing tools for Polish. Submitted tools compete against one another within certain tasks selected by organisers, using available data and are evaluated according to pre-established procedures.
Line 4: Line 4:
[[http://www.poleval.pl|Official website]] [[http://poleval.pl|Official website]]
Line 10: Line 10:
This is a corrected version of corpus created for the evaluation of Task 1 during PolEval 2017 competition. It contains ca. 55k segments, annotated manually by two qualified linguists. The annotation has been conducted in two phases: in the first phase the source raw text --- coming from the Polish Coreference Corpus --- has been annotated in parallel a) manually, by two qualified linguists; b) automatically, using the most recent version of the Concraft tagger, trained on the hand-annotated portion of the NCP. In the second phase, the differences between human annotators and the tagger have been found and cross-corrected by the annotator, which has not previously worked on this text part. This is a corrected version of the corpus created for the evaluation of Task 1 during PolEval 2017 competition. It contains 54 906 segments, annotated manually by two qualified linguists. The annotation has been conducted in two phases: in the first phase the source raw text coming from the Polish Coreference Corpus has been annotated in parallel a) manually, by two qualified linguists; b) automatically, using the most recent version of the Concraft tagger, trained on the hand-annotated portion of the NCP. In the second phase, the differences between human annotators and the tagger have been found and cross-corrected by the annotator, which has not previously worked on this text part.
Line 12: Line 12:
 * [[attachment:poleval-2017-task-1ab-gold-2.0-tei.tar.gz]]

PolEval

PolEval is a SemEval-inspired evaluation campaign for natural language processing tools for Polish. Submitted tools compete against one another within certain tasks selected by organisers, using available data and are evaluated according to pre-established procedures.

Official website

Below you can find language resources associated with PolEval, which may be useful for other projects.

PolEval 2017 POS Tagging Shared Task

This is a corrected version of the corpus created for the evaluation of Task 1 during PolEval 2017 competition. It contains 54 906 segments, annotated manually by two qualified linguists. The annotation has been conducted in two phases: in the first phase the source raw text – coming from the Polish Coreference Corpus – has been annotated in parallel a) manually, by two qualified linguists; b) automatically, using the most recent version of the Concraft tagger, trained on the hand-annotated portion of the NCP. In the second phase, the differences between human annotators and the tagger have been found and cross-corrected by the annotator, which has not previously worked on this text part.