language:
- en
paperswithcode_id: conll-2000-1
pretty_name: CoNLL-2000
dataset_info:
features:
- name: id
dtype: string
- name: tokens
sequence: string
- name: pos_tags
sequence:
class_label:
names:
'0': ''''''
'1': '#'
'2': $
'3': (
'4': )
'5': ','
'6': .
'7': ':'
'8': '``'
'9': CC
'10': CD
'11': DT
'12': EX
'13': FW
'14': IN
'15': JJ
'16': JJR
'17': JJS
'18': MD
'19': NN
'20': NNP
'21': NNPS
'22': NNS
'23': PDT
'24': POS
'25': PRP
'26': PRP$
'27': RB
'28': RBR
'29': RBS
'30': RP
'31': SYM
'32': TO
'33': UH
'34': VB
'35': VBD
'36': VBG
'37': VBN
'38': VBP
'39': VBZ
'40': WDT
'41': WP
'42': WP$
'43': WRB
- name: chunk_tags
sequence:
class_label:
names:
'0': O
'1': B-ADJP
'2': I-ADJP
'3': B-ADVP
'4': I-ADVP
'5': B-CONJP
'6': I-CONJP
'7': B-INTJ
'8': I-INTJ
'9': B-LST
'10': I-LST
'11': B-NP
'12': I-NP
'13': B-PP
'14': I-PP
'15': B-PRT
'16': I-PRT
'17': B-SBAR
'18': I-SBAR
'19': B-UCP
'20': I-UCP
'21': B-VP
'22': I-VP
splits:
- name: train
num_bytes: 5356965
num_examples: 8937
- name: test
num_bytes: 1201151
num_examples: 2013
download_size: 3481560
dataset_size: 6558116
Dataset Card for "conll2000"
Table of Contents
- Dataset Description
- Dataset Structure
- Dataset Creation
- Considerations for Using the Data
- Additional Information
Dataset Description
- Homepage: https://www.clips.uantwerpen.be/conll2000/chunking/
- Repository: More Information Needed
- Paper: More Information Needed
- Point of Contact: More Information Needed
- Size of downloaded dataset files: 3.48 MB
- Size of the generated dataset: 6.55 MB
- Total amount of disk used: 10.03 MB
Dataset Summary
Text chunking consists of dividing a text in syntactically correlated parts of words. For example, the sentence He reckons the current account deficit will narrow to only # 1.8 billion in September . can be divided as follows: [NP He ] [VP reckons ] [NP the current account deficit ] [VP will narrow ] [PP to ] [NP only # 1.8 billion ] [PP in ] [NP September ] .
Text chunking is an intermediate step towards full parsing. It was the shared task for CoNLL-2000. Training and test data for this task is available. This data consists of the same partitions of the Wall Street Journal corpus (WSJ) as the widely used data for noun phrase chunking: sections 15-18 as training data (211727 tokens) and section 20 as test data (47377 tokens). The annotation of the data has been derived from the WSJ corpus by a program written by Sabine Buchholz from Tilburg University, The Netherlands.
Supported Tasks and Leaderboards
Languages
Dataset Structure
Data Instances
conll2000
- Size of downloaded dataset files: 3.48 MB
- Size of the generated dataset: 6.55 MB
- Total amount of disk used: 10.03 MB
An example of 'train' looks as follows.
This example was too long and was cropped:
{
"chunk_tags": [11, 13, 11, 12, 21, 22, 22, 22, 22, 11, 12, 12, 17, 11, 12, 13, 11, 0, 1, 13, 11, 11, 0, 21, 22, 22, 11, 12, 12, 13, 11, 12, 12, 11, 12, 12, 0],
"id": "0",
"pos_tags": [19, 14, 11, 19, 39, 27, 37, 32, 34, 11, 15, 19, 14, 19, 22, 14, 20, 5, 15, 14, 19, 19, 5, 34, 32, 34, 11, 15, 19, 14, 20, 9, 20, 24, 15, 22, 6],
"tokens": "[\"Confidence\", \"in\", \"the\", \"pound\", \"is\", \"widely\", \"expected\", \"to\", \"take\", \"another\", \"sharp\", \"dive\", \"if\", \"trade\", \"figur..."
}
Data Fields
The data fields are the same among all splits.
conll2000
id
: astring
feature.tokens
: alist
ofstring
features.pos_tags
: alist
of classification labels, with possible values including''
(0),#
(1),$
(2),(
(3),)
(4).chunk_tags
: alist
of classification labels, with possible values includingO
(0),B-ADJP
(1),I-ADJP
(2),B-ADVP
(3),I-ADVP
(4).
Data Splits
name | train | test |
---|---|---|
conll2000 | 8937 | 2013 |
Dataset Creation
Curation Rationale
Source Data
Initial Data Collection and Normalization
Who are the source language producers?
Annotations
Annotation process
Who are the annotators?
Personal and Sensitive Information
Considerations for Using the Data
Social Impact of Dataset
Discussion of Biases
Other Known Limitations
Additional Information
Dataset Curators
Licensing Information
Citation Information
@inproceedings{tksbuchholz2000conll,
author = "Tjong Kim Sang, Erik F. and Sabine Buchholz",
title = "Introduction to the CoNLL-2000 Shared Task: Chunking",
editor = "Claire Cardie and Walter Daelemans and Claire
Nedellec and Tjong Kim Sang, Erik",
booktitle = "Proceedings of CoNLL-2000 and LLL-2000",
publisher = "Lisbon, Portugal",
pages = "127--132",
year = "2000"
}