annotations_creators:
- expert-generated
language_creators:
- expert-generated
language:
- afr
- af
- aaa
- abc
- ada
- adq
- aeu
- agq
- ags
- ahk
- aia
- ajz
- aka
- ak
- ame
- amh
- am
- amp
- amu
- ann
- aph
- awa
- awb
- azn
- azo
- bag
- bam
- bm
- baw
- bax
- bbk
- bcc
- bce
- bec
- bef
- ben
- bn
- bfd
- bfm
- bfn
- bgf
- bho
- bhs
- bis
- bi
- bjn
- bjr
- bkc
- bkh
- bkm
- bkx
- bob
- bod
- bo
- boz
- bqm
- bra
- brb
- bri
- brv
- bss
- bud
- buo
- bwt
- bwx
- bxa
- bya
- bze
- bzi
- cak
- cbr
- ceb
- cgc
- chd
- chp
- cim
- clo
- cmn
- zh
- cmo
- csw
- cuh
- cuv
- dag
- ddg
- ded
- deu
- de
- dig
- dje
- dmg
- dnw
- dtp
- dtr
- dty
- dug
- eee
- ekm
- enb
- enc
- eng
- en
- ewo
- fas
- fa
- fil
- fli
- fon
- fra
- fr
- fub
- fuh
- gal
- gbj
- gou
- gsw
- guc
- guj
- gu
- guz
- gwc
- hao
- hat
- ht
- hau
- ha
- hbb
- hig
- hil
- hin
- hi
- hla
- hna
- hre
- hro
- idt
- ilo
- ind
- id
- ino
- isu
- ita
- it
- jgo
- jmx
- jpn
- ja
- jra
- kak
- kam
- kan
- kn
- kau
- kr
- kbq
- kbx
- kby
- kek
- ken
- khb
- khm
- km
- kik
- ki
- kin
- rw
- kir
- ky
- kjb
- kmg
- kmr
- ku
- kms
- kmu
- kor
- ko
- kqr
- krr
- ksw
- kur
- ku
- kvt
- kwd
- kwu
- kwx
- kxp
- kyq
- laj
- lan
- lao
- lo
- lbr
- lfa
- lgg
- lgr
- lhm
- lhu
- lkb
- llg
- lmp
- lns
- loh
- lsi
- lts
- lug
- lg
- luy
- lwl
- mai
- mal
- ml
- mam
- mar
- mr
- mdr
- mfh
- mfj
- mgg
- mgm
- mgo
- mgq
- mhx
- miy
- mkz
- mle
- mlk
- mlw
- mmu
- mne
- mnf
- mnw
- mot
- mqj
- mrn
- mry
- msb
- muv
- mve
- mxu
- mya
- my
- myk
- myx
- mzm
- nas
- nco
- nep
- ne
- new
- nge
- ngn
- nhx
- njy
- nla
- nld
- nl
- nlv
- nod
- nsk
- nsn
- nso
- nst
- nuj
- nwe
- nwi
- nxa
- nxl
- nya
- ny
- nyo
- nyu
- nza
- odk
- oji
- oj
- oki
- omw
- ori
- or
- ozm
- pae
- pag
- pan
- pa
- pbt
- pce
- pcg
- pdu
- pea
- pex
- pis
- pkb
- pmf
- pnz
- por
- pt
- psp
- pwg
- qaa
- qub
- quc
- quf
- quz
- qve
- qvh
- qvm
- qvo
- qxh
- rel
- rnl
- ron
- ro
- roo
- rue
- rug
- rus
- ru
- san
- sa
- saq
- sat
- sdk
- sea
- sgd
- shn
- sml
- snk
- snl
- som
- so
- sot
- st
- sox
- spa
- es
- sps
- ssn
- stk
- swa
- sw
- swh
- sxb
- syw
- taj
- tam
- ta
- tbj
- tdb
- tdg
- tdt
- teo
- tet
- tgk
- tg
- tha
- th
- the
- thk
- thl
- thy
- tio
- tkd
- tnl
- tnn
- tnp
- tnt
- tod
- tom
- tpi
- tpl
- tpu
- tsb
- tsn
- tn
- tso
- ts
- tuv
- tuz
- tvs
- udg
- unr
- urd
- ur
- uzb
- uz
- ven
- ve
- vie
- vi
- vif
- war
- wbm
- wbr
- wms
- wni
- wnk
- wtk
- xho
- xh
- xkg
- xmd
- xmg
- xmm
- xog
- xty
- yas
- yav
- ybb
- ybh
- ybi
- ydd
- yea
- yet
- yid
- yi
- yin
- ymp
- zaw
- zho
- zh
- zlm
- zuh
- zul
- zu
license:
- cc-by-nc-4.0
multilinguality:
- multilingual
size_categories:
- 10K<n<100K
source_datasets:
- original
task_ids:
- image-captioning
paperswithcode_id: null
pretty_name: BloomCaptioning
extra_gated_prompt: >-
One more step before getting this dataset. This dataset is open access and
available only for non-commercial use (except for portions of the dataset
labeled explicitly with a `cc-by-sa` license). A "license" field paired with
each of the dataset entries/samples specifies the Creative Commons license for
that entry/sample.
These [Creative Commons
licenses](https://creativecommons.org/about/cclicenses/) specify that:
1. You cannot use the dataset for or directed toward commercial advantage or
monetary compensation (except for those portions of the dataset labeled
specifically with a `cc-by-sa` license. If you would like to ask about
commercial uses of this dataset, please [email us](mailto:[email protected]).
2. Any public, non-commercial use of the data must give appropriate credit,
provide a link to the license, and indicate if changes were made. You may do
so in any reasonable manner, but not in any way that suggests the licensor
endorses you or your use.
3. For those portions of the dataset marked with an ND license, you cannot
remix, transform, or build upon the material, and you may not distribute
modified material.
In addition to the above implied by Creative Commons and when clicking "Access
Repository" below, you agree:
1. Not to use the dataset for any use intended to or which has the effect of
harming or enabling discrimination against individuals or groups based on
legally protected characteristics or categories, including but not limited to
discrimination against Indigenous People as outlined in Articles 2; 13-16; and
31 of the United Nations Declaration on the Rights of Indigenous People, 13
September 2007 and as subsequently amended and revised.
2. That your *contact information* (email address and username) can be shared
with the model authors as well.
extra_gated_fields:
I have read the License and agree with its terms: checkbox
Dataset Description
- Homepage: SIL AI
- Point of Contact: SIL AI email
- Source Data: Bloom Library
Dataset Summary
Bloom is free, open-source software and an associated website Bloom Library, app, and services developed by SIL International. Bloom’s primary goal is to equip non-dominant language communities and their members to create the literature they want for their community and children. Bloom also serves organizations that help such communities develop literature and education or other aspects of community development.
This version of the Bloom Library data is developed specifically for the image captioning task. It includes data from 351 languages across 31 language families. There is a mean of 32 stories and 319 image-caption pairs per language.
Note: If you speak one of these languages and can help provide feedback or corrections, please let us know!
Note: Although this data was used in the training of the BLOOM model, this dataset only represents a small portion of the data used to train that model. Data from "Bloom Library" was combined with a large number of other datasets to train that model. "Bloom Library" is a project that existed prior to the BLOOM model, and is something separate. All that to say... We were using the "Bloom" name before it was cool. 😉
Languages
Of the 500+ languages listed at BloomLibrary.org, there are 351 languages available in this dataset. Here are the corresponding ISO 639-3 codes:
aaa, abc, ada, adq, aeu, afr, agq, ags, ahk, aia, ajz, aka, ame, amh, amp, amu, ann, aph, awa, awb, azn, azo, bag, bam, baw, bax, bbk, bcc, bce, bec, bef, ben, bfd, bfm, bfn, bgf, bho, bhs, bis, bjn, bjr, bkc, bkh, bkm, bkx, bob, bod, boz, bqm, bra, brb, bri, brv, bss, bud, buo, bwt, bwx, bxa, bya, bze, bzi, cak, cbr, ceb, cgc, chd, chp, cim, clo, cmn, cmo, csw, cuh, cuv, dag, ddg, ded, deu, dig, dje, dmg, dnw, dtp, dtr, dty, dug, eee, ekm, enb, enc, eng, ewo, fas, fil, fli, fon, fra, fub, fuh, gal, gbj, gou, gsw, guc, guj, guz, gwc, hao, hat, hau, hbb, hig, hil, hin, hla, hna, hre, hro, idt, ilo, ind, ino, isu, ita, jgo, jmx, jpn, jra, kak, kam, kan, kau, kbq, kbx, kby, kek, ken, khb, khm, kik, kin, kir, kjb, kmg, kmr, kms, kmu, kor, kqr, krr, ksw, kur, kvt, kwd, kwu, kwx, kxp, kyq, laj, lan, lao, lbr, lfa, lgg, lgr, lhm, lhu, lkb, llg, lmp, lns, loh, lsi, lts, lug, luy, lwl, mai, mal, mam, mar, mdr, mfh, mfj, mgg, mgm, mgo, mgq, mhx, miy, mkz, mle, mlk, mlw, mmu, mne, mnf, mnw, mot, mqj, mrn, mry, msb, muv, mve, mxu, mya, myk, myx, mzm, nas, nco, nep, new, nge, ngn, nhx, njy, nla, nld, nlv, nod, nsk, nsn, nso, nst, nuj, nwe, nwi, nxa, nxl, nya, nyo, nyu, nza, odk, oji, oki, omw, ori, ozm, pae, pag, pan, pbt, pce, pcg, pdu, pea, pex, pis, pkb, pmf, pnz, por, psp, pwg, qub, quc, quf, quz, qve, qvh, qvm, qvo, qxh, rel, rnl, ron, roo, rue, rug, rus, san, saq, sat, sdk, sea, sgd, shn, sml, snk, snl, som, sot, sox, spa, sps, ssn, stk, swa, swh, sxb, syw, taj, tam, tbj, tdb, tdg, tdt, teo, tet, tgk, tha, the, thk, thl, thy, tio, tkd, tnl, tnn, tnp, tnt, tod, tom, tpi, tpl, tpu, tsb, tsn, tso, tuv, tuz, tvs, udg, unr, urd, uzb, ven, vie, vif, war, wbm, wbr, wms, wni, wnk, wtk, xho, xkg, xmd, xmg, xmm, xog, xty, yas, yav, ybb, ybh, ybi, ydd, yea, yet, yid, yin, ymp, zaw, zho, zlm, zuh, zul
Dataset Statistics
Some of the languages included in the dataset just include 1 or a couple of "stories." These are not split between training, validation, and test. For those with higher numbers of available stories we include the following statistics:
ISO 639-3 | stories | image-caption pairs |
---|---|---|
ahk | 101 | 907 |
awa | 163 | 1200 |
bam | 4 | 86 |
ben | 251 | 2235 |
bho | 173 | 1172 |
boz | 5 | 102 |
bzi | 66 | 497 |
cak | 67 | 817 |
ceb | 418 | 2953 |
cgc | 197 | 1638 |
chd | 1 | 84 |
dty | 172 | 1310 |
eng | 2633 | 28618 |
fas | 129 | 631 |
fra | 403 | 5278 |
hat | 260 | 2411 |
hau | 256 | 1865 |
hbb | 27 | 273 |
ind | 259 | 2177 |
jra | 139 | 1423 |
kak | 195 | 1416 |
kan | 21 | 168 |
kek | 36 | 621 |
kir | 382 | 4026 |
kjb | 102 | 984 |
kor | 132 | 2773 |
mai | 180 | 1211 |
mai | 180 | 1211 |
mam | 134 | 1317 |
mhx | 98 | 945 |
mya | 38 | 421 |
myk | 34 | 341 |
nep | 200 | 1507 |
new | 177 | 1225 |
por | 163 | 3101 |
quc | 99 | 817 |
rus | 353 | 3933 |
sdk | 11 | 153 |
snk | 35 | 356 |
spa | 528 | 6111 |
stk | 7 | 113 |
tgl | 0 | 0 |
tha | 285 | 3023 |
thl | 185 | 1464 |
tpi | 201 | 2162 |
Dataset Structure
Data Instances
The examples look like this for Hausa:
from datasets import load_dataset
# Specify the language code.
dataset = load_dataset("sil-ai/bloom-captioning", iso639_3_letter_code,
use_auth_token=True, download_mode='force_redownload')
# An entry in the dataset consists of a image caption along with
# a link to the corresponding image (and various pieces of metadata).
print(dataset['train'][0])
This would produce an output:
{'image_id': '5e7e2ab6-493f-4430-a635-695fbff76cf0',
'image_url': 'https://bloom-vist.s3.amazonaws.com/%E0%A4%AF%E0%A5%87%E0%A4%B8%E0%A5%81%20%E0%A4%9A%E0%A5%81%E0%A4%B5%E0%A4%BE%20%E0%A4%89%E0%A4%A0%E0%A5%81%E0%A4%99%E0%A5%8D%E2%80%8C%E0%A4%99%E0%A4%BF%20%E0%A4%B2%E0%A4%BE%E0%A4%AE%E0%A5%8D%E2%80%8C%E0%A4%9F%E0%A4%BF%E0%A4%AF%E0%A4%BE%E0%A4%A8%E0%A4%BE/image2.jpg',
'caption': 'Lokacinan almajiran suna tuƙa jirgin ruwansu, amma can cikin dare sun kai tsakiyar tafkin kaɗai. Suna tuƙi da wahala saboda iska tana busawa da ƙarfi gaba da su.',
'story_id': 'cd17125d-66c6-467c-b6c3-7463929faff9',
'album_id': 'a3074fc4-b88f-4769-a6de-dc952fdb35f0',
'original_bloom_language_tag': 'ha',
'index_in_story': 0}
To download all of the images locally directory images
, you can do something similar to the following:
from PIL import Image
import urllib
from datasets.utils.file_utils import get_datasets_user_agent
USER_AGENT = get_datasets_user_agent()
def fetch_single_image(image_url, timeout=None, retries=0):
request = urllib.request.Request(
image_url,
data=None,
headers={"user-agent": USER_AGENT},
)
with urllib.request.urlopen(request, timeout=timeout) as req:
if 'png' in image_url:
png = Image.open(io.BytesIO(req.read())).convert('RGBA')
png.load() # required for png.split()
background = Image.new("RGB", png.size, (255, 255, 255))
background.paste(png, mask=png.split()[3]) # 3 is the alpha channel
image_id = str(uuid.uuid4())
image_path = "images/" + image_id + ".jpg"
background.save(image_path, 'JPEG', quality=80)
else:
image = Image.open(io.BytesIO(req.read()))
image_id = str(uuid.uuid4())
image_path = "images/" + image_id + ".jpg"
image.save(image_path)
return image_path
def fetch_images(batch, num_threads, timeout=None, retries=3):
fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries)
with ThreadPoolExecutor(max_workers=num_threads) as executor:
batch["image_path"] = list(executor.map(fetch_single_image_with_args, batch["image_url"]))
return batch
num_threads = 20
dataset = dataset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads})
Data Fields
The metadata fields below are available:
- image_id: a unique ID for the image
- image_url: a link for downloading the image
- caption: a caption corresponding to the image
- story_id: a unique ID for the corresponding story in which the caption appears
- album_id: a unique ID for the corresponding album in which the image appears
- original_bloom_language_tag: the original language identification from the Bloom library
- index_in_story: an index corresponding to the order of the image-caption pair in the corresponding story
Data Splits
All languages include a train, validation, and test split. However, for language having a small number of stories, certain of these splits maybe empty. In such cases, we recommend using any data for testing only or for zero-shot experiments.
NOTE: The captions for the test split are currently hidden due to on ongoing shared task competition. They have been replaced by a placeholder <hidden>
token.
Changelog
- 25 October 2022 - Initial release
- 25 October 2022 - Update to include licenses on each data item.