metadata
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 12954169685
num_examples: 3435343
download_size: 6525186931
dataset_size: 12954169685
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
A dataset of translated wikipedia en->ar. Translated ~70gb worth of english wikipedia text for pretraining. Native arabic wikipedia is only 20gb. The dataset size you see in HF is compressed.
Translated using nllb-600m-distilled