--- dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 12954169685 num_examples: 3435343 download_size: 6525186931 dataset_size: 12954169685 configs: - config_name: default data_files: - split: train path: data/train-* --- A dataset of translated wikipedia en->ar. Translated ~70gb worth of english wikipedia text for pretraining. Native arabic wikipedia is only 20gb. The dataset size you see in HF is compressed. Translated using nllb-600m-distilled