Datasets:

Size:
n>1T
ArXiv:
License:
drschwenk commited on
Commit
85601d1
1 Parent(s): 2272d6b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -28,19 +28,19 @@ dataset is released in a document-level form that has been deduplicated.
28
  You can load both the clean and noisy versions of any language by specifing its LangID:
29
 
30
  ~~~
31
- madlad_ape = load_dataset("allenai/madlad-400", "ape")
32
  ~~~
33
 
34
  A list of langagues can also be supplied with a keyword argument:
35
 
36
  ~~~
37
- madlad_multilang = load_dataset("allenai/madlad-400", languages=["ape", "abt", "ace"])
38
  ~~~
39
 
40
  Additionally, you can load the noisy and clean subsets seperately with the split keyword argument:
41
 
42
  ~~~
43
- madlad_multilang_clean = load_dataset("allenai/madlad-400", languages=["ape", "abt", "ace"], split="clean")
44
  ~~~
45
 
46
 
@@ -951,9 +951,9 @@ A few comments too long to fit in the table above:
951
  The number of documents, sentences, tokens, characters, and bytes for the noisy
952
  and clean splits of the data. Note that the "toks" field below uses whitespace
953
  for tokenization, so is not appropriate for non-whitespace-separating languages
954
- like Chinese (see section above). Note that the english subset in this version of
955
  is missing 18% of documents that were included in the published analysis of the dataset.
956
- These documents will be added in an update coming soon.
957
 
958
  BCP-47 | docs (noisy) | docs (clean) | sents (noisy) | sents (clean) | toks (noisy) | toks (clean) | chars (noisy) | chars (clean) | clean | noisy |
959
  ----------------|:---------------|:---------------|:----------------|:----------------|:---------------|:---------------|:----------------|:----------------|:---------|:---------|
 
28
  You can load both the clean and noisy versions of any language by specifing its LangID:
29
 
30
  ~~~
31
+ madlad_abt = load_dataset("allenai/madlad-400", "abt")
32
  ~~~
33
 
34
  A list of langagues can also be supplied with a keyword argument:
35
 
36
  ~~~
37
+ madlad_multilang = load_dataset("allenai/madlad-400", languages=["abt", "ace"])
38
  ~~~
39
 
40
  Additionally, you can load the noisy and clean subsets seperately with the split keyword argument:
41
 
42
  ~~~
43
+ madlad_multilang_clean = load_dataset("allenai/madlad-400", languages=["abt", "ace"], split="clean")
44
  ~~~
45
 
46
 
 
951
  The number of documents, sentences, tokens, characters, and bytes for the noisy
952
  and clean splits of the data. Note that the "toks" field below uses whitespace
953
  for tokenization, so is not appropriate for non-whitespace-separating languages
954
+ like Chinese (see section above). Note that the english subset in this version
955
  is missing 18% of documents that were included in the published analysis of the dataset.
956
+ These documents will be incoporated in an update coming soon.
957
 
958
  BCP-47 | docs (noisy) | docs (clean) | sents (noisy) | sents (clean) | toks (noisy) | toks (clean) | chars (noisy) | chars (clean) | clean | noisy |
959
  ----------------|:---------------|:---------------|:----------------|:----------------|:---------------|:---------------|:----------------|:----------------|:---------|:---------|