File size: 5,350 Bytes
9605203
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
import os
import math
import json
import argparse
import subprocess
from collections import OrderedDict
from new_to_old_format_data_path import output_sampling_probs_new_format

SPLIT = [0, 0.949, 0.999, 1.0]


def calc_multinomial_sampling_prob_with_penalty(dataset_size, alpha=.5):
    """
    Calculate multinomial probability distribution based on https://arxiv.org/pdf/1901.07291.pdf (section 3.1)
    :dataset_size: A dictionary contains the size (value) of each of the language (key).
    """
    tot_size = 0
    probs = OrderedDict()
    for lang, size in dataset_size.items():
        tot_size += size
    for lang, size in dataset_size.items():
        probs[lang] = size / tot_size

    pen_prob = OrderedDict()
    tot_pen_prob = 0.0
    for lang, prob in probs.items():
        tot_pen_prob += (prob ** alpha)
    sum_ = 0.0
    for lang, prob in probs.items():
        pen_prob[lang] = (prob ** alpha) / tot_pen_prob
        sum_ += pen_prob[lang]
    assert math.fabs(1 - sum_) < 1e-6
    return pen_prob


def get_size_stats(args):
    """
    Calculate size for each of the iterator.
    It recusively iterate though a directory to find a specific extension file and report their size in preferred format.
    """
    lang_size_dict = {}
    for (dirpath, dirnames, filenames) in os.walk(args.data_folder_path):
        for filename in filenames:
            if not (filename.startswith(args.name_prefix) and filename.endswith(args.extension_name)):
                continue
            full_file_path = os.path.join(dirpath, filename)
            lang_size = subprocess.check_output("du -s {}".format(full_file_path), shell=True)
            lang_size = int(lang_size.decode("utf-8").split("\t")[0])
            if args.size_format == 'KB':
                _conv = 1
            elif args.size_format == 'MB':
                _conv = 1024
            elif args.size_format == 'GB':
                _conv = 1024 * 1024
            elif args.size_format == 'TB':
                _conv = 1024 * 1024 * 1024
            lang_size_ = round(lang_size / float(_conv), 2)
            lang_size_dict[full_file_path] = lang_size_
    return lang_size_dict


def print_stat(args, lang_size_dict, value_name='size'):
    """
    Print size statistics.
    """
    lang_list = sorted([(k, v) for k, v in lang_size_dict.items()], key=lambda tup: tup[1])
    total_size = 0
    print("\nLanguage : ({})".format(value_name))
    print("-" * 20)
    for lang, size in lang_list:
        print("{} :   {}".format(lang, size))
        total_size += size
    print("-" * 20)
    print("Total size : {}".format(total_size))


def removesuffix(string, suffix):
    if string.endswith(suffix):
        string = string[:-len(suffix)]
    return string


def main():
    parser = argparse.ArgumentParser()
    parser.add_argument('--data-folder-path', type=str, required=True,
                        help='Path to the data folder')
    parser.add_argument('--size-format', type=str, required=True,
                        help='Calculation will be done in byte, mega-byte, giga-byte or tera-byte',
                        choices=['KB', 'MB', 'GB', 'TB'])
    parser.add_argument('--alpha', type=float, required=True,
                        help='Sampling penalty.')
    parser.add_argument('--output-dir', type=str, required=True,
                        help='Output directory where sampling prob_dict will be saved.')
    parser.add_argument('--name-prefix', type=str, required=True,
                        help='File name prefix to match. Combination of `--name-prefix` and --extension-name will be used to select file.')
    parser.add_argument('--extension-name', type=str, required=True,
                        help='Extension of the file to match. Combination of `--name-prefix` and --extension-name will be used to select file')
    parser.add_argument('--old-format', action="store_true",
                        help='Legacy option')

    args = parser.parse_args()
    size_dict = get_size_stats(args)
    print_stat(args, size_dict, value_name=args.size_format)
    sampling_probability = calc_multinomial_sampling_prob_with_penalty(
        size_dict, alpha=args.alpha
    )
    print_stat(args, sampling_probability, 'probability')
    total_contrib = 0
    print("\nLanguage : Per epoch contribution in {}".format(args.size_format))
    print("-" * 50)
    for lang, prob in sampling_probability.items():
        sampling_probability[lang] = (prob, size_dict[lang])
        lang_contrib_size = round(size_dict[lang] * prob, 2)
        print("{} :   {}  ({}  ->  {})".format(lang, prob, size_dict[lang], lang_contrib_size))
        total_contrib += lang_contrib_size
    print("-" * 50)
    print("Total size : {}".format(total_contrib))

    open(os.path.join(args.output_dir, 'iterator_selection_prob.{}.json'.format(args.alpha)), "w").write(
        json.dumps(sampling_probability, indent=4)
    )

    if args.old_format:
        with open(os.path.join(args.output_dir, "dataset_probabilities.{}.txt".format(args.alpha)), "w") as fout:
            fout.write(
                " ".join([f"{prob[0]} {removesuffix(path, '.bin')}" for path, prob in sampling_probability.items()]))
        pass
    else:
        output_sampling_probs_new_format(sampling_probability, args.output_dir, args.alpha)

if __name__ == '__main__':
    main()