peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/nltk
/probability.py
| # Natural Language Toolkit: Probability and Statistics | |
| # | |
| # Copyright (C) 2001-2023 NLTK Project | |
| # Author: Edward Loper <[email protected]> | |
| # Steven Bird <[email protected]> (additions) | |
| # Trevor Cohn <[email protected]> (additions) | |
| # Peter Ljunglöf <[email protected]> (additions) | |
| # Liang Dong <[email protected]> (additions) | |
| # Geoffrey Sampson <[email protected]> (additions) | |
| # Ilia Kurenkov <[email protected]> (additions) | |
| # | |
| # URL: <https://www.nltk.org/> | |
| # For license information, see LICENSE.TXT | |
| """ | |
| Classes for representing and processing probabilistic information. | |
| The ``FreqDist`` class is used to encode "frequency distributions", | |
| which count the number of times that each outcome of an experiment | |
| occurs. | |
| The ``ProbDistI`` class defines a standard interface for "probability | |
| distributions", which encode the probability of each outcome for an | |
| experiment. There are two types of probability distribution: | |
| - "derived probability distributions" are created from frequency | |
| distributions. They attempt to model the probability distribution | |
| that generated the frequency distribution. | |
| - "analytic probability distributions" are created directly from | |
| parameters (such as variance). | |
| The ``ConditionalFreqDist`` class and ``ConditionalProbDistI`` interface | |
| are used to encode conditional distributions. Conditional probability | |
| distributions can be derived or analytic; but currently the only | |
| implementation of the ``ConditionalProbDistI`` interface is | |
| ``ConditionalProbDist``, a derived distribution. | |
| """ | |
| import array | |
| import math | |
| import random | |
| import warnings | |
| from abc import ABCMeta, abstractmethod | |
| from collections import Counter, defaultdict | |
| from functools import reduce | |
| from nltk.internals import raise_unorderable_types | |
| _NINF = float("-1e300") | |
| ##////////////////////////////////////////////////////// | |
| ## Frequency Distributions | |
| ##////////////////////////////////////////////////////// | |
| class FreqDist(Counter): | |
| """ | |
| A frequency distribution for the outcomes of an experiment. A | |
| frequency distribution records the number of times each outcome of | |
| an experiment has occurred. For example, a frequency distribution | |
| could be used to record the frequency of each word type in a | |
| document. Formally, a frequency distribution can be defined as a | |
| function mapping from each sample to the number of times that | |
| sample occurred as an outcome. | |
| Frequency distributions are generally constructed by running a | |
| number of experiments, and incrementing the count for a sample | |
| every time it is an outcome of an experiment. For example, the | |
| following code will produce a frequency distribution that encodes | |
| how often each word occurs in a text: | |
| >>> from nltk.tokenize import word_tokenize | |
| >>> from nltk.probability import FreqDist | |
| >>> sent = 'This is an example sentence' | |
| >>> fdist = FreqDist() | |
| >>> for word in word_tokenize(sent): | |
| ... fdist[word.lower()] += 1 | |
| An equivalent way to do this is with the initializer: | |
| >>> fdist = FreqDist(word.lower() for word in word_tokenize(sent)) | |
| """ | |
| def __init__(self, samples=None): | |
| """ | |
| Construct a new frequency distribution. If ``samples`` is | |
| given, then the frequency distribution will be initialized | |
| with the count of each object in ``samples``; otherwise, it | |
| will be initialized to be empty. | |
| In particular, ``FreqDist()`` returns an empty frequency | |
| distribution; and ``FreqDist(samples)`` first creates an empty | |
| frequency distribution, and then calls ``update`` with the | |
| list ``samples``. | |
| :param samples: The samples to initialize the frequency | |
| distribution with. | |
| :type samples: Sequence | |
| """ | |
| Counter.__init__(self, samples) | |
| # Cached number of samples in this FreqDist | |
| self._N = None | |
| def N(self): | |
| """ | |
| Return the total number of sample outcomes that have been | |
| recorded by this FreqDist. For the number of unique | |
| sample values (or bins) with counts greater than zero, use | |
| ``FreqDist.B()``. | |
| :rtype: int | |
| """ | |
| if self._N is None: | |
| # Not already cached, or cache has been invalidated | |
| self._N = sum(self.values()) | |
| return self._N | |
| def __setitem__(self, key, val): | |
| """ | |
| Override ``Counter.__setitem__()`` to invalidate the cached N | |
| """ | |
| self._N = None | |
| super().__setitem__(key, val) | |
| def __delitem__(self, key): | |
| """ | |
| Override ``Counter.__delitem__()`` to invalidate the cached N | |
| """ | |
| self._N = None | |
| super().__delitem__(key) | |
| def update(self, *args, **kwargs): | |
| """ | |
| Override ``Counter.update()`` to invalidate the cached N | |
| """ | |
| self._N = None | |
| super().update(*args, **kwargs) | |
| def setdefault(self, key, val): | |
| """ | |
| Override ``Counter.setdefault()`` to invalidate the cached N | |
| """ | |
| self._N = None | |
| super().setdefault(key, val) | |
| def B(self): | |
| """ | |
| Return the total number of sample values (or "bins") that | |
| have counts greater than zero. For the total | |
| number of sample outcomes recorded, use ``FreqDist.N()``. | |
| (FreqDist.B() is the same as len(FreqDist).) | |
| :rtype: int | |
| """ | |
| return len(self) | |
| def hapaxes(self): | |
| """ | |
| Return a list of all samples that occur once (hapax legomena) | |
| :rtype: list | |
| """ | |
| return [item for item in self if self[item] == 1] | |
| def Nr(self, r, bins=None): | |
| return self.r_Nr(bins)[r] | |
| def r_Nr(self, bins=None): | |
| """ | |
| Return the dictionary mapping r to Nr, the number of samples with frequency r, where Nr > 0. | |
| :type bins: int | |
| :param bins: The number of possible sample outcomes. ``bins`` | |
| is used to calculate Nr(0). In particular, Nr(0) is | |
| ``bins-self.B()``. If ``bins`` is not specified, it | |
| defaults to ``self.B()`` (so Nr(0) will be 0). | |
| :rtype: int | |
| """ | |
| _r_Nr = defaultdict(int) | |
| for count in self.values(): | |
| _r_Nr[count] += 1 | |
| # Special case for Nr[0]: | |
| _r_Nr[0] = bins - self.B() if bins is not None else 0 | |
| return _r_Nr | |
| def _cumulative_frequencies(self, samples): | |
| """ | |
| Return the cumulative frequencies of the specified samples. | |
| If no samples are specified, all counts are returned, starting | |
| with the largest. | |
| :param samples: the samples whose frequencies should be returned. | |
| :type samples: any | |
| :rtype: list(float) | |
| """ | |
| cf = 0.0 | |
| for sample in samples: | |
| cf += self[sample] | |
| yield cf | |
| # slightly odd nomenclature freq() if FreqDist does counts and ProbDist does probs, | |
| # here, freq() does probs | |
| def freq(self, sample): | |
| """ | |
| Return the frequency of a given sample. The frequency of a | |
| sample is defined as the count of that sample divided by the | |
| total number of sample outcomes that have been recorded by | |
| this FreqDist. The count of a sample is defined as the | |
| number of times that sample outcome was recorded by this | |
| FreqDist. Frequencies are always real numbers in the range | |
| [0, 1]. | |
| :param sample: the sample whose frequency | |
| should be returned. | |
| :type sample: any | |
| :rtype: float | |
| """ | |
| n = self.N() | |
| if n == 0: | |
| return 0 | |
| return self[sample] / n | |
| def max(self): | |
| """ | |
| Return the sample with the greatest number of outcomes in this | |
| frequency distribution. If two or more samples have the same | |
| number of outcomes, return one of them; which sample is | |
| returned is undefined. If no outcomes have occurred in this | |
| frequency distribution, return None. | |
| :return: The sample with the maximum number of outcomes in this | |
| frequency distribution. | |
| :rtype: any or None | |
| """ | |
| if len(self) == 0: | |
| raise ValueError( | |
| "A FreqDist must have at least one sample before max is defined." | |
| ) | |
| return self.most_common(1)[0][0] | |
| def plot( | |
| self, *args, title="", cumulative=False, percents=False, show=True, **kwargs | |
| ): | |
| """ | |
| Plot samples from the frequency distribution | |
| displaying the most frequent sample first. If an integer | |
| parameter is supplied, stop after this many samples have been | |
| plotted. For a cumulative plot, specify cumulative=True. Additional | |
| ``**kwargs`` are passed to matplotlib's plot function. | |
| (Requires Matplotlib to be installed.) | |
| :param title: The title for the graph. | |
| :type title: str | |
| :param cumulative: Whether the plot is cumulative. (default = False) | |
| :type cumulative: bool | |
| :param percents: Whether the plot uses percents instead of counts. (default = False) | |
| :type percents: bool | |
| :param show: Whether to show the plot, or only return the ax. | |
| :type show: bool | |
| """ | |
| try: | |
| import matplotlib.pyplot as plt | |
| except ImportError as e: | |
| raise ValueError( | |
| "The plot function requires matplotlib to be installed." | |
| "See https://matplotlib.org/" | |
| ) from e | |
| if len(args) == 0: | |
| args = [len(self)] | |
| samples = [item for item, _ in self.most_common(*args)] | |
| if cumulative: | |
| freqs = list(self._cumulative_frequencies(samples)) | |
| ylabel = "Cumulative " | |
| else: | |
| freqs = [self[sample] for sample in samples] | |
| ylabel = "" | |
| if percents: | |
| freqs = [f / self.N() * 100 for f in freqs] | |
| ylabel += "Percents" | |
| else: | |
| ylabel += "Counts" | |
| ax = plt.gca() | |
| ax.grid(True, color="silver") | |
| if "linewidth" not in kwargs: | |
| kwargs["linewidth"] = 2 | |
| if title: | |
| ax.set_title(title) | |
| ax.plot(freqs, **kwargs) | |
| ax.set_xticks(range(len(samples))) | |
| ax.set_xticklabels([str(s) for s in samples], rotation=90) | |
| ax.set_xlabel("Samples") | |
| ax.set_ylabel(ylabel) | |
| if show: | |
| plt.show() | |
| return ax | |
| def tabulate(self, *args, **kwargs): | |
| """ | |
| Tabulate the given samples from the frequency distribution (cumulative), | |
| displaying the most frequent sample first. If an integer | |
| parameter is supplied, stop after this many samples have been | |
| plotted. | |
| :param samples: The samples to plot (default is all samples) | |
| :type samples: list | |
| :param cumulative: A flag to specify whether the freqs are cumulative (default = False) | |
| :type title: bool | |
| """ | |
| if len(args) == 0: | |
| args = [len(self)] | |
| samples = _get_kwarg( | |
| kwargs, "samples", [item for item, _ in self.most_common(*args)] | |
| ) | |
| cumulative = _get_kwarg(kwargs, "cumulative", False) | |
| if cumulative: | |
| freqs = list(self._cumulative_frequencies(samples)) | |
| else: | |
| freqs = [self[sample] for sample in samples] | |
| # percents = [f * 100 for f in freqs] only in ProbDist? | |
| width = max(len(f"{s}") for s in samples) | |
| width = max(width, max(len("%d" % f) for f in freqs)) | |
| for i in range(len(samples)): | |
| print("%*s" % (width, samples[i]), end=" ") | |
| print() | |
| for i in range(len(samples)): | |
| print("%*d" % (width, freqs[i]), end=" ") | |
| print() | |
| def copy(self): | |
| """ | |
| Create a copy of this frequency distribution. | |
| :rtype: FreqDist | |
| """ | |
| return self.__class__(self) | |
| # Mathematical operatiors | |
| def __add__(self, other): | |
| """ | |
| Add counts from two counters. | |
| >>> FreqDist('abbb') + FreqDist('bcc') | |
| FreqDist({'b': 4, 'c': 2, 'a': 1}) | |
| """ | |
| return self.__class__(super().__add__(other)) | |
| def __sub__(self, other): | |
| """ | |
| Subtract count, but keep only results with positive counts. | |
| >>> FreqDist('abbbc') - FreqDist('bccd') | |
| FreqDist({'b': 2, 'a': 1}) | |
| """ | |
| return self.__class__(super().__sub__(other)) | |
| def __or__(self, other): | |
| """ | |
| Union is the maximum of value in either of the input counters. | |
| >>> FreqDist('abbb') | FreqDist('bcc') | |
| FreqDist({'b': 3, 'c': 2, 'a': 1}) | |
| """ | |
| return self.__class__(super().__or__(other)) | |
| def __and__(self, other): | |
| """ | |
| Intersection is the minimum of corresponding counts. | |
| >>> FreqDist('abbb') & FreqDist('bcc') | |
| FreqDist({'b': 1}) | |
| """ | |
| return self.__class__(super().__and__(other)) | |
| def __le__(self, other): | |
| """ | |
| Returns True if this frequency distribution is a subset of the other | |
| and for no key the value exceeds the value of the same key from | |
| the other frequency distribution. | |
| The <= operator forms partial order and satisfying the axioms | |
| reflexivity, antisymmetry and transitivity. | |
| >>> FreqDist('a') <= FreqDist('a') | |
| True | |
| >>> a = FreqDist('abc') | |
| >>> b = FreqDist('aabc') | |
| >>> (a <= b, b <= a) | |
| (True, False) | |
| >>> FreqDist('a') <= FreqDist('abcd') | |
| True | |
| >>> FreqDist('abc') <= FreqDist('xyz') | |
| False | |
| >>> FreqDist('xyz') <= FreqDist('abc') | |
| False | |
| >>> c = FreqDist('a') | |
| >>> d = FreqDist('aa') | |
| >>> e = FreqDist('aaa') | |
| >>> c <= d and d <= e and c <= e | |
| True | |
| """ | |
| if not isinstance(other, FreqDist): | |
| raise_unorderable_types("<=", self, other) | |
| return set(self).issubset(other) and all( | |
| self[key] <= other[key] for key in self | |
| ) | |
| def __ge__(self, other): | |
| if not isinstance(other, FreqDist): | |
| raise_unorderable_types(">=", self, other) | |
| return set(self).issuperset(other) and all( | |
| self[key] >= other[key] for key in other | |
| ) | |
| __lt__ = lambda self, other: self <= other and not self == other | |
| __gt__ = lambda self, other: self >= other and not self == other | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this FreqDist. | |
| :rtype: string | |
| """ | |
| return self.pformat() | |
| def pprint(self, maxlen=10, stream=None): | |
| """ | |
| Print a string representation of this FreqDist to 'stream' | |
| :param maxlen: The maximum number of items to print | |
| :type maxlen: int | |
| :param stream: The stream to print to. stdout by default | |
| """ | |
| print(self.pformat(maxlen=maxlen), file=stream) | |
| def pformat(self, maxlen=10): | |
| """ | |
| Return a string representation of this FreqDist. | |
| :param maxlen: The maximum number of items to display | |
| :type maxlen: int | |
| :rtype: string | |
| """ | |
| items = ["{!r}: {!r}".format(*item) for item in self.most_common(maxlen)] | |
| if len(self) > maxlen: | |
| items.append("...") | |
| return "FreqDist({{{0}}})".format(", ".join(items)) | |
| def __str__(self): | |
| """ | |
| Return a string representation of this FreqDist. | |
| :rtype: string | |
| """ | |
| return "<FreqDist with %d samples and %d outcomes>" % (len(self), self.N()) | |
| def __iter__(self): | |
| """ | |
| Return an iterator which yields tokens ordered by frequency. | |
| :rtype: iterator | |
| """ | |
| for token, _ in self.most_common(self.B()): | |
| yield token | |
| ##////////////////////////////////////////////////////// | |
| ## Probability Distributions | |
| ##////////////////////////////////////////////////////// | |
| class ProbDistI(metaclass=ABCMeta): | |
| """ | |
| A probability distribution for the outcomes of an experiment. A | |
| probability distribution specifies how likely it is that an | |
| experiment will have any given outcome. For example, a | |
| probability distribution could be used to predict the probability | |
| that a token in a document will have a given type. Formally, a | |
| probability distribution can be defined as a function mapping from | |
| samples to nonnegative real numbers, such that the sum of every | |
| number in the function's range is 1.0. A ``ProbDist`` is often | |
| used to model the probability distribution of the experiment used | |
| to generate a frequency distribution. | |
| """ | |
| SUM_TO_ONE = True | |
| """True if the probabilities of the samples in this probability | |
| distribution will always sum to one.""" | |
| def __init__(self): | |
| """ | |
| Classes inheriting from ProbDistI should implement __init__. | |
| """ | |
| def prob(self, sample): | |
| """ | |
| Return the probability for a given sample. Probabilities | |
| are always real numbers in the range [0, 1]. | |
| :param sample: The sample whose probability | |
| should be returned. | |
| :type sample: any | |
| :rtype: float | |
| """ | |
| def logprob(self, sample): | |
| """ | |
| Return the base 2 logarithm of the probability for a given sample. | |
| :param sample: The sample whose probability | |
| should be returned. | |
| :type sample: any | |
| :rtype: float | |
| """ | |
| # Default definition, in terms of prob() | |
| p = self.prob(sample) | |
| return math.log(p, 2) if p != 0 else _NINF | |
| def max(self): | |
| """ | |
| Return the sample with the greatest probability. If two or | |
| more samples have the same probability, return one of them; | |
| which sample is returned is undefined. | |
| :rtype: any | |
| """ | |
| def samples(self): | |
| """ | |
| Return a list of all samples that have nonzero probabilities. | |
| Use ``prob`` to find the probability of each sample. | |
| :rtype: list | |
| """ | |
| # cf self.SUM_TO_ONE | |
| def discount(self): | |
| """ | |
| Return the ratio by which counts are discounted on average: c*/c | |
| :rtype: float | |
| """ | |
| return 0.0 | |
| # Subclasses should define more efficient implementations of this, | |
| # where possible. | |
| def generate(self): | |
| """ | |
| Return a randomly selected sample from this probability distribution. | |
| The probability of returning each sample ``samp`` is equal to | |
| ``self.prob(samp)``. | |
| """ | |
| p = random.random() | |
| p_init = p | |
| for sample in self.samples(): | |
| p -= self.prob(sample) | |
| if p <= 0: | |
| return sample | |
| # allow for some rounding error: | |
| if p < 0.0001: | |
| return sample | |
| # we *should* never get here | |
| if self.SUM_TO_ONE: | |
| warnings.warn( | |
| "Probability distribution %r sums to %r; generate()" | |
| " is returning an arbitrary sample." % (self, p_init - p) | |
| ) | |
| return random.choice(list(self.samples())) | |
| class UniformProbDist(ProbDistI): | |
| """ | |
| A probability distribution that assigns equal probability to each | |
| sample in a given set; and a zero probability to all other | |
| samples. | |
| """ | |
| def __init__(self, samples): | |
| """ | |
| Construct a new uniform probability distribution, that assigns | |
| equal probability to each sample in ``samples``. | |
| :param samples: The samples that should be given uniform | |
| probability. | |
| :type samples: list | |
| :raise ValueError: If ``samples`` is empty. | |
| """ | |
| if len(samples) == 0: | |
| raise ValueError( | |
| "A Uniform probability distribution must " + "have at least one sample." | |
| ) | |
| self._sampleset = set(samples) | |
| self._prob = 1.0 / len(self._sampleset) | |
| self._samples = list(self._sampleset) | |
| def prob(self, sample): | |
| return self._prob if sample in self._sampleset else 0 | |
| def max(self): | |
| return self._samples[0] | |
| def samples(self): | |
| return self._samples | |
| def __repr__(self): | |
| return "<UniformProbDist with %d samples>" % len(self._sampleset) | |
| class RandomProbDist(ProbDistI): | |
| """ | |
| Generates a random probability distribution whereby each sample | |
| will be between 0 and 1 with equal probability (uniform random distribution. | |
| Also called a continuous uniform distribution). | |
| """ | |
| def __init__(self, samples): | |
| if len(samples) == 0: | |
| raise ValueError( | |
| "A probability distribution must " + "have at least one sample." | |
| ) | |
| self._probs = self.unirand(samples) | |
| self._samples = list(self._probs.keys()) | |
| def unirand(cls, samples): | |
| """ | |
| The key function that creates a randomized initial distribution | |
| that still sums to 1. Set as a dictionary of prob values so that | |
| it can still be passed to MutableProbDist and called with identical | |
| syntax to UniformProbDist | |
| """ | |
| samples = set(samples) | |
| randrow = [random.random() for i in range(len(samples))] | |
| total = sum(randrow) | |
| for i, x in enumerate(randrow): | |
| randrow[i] = x / total | |
| total = sum(randrow) | |
| if total != 1: | |
| # this difference, if present, is so small (near NINF) that it | |
| # can be subtracted from any element without risking probs not (0 1) | |
| randrow[-1] -= total - 1 | |
| return {s: randrow[i] for i, s in enumerate(samples)} | |
| def max(self): | |
| if not hasattr(self, "_max"): | |
| self._max = max((p, v) for (v, p) in self._probs.items())[1] | |
| return self._max | |
| def prob(self, sample): | |
| return self._probs.get(sample, 0) | |
| def samples(self): | |
| return self._samples | |
| def __repr__(self): | |
| return "<RandomUniformProbDist with %d samples>" % len(self._probs) | |
| class DictionaryProbDist(ProbDistI): | |
| """ | |
| A probability distribution whose probabilities are directly | |
| specified by a given dictionary. The given dictionary maps | |
| samples to probabilities. | |
| """ | |
| def __init__(self, prob_dict=None, log=False, normalize=False): | |
| """ | |
| Construct a new probability distribution from the given | |
| dictionary, which maps values to probabilities (or to log | |
| probabilities, if ``log`` is true). If ``normalize`` is | |
| true, then the probability values are scaled by a constant | |
| factor such that they sum to 1. | |
| If called without arguments, the resulting probability | |
| distribution assigns zero probability to all values. | |
| """ | |
| self._prob_dict = prob_dict.copy() if prob_dict is not None else {} | |
| self._log = log | |
| # Normalize the distribution, if requested. | |
| if normalize: | |
| if len(prob_dict) == 0: | |
| raise ValueError( | |
| "A DictionaryProbDist must have at least one sample " | |
| + "before it can be normalized." | |
| ) | |
| if log: | |
| value_sum = sum_logs(list(self._prob_dict.values())) | |
| if value_sum <= _NINF: | |
| logp = math.log(1.0 / len(prob_dict), 2) | |
| for x in prob_dict: | |
| self._prob_dict[x] = logp | |
| else: | |
| for (x, p) in self._prob_dict.items(): | |
| self._prob_dict[x] -= value_sum | |
| else: | |
| value_sum = sum(self._prob_dict.values()) | |
| if value_sum == 0: | |
| p = 1.0 / len(prob_dict) | |
| for x in prob_dict: | |
| self._prob_dict[x] = p | |
| else: | |
| norm_factor = 1.0 / value_sum | |
| for (x, p) in self._prob_dict.items(): | |
| self._prob_dict[x] *= norm_factor | |
| def prob(self, sample): | |
| if self._log: | |
| return 2 ** (self._prob_dict[sample]) if sample in self._prob_dict else 0 | |
| else: | |
| return self._prob_dict.get(sample, 0) | |
| def logprob(self, sample): | |
| if self._log: | |
| return self._prob_dict.get(sample, _NINF) | |
| else: | |
| if sample not in self._prob_dict: | |
| return _NINF | |
| elif self._prob_dict[sample] == 0: | |
| return _NINF | |
| else: | |
| return math.log(self._prob_dict[sample], 2) | |
| def max(self): | |
| if not hasattr(self, "_max"): | |
| self._max = max((p, v) for (v, p) in self._prob_dict.items())[1] | |
| return self._max | |
| def samples(self): | |
| return self._prob_dict.keys() | |
| def __repr__(self): | |
| return "<ProbDist with %d samples>" % len(self._prob_dict) | |
| class MLEProbDist(ProbDistI): | |
| """ | |
| The maximum likelihood estimate for the probability distribution | |
| of the experiment used to generate a frequency distribution. The | |
| "maximum likelihood estimate" approximates the probability of | |
| each sample as the frequency of that sample in the frequency | |
| distribution. | |
| """ | |
| def __init__(self, freqdist, bins=None): | |
| """ | |
| Use the maximum likelihood estimate to create a probability | |
| distribution for the experiment used to generate ``freqdist``. | |
| :type freqdist: FreqDist | |
| :param freqdist: The frequency distribution that the | |
| probability estimates should be based on. | |
| """ | |
| self._freqdist = freqdist | |
| def freqdist(self): | |
| """ | |
| Return the frequency distribution that this probability | |
| distribution is based on. | |
| :rtype: FreqDist | |
| """ | |
| return self._freqdist | |
| def prob(self, sample): | |
| return self._freqdist.freq(sample) | |
| def max(self): | |
| return self._freqdist.max() | |
| def samples(self): | |
| return self._freqdist.keys() | |
| def __repr__(self): | |
| """ | |
| :rtype: str | |
| :return: A string representation of this ``ProbDist``. | |
| """ | |
| return "<MLEProbDist based on %d samples>" % self._freqdist.N() | |
| class LidstoneProbDist(ProbDistI): | |
| """ | |
| The Lidstone estimate for the probability distribution of the | |
| experiment used to generate a frequency distribution. The | |
| "Lidstone estimate" is parameterized by a real number *gamma*, | |
| which typically ranges from 0 to 1. The Lidstone estimate | |
| approximates the probability of a sample with count *c* from an | |
| experiment with *N* outcomes and *B* bins as | |
| ``c+gamma)/(N+B*gamma)``. This is equivalent to adding | |
| *gamma* to the count for each bin, and taking the maximum | |
| likelihood estimate of the resulting frequency distribution. | |
| """ | |
| SUM_TO_ONE = False | |
| def __init__(self, freqdist, gamma, bins=None): | |
| """ | |
| Use the Lidstone estimate to create a probability distribution | |
| for the experiment used to generate ``freqdist``. | |
| :type freqdist: FreqDist | |
| :param freqdist: The frequency distribution that the | |
| probability estimates should be based on. | |
| :type gamma: float | |
| :param gamma: A real number used to parameterize the | |
| estimate. The Lidstone estimate is equivalent to adding | |
| *gamma* to the count for each bin, and taking the | |
| maximum likelihood estimate of the resulting frequency | |
| distribution. | |
| :type bins: int | |
| :param bins: The number of sample values that can be generated | |
| by the experiment that is described by the probability | |
| distribution. This value must be correctly set for the | |
| probabilities of the sample values to sum to one. If | |
| ``bins`` is not specified, it defaults to ``freqdist.B()``. | |
| """ | |
| if (bins == 0) or (bins is None and freqdist.N() == 0): | |
| name = self.__class__.__name__[:-8] | |
| raise ValueError( | |
| "A %s probability distribution " % name + "must have at least one bin." | |
| ) | |
| if (bins is not None) and (bins < freqdist.B()): | |
| name = self.__class__.__name__[:-8] | |
| raise ValueError( | |
| "\nThe number of bins in a %s distribution " % name | |
| + "(%d) must be greater than or equal to\n" % bins | |
| + "the number of bins in the FreqDist used " | |
| + "to create it (%d)." % freqdist.B() | |
| ) | |
| self._freqdist = freqdist | |
| self._gamma = float(gamma) | |
| self._N = self._freqdist.N() | |
| if bins is None: | |
| bins = freqdist.B() | |
| self._bins = bins | |
| self._divisor = self._N + bins * gamma | |
| if self._divisor == 0.0: | |
| # In extreme cases we force the probability to be 0, | |
| # which it will be, since the count will be 0: | |
| self._gamma = 0 | |
| self._divisor = 1 | |
| def freqdist(self): | |
| """ | |
| Return the frequency distribution that this probability | |
| distribution is based on. | |
| :rtype: FreqDist | |
| """ | |
| return self._freqdist | |
| def prob(self, sample): | |
| c = self._freqdist[sample] | |
| return (c + self._gamma) / self._divisor | |
| def max(self): | |
| # For Lidstone distributions, probability is monotonic with | |
| # frequency, so the most probable sample is the one that | |
| # occurs most frequently. | |
| return self._freqdist.max() | |
| def samples(self): | |
| return self._freqdist.keys() | |
| def discount(self): | |
| gb = self._gamma * self._bins | |
| return gb / (self._N + gb) | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ProbDist``. | |
| :rtype: str | |
| """ | |
| return "<LidstoneProbDist based on %d samples>" % self._freqdist.N() | |
| class LaplaceProbDist(LidstoneProbDist): | |
| """ | |
| The Laplace estimate for the probability distribution of the | |
| experiment used to generate a frequency distribution. The | |
| "Laplace estimate" approximates the probability of a sample with | |
| count *c* from an experiment with *N* outcomes and *B* bins as | |
| *(c+1)/(N+B)*. This is equivalent to adding one to the count for | |
| each bin, and taking the maximum likelihood estimate of the | |
| resulting frequency distribution. | |
| """ | |
| def __init__(self, freqdist, bins=None): | |
| """ | |
| Use the Laplace estimate to create a probability distribution | |
| for the experiment used to generate ``freqdist``. | |
| :type freqdist: FreqDist | |
| :param freqdist: The frequency distribution that the | |
| probability estimates should be based on. | |
| :type bins: int | |
| :param bins: The number of sample values that can be generated | |
| by the experiment that is described by the probability | |
| distribution. This value must be correctly set for the | |
| probabilities of the sample values to sum to one. If | |
| ``bins`` is not specified, it defaults to ``freqdist.B()``. | |
| """ | |
| LidstoneProbDist.__init__(self, freqdist, 1, bins) | |
| def __repr__(self): | |
| """ | |
| :rtype: str | |
| :return: A string representation of this ``ProbDist``. | |
| """ | |
| return "<LaplaceProbDist based on %d samples>" % self._freqdist.N() | |
| class ELEProbDist(LidstoneProbDist): | |
| """ | |
| The expected likelihood estimate for the probability distribution | |
| of the experiment used to generate a frequency distribution. The | |
| "expected likelihood estimate" approximates the probability of a | |
| sample with count *c* from an experiment with *N* outcomes and | |
| *B* bins as *(c+0.5)/(N+B/2)*. This is equivalent to adding 0.5 | |
| to the count for each bin, and taking the maximum likelihood | |
| estimate of the resulting frequency distribution. | |
| """ | |
| def __init__(self, freqdist, bins=None): | |
| """ | |
| Use the expected likelihood estimate to create a probability | |
| distribution for the experiment used to generate ``freqdist``. | |
| :type freqdist: FreqDist | |
| :param freqdist: The frequency distribution that the | |
| probability estimates should be based on. | |
| :type bins: int | |
| :param bins: The number of sample values that can be generated | |
| by the experiment that is described by the probability | |
| distribution. This value must be correctly set for the | |
| probabilities of the sample values to sum to one. If | |
| ``bins`` is not specified, it defaults to ``freqdist.B()``. | |
| """ | |
| LidstoneProbDist.__init__(self, freqdist, 0.5, bins) | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ProbDist``. | |
| :rtype: str | |
| """ | |
| return "<ELEProbDist based on %d samples>" % self._freqdist.N() | |
| class HeldoutProbDist(ProbDistI): | |
| """ | |
| The heldout estimate for the probability distribution of the | |
| experiment used to generate two frequency distributions. These | |
| two frequency distributions are called the "heldout frequency | |
| distribution" and the "base frequency distribution." The | |
| "heldout estimate" uses uses the "heldout frequency | |
| distribution" to predict the probability of each sample, given its | |
| frequency in the "base frequency distribution". | |
| In particular, the heldout estimate approximates the probability | |
| for a sample that occurs *r* times in the base distribution as | |
| the average frequency in the heldout distribution of all samples | |
| that occur *r* times in the base distribution. | |
| This average frequency is *Tr[r]/(Nr[r].N)*, where: | |
| - *Tr[r]* is the total count in the heldout distribution for | |
| all samples that occur *r* times in the base distribution. | |
| - *Nr[r]* is the number of samples that occur *r* times in | |
| the base distribution. | |
| - *N* is the number of outcomes recorded by the heldout | |
| frequency distribution. | |
| In order to increase the efficiency of the ``prob`` member | |
| function, *Tr[r]/(Nr[r].N)* is precomputed for each value of *r* | |
| when the ``HeldoutProbDist`` is created. | |
| :type _estimate: list(float) | |
| :ivar _estimate: A list mapping from *r*, the number of | |
| times that a sample occurs in the base distribution, to the | |
| probability estimate for that sample. ``_estimate[r]`` is | |
| calculated by finding the average frequency in the heldout | |
| distribution of all samples that occur *r* times in the base | |
| distribution. In particular, ``_estimate[r]`` = | |
| *Tr[r]/(Nr[r].N)*. | |
| :type _max_r: int | |
| :ivar _max_r: The maximum number of times that any sample occurs | |
| in the base distribution. ``_max_r`` is used to decide how | |
| large ``_estimate`` must be. | |
| """ | |
| SUM_TO_ONE = False | |
| def __init__(self, base_fdist, heldout_fdist, bins=None): | |
| """ | |
| Use the heldout estimate to create a probability distribution | |
| for the experiment used to generate ``base_fdist`` and | |
| ``heldout_fdist``. | |
| :type base_fdist: FreqDist | |
| :param base_fdist: The base frequency distribution. | |
| :type heldout_fdist: FreqDist | |
| :param heldout_fdist: The heldout frequency distribution. | |
| :type bins: int | |
| :param bins: The number of sample values that can be generated | |
| by the experiment that is described by the probability | |
| distribution. This value must be correctly set for the | |
| probabilities of the sample values to sum to one. If | |
| ``bins`` is not specified, it defaults to ``freqdist.B()``. | |
| """ | |
| self._base_fdist = base_fdist | |
| self._heldout_fdist = heldout_fdist | |
| # The max number of times any sample occurs in base_fdist. | |
| self._max_r = base_fdist[base_fdist.max()] | |
| # Calculate Tr, Nr, and N. | |
| Tr = self._calculate_Tr() | |
| r_Nr = base_fdist.r_Nr(bins) | |
| Nr = [r_Nr[r] for r in range(self._max_r + 1)] | |
| N = heldout_fdist.N() | |
| # Use Tr, Nr, and N to compute the probability estimate for | |
| # each value of r. | |
| self._estimate = self._calculate_estimate(Tr, Nr, N) | |
| def _calculate_Tr(self): | |
| """ | |
| Return the list *Tr*, where *Tr[r]* is the total count in | |
| ``heldout_fdist`` for all samples that occur *r* | |
| times in ``base_fdist``. | |
| :rtype: list(float) | |
| """ | |
| Tr = [0.0] * (self._max_r + 1) | |
| for sample in self._heldout_fdist: | |
| r = self._base_fdist[sample] | |
| Tr[r] += self._heldout_fdist[sample] | |
| return Tr | |
| def _calculate_estimate(self, Tr, Nr, N): | |
| """ | |
| Return the list *estimate*, where *estimate[r]* is the probability | |
| estimate for any sample that occurs *r* times in the base frequency | |
| distribution. In particular, *estimate[r]* is *Tr[r]/(N[r].N)*. | |
| In the special case that *N[r]=0*, *estimate[r]* will never be used; | |
| so we define *estimate[r]=None* for those cases. | |
| :rtype: list(float) | |
| :type Tr: list(float) | |
| :param Tr: the list *Tr*, where *Tr[r]* is the total count in | |
| the heldout distribution for all samples that occur *r* | |
| times in base distribution. | |
| :type Nr: list(float) | |
| :param Nr: The list *Nr*, where *Nr[r]* is the number of | |
| samples that occur *r* times in the base distribution. | |
| :type N: int | |
| :param N: The total number of outcomes recorded by the heldout | |
| frequency distribution. | |
| """ | |
| estimate = [] | |
| for r in range(self._max_r + 1): | |
| if Nr[r] == 0: | |
| estimate.append(None) | |
| else: | |
| estimate.append(Tr[r] / (Nr[r] * N)) | |
| return estimate | |
| def base_fdist(self): | |
| """ | |
| Return the base frequency distribution that this probability | |
| distribution is based on. | |
| :rtype: FreqDist | |
| """ | |
| return self._base_fdist | |
| def heldout_fdist(self): | |
| """ | |
| Return the heldout frequency distribution that this | |
| probability distribution is based on. | |
| :rtype: FreqDist | |
| """ | |
| return self._heldout_fdist | |
| def samples(self): | |
| return self._base_fdist.keys() | |
| def prob(self, sample): | |
| # Use our precomputed probability estimate. | |
| r = self._base_fdist[sample] | |
| return self._estimate[r] | |
| def max(self): | |
| # Note: the Heldout estimation is *not* necessarily monotonic; | |
| # so this implementation is currently broken. However, it | |
| # should give the right answer *most* of the time. :) | |
| return self._base_fdist.max() | |
| def discount(self): | |
| raise NotImplementedError() | |
| def __repr__(self): | |
| """ | |
| :rtype: str | |
| :return: A string representation of this ``ProbDist``. | |
| """ | |
| s = "<HeldoutProbDist: %d base samples; %d heldout samples>" | |
| return s % (self._base_fdist.N(), self._heldout_fdist.N()) | |
| class CrossValidationProbDist(ProbDistI): | |
| """ | |
| The cross-validation estimate for the probability distribution of | |
| the experiment used to generate a set of frequency distribution. | |
| The "cross-validation estimate" for the probability of a sample | |
| is found by averaging the held-out estimates for the sample in | |
| each pair of frequency distributions. | |
| """ | |
| SUM_TO_ONE = False | |
| def __init__(self, freqdists, bins): | |
| """ | |
| Use the cross-validation estimate to create a probability | |
| distribution for the experiment used to generate | |
| ``freqdists``. | |
| :type freqdists: list(FreqDist) | |
| :param freqdists: A list of the frequency distributions | |
| generated by the experiment. | |
| :type bins: int | |
| :param bins: The number of sample values that can be generated | |
| by the experiment that is described by the probability | |
| distribution. This value must be correctly set for the | |
| probabilities of the sample values to sum to one. If | |
| ``bins`` is not specified, it defaults to ``freqdist.B()``. | |
| """ | |
| self._freqdists = freqdists | |
| # Create a heldout probability distribution for each pair of | |
| # frequency distributions in freqdists. | |
| self._heldout_probdists = [] | |
| for fdist1 in freqdists: | |
| for fdist2 in freqdists: | |
| if fdist1 is not fdist2: | |
| probdist = HeldoutProbDist(fdist1, fdist2, bins) | |
| self._heldout_probdists.append(probdist) | |
| def freqdists(self): | |
| """ | |
| Return the list of frequency distributions that this ``ProbDist`` is based on. | |
| :rtype: list(FreqDist) | |
| """ | |
| return self._freqdists | |
| def samples(self): | |
| # [xx] nb: this is not too efficient | |
| return set(sum((list(fd) for fd in self._freqdists), [])) | |
| def prob(self, sample): | |
| # Find the average probability estimate returned by each | |
| # heldout distribution. | |
| prob = 0.0 | |
| for heldout_probdist in self._heldout_probdists: | |
| prob += heldout_probdist.prob(sample) | |
| return prob / len(self._heldout_probdists) | |
| def discount(self): | |
| raise NotImplementedError() | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ProbDist``. | |
| :rtype: str | |
| """ | |
| return "<CrossValidationProbDist: %d-way>" % len(self._freqdists) | |
| class WittenBellProbDist(ProbDistI): | |
| """ | |
| The Witten-Bell estimate of a probability distribution. This distribution | |
| allocates uniform probability mass to as yet unseen events by using the | |
| number of events that have only been seen once. The probability mass | |
| reserved for unseen events is equal to *T / (N + T)* | |
| where *T* is the number of observed event types and *N* is the total | |
| number of observed events. This equates to the maximum likelihood estimate | |
| of a new type event occurring. The remaining probability mass is discounted | |
| such that all probability estimates sum to one, yielding: | |
| - *p = T / Z (N + T)*, if count = 0 | |
| - *p = c / (N + T)*, otherwise | |
| """ | |
| def __init__(self, freqdist, bins=None): | |
| """ | |
| Creates a distribution of Witten-Bell probability estimates. This | |
| distribution allocates uniform probability mass to as yet unseen | |
| events by using the number of events that have only been seen once. The | |
| probability mass reserved for unseen events is equal to *T / (N + T)* | |
| where *T* is the number of observed event types and *N* is the total | |
| number of observed events. This equates to the maximum likelihood | |
| estimate of a new type event occurring. The remaining probability mass | |
| is discounted such that all probability estimates sum to one, | |
| yielding: | |
| - *p = T / Z (N + T)*, if count = 0 | |
| - *p = c / (N + T)*, otherwise | |
| The parameters *T* and *N* are taken from the ``freqdist`` parameter | |
| (the ``B()`` and ``N()`` values). The normalizing factor *Z* is | |
| calculated using these values along with the ``bins`` parameter. | |
| :param freqdist: The frequency counts upon which to base the | |
| estimation. | |
| :type freqdist: FreqDist | |
| :param bins: The number of possible event types. This must be at least | |
| as large as the number of bins in the ``freqdist``. If None, then | |
| it's assumed to be equal to that of the ``freqdist`` | |
| :type bins: int | |
| """ | |
| assert bins is None or bins >= freqdist.B(), ( | |
| "bins parameter must not be less than %d=freqdist.B()" % freqdist.B() | |
| ) | |
| if bins is None: | |
| bins = freqdist.B() | |
| self._freqdist = freqdist | |
| self._T = self._freqdist.B() | |
| self._Z = bins - self._freqdist.B() | |
| self._N = self._freqdist.N() | |
| # self._P0 is P(0), precalculated for efficiency: | |
| if self._N == 0: | |
| # if freqdist is empty, we approximate P(0) by a UniformProbDist: | |
| self._P0 = 1.0 / self._Z | |
| else: | |
| self._P0 = self._T / (self._Z * (self._N + self._T)) | |
| def prob(self, sample): | |
| # inherit docs from ProbDistI | |
| c = self._freqdist[sample] | |
| return c / (self._N + self._T) if c != 0 else self._P0 | |
| def max(self): | |
| return self._freqdist.max() | |
| def samples(self): | |
| return self._freqdist.keys() | |
| def freqdist(self): | |
| return self._freqdist | |
| def discount(self): | |
| raise NotImplementedError() | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ProbDist``. | |
| :rtype: str | |
| """ | |
| return "<WittenBellProbDist based on %d samples>" % self._freqdist.N() | |
| ##////////////////////////////////////////////////////// | |
| ## Good-Turing Probability Distributions | |
| ##////////////////////////////////////////////////////// | |
| # Good-Turing frequency estimation was contributed by Alan Turing and | |
| # his statistical assistant I.J. Good, during their collaboration in | |
| # the WWII. It is a statistical technique for predicting the | |
| # probability of occurrence of objects belonging to an unknown number | |
| # of species, given past observations of such objects and their | |
| # species. (In drawing balls from an urn, the 'objects' would be balls | |
| # and the 'species' would be the distinct colors of the balls (finite | |
| # but unknown in number). | |
| # | |
| # Good-Turing method calculates the probability mass to assign to | |
| # events with zero or low counts based on the number of events with | |
| # higher counts. It does so by using the adjusted count *c\**: | |
| # | |
| # - *c\* = (c + 1) N(c + 1) / N(c)* for c >= 1 | |
| # - *things with frequency zero in training* = N(1) for c == 0 | |
| # | |
| # where *c* is the original count, *N(i)* is the number of event types | |
| # observed with count *i*. We can think the count of unseen as the count | |
| # of frequency one (see Jurafsky & Martin 2nd Edition, p101). | |
| # | |
| # This method is problematic because the situation ``N(c+1) == 0`` | |
| # is quite common in the original Good-Turing estimation; smoothing or | |
| # interpolation of *N(i)* values is essential in practice. | |
| # | |
| # Bill Gale and Geoffrey Sampson present a simple and effective approach, | |
| # Simple Good-Turing. As a smoothing curve they simply use a power curve: | |
| # | |
| # Nr = a*r^b (with b < -1 to give the appropriate hyperbolic | |
| # relationship) | |
| # | |
| # They estimate a and b by simple linear regression technique on the | |
| # logarithmic form of the equation: | |
| # | |
| # log Nr = a + b*log(r) | |
| # | |
| # However, they suggest that such a simple curve is probably only | |
| # appropriate for high values of r. For low values of r, they use the | |
| # measured Nr directly. (see M&S, p.213) | |
| # | |
| # Gale and Sampson propose to use r while the difference between r and | |
| # r* is 1.96 greater than the standard deviation, and switch to r* if | |
| # it is less or equal: | |
| # | |
| # |r - r*| > 1.96 * sqrt((r + 1)^2 (Nr+1 / Nr^2) (1 + Nr+1 / Nr)) | |
| # | |
| # The 1.96 coefficient correspond to a 0.05 significance criterion, | |
| # some implementations can use a coefficient of 1.65 for a 0.1 | |
| # significance criterion. | |
| # | |
| ##////////////////////////////////////////////////////// | |
| ## Simple Good-Turing Probablity Distributions | |
| ##////////////////////////////////////////////////////// | |
| class SimpleGoodTuringProbDist(ProbDistI): | |
| """ | |
| SimpleGoodTuring ProbDist approximates from frequency to frequency of | |
| frequency into a linear line under log space by linear regression. | |
| Details of Simple Good-Turing algorithm can be found in: | |
| - Good Turing smoothing without tears" (Gale & Sampson 1995), | |
| Journal of Quantitative Linguistics, vol. 2 pp. 217-237. | |
| - "Speech and Language Processing (Jurafsky & Martin), | |
| 2nd Edition, Chapter 4.5 p103 (log(Nc) = a + b*log(c)) | |
| - https://www.grsampson.net/RGoodTur.html | |
| Given a set of pair (xi, yi), where the xi denotes the frequency and | |
| yi denotes the frequency of frequency, we want to minimize their | |
| square variation. E(x) and E(y) represent the mean of xi and yi. | |
| - slope: b = sigma ((xi-E(x)(yi-E(y))) / sigma ((xi-E(x))(xi-E(x))) | |
| - intercept: a = E(y) - b.E(x) | |
| """ | |
| SUM_TO_ONE = False | |
| def __init__(self, freqdist, bins=None): | |
| """ | |
| :param freqdist: The frequency counts upon which to base the | |
| estimation. | |
| :type freqdist: FreqDist | |
| :param bins: The number of possible event types. This must be | |
| larger than the number of bins in the ``freqdist``. If None, | |
| then it's assumed to be equal to ``freqdist``.B() + 1 | |
| :type bins: int | |
| """ | |
| assert ( | |
| bins is None or bins > freqdist.B() | |
| ), "bins parameter must not be less than %d=freqdist.B()+1" % (freqdist.B() + 1) | |
| if bins is None: | |
| bins = freqdist.B() + 1 | |
| self._freqdist = freqdist | |
| self._bins = bins | |
| r, nr = self._r_Nr() | |
| self.find_best_fit(r, nr) | |
| self._switch(r, nr) | |
| self._renormalize(r, nr) | |
| def _r_Nr_non_zero(self): | |
| r_Nr = self._freqdist.r_Nr() | |
| del r_Nr[0] | |
| return r_Nr | |
| def _r_Nr(self): | |
| """ | |
| Split the frequency distribution in two list (r, Nr), where Nr(r) > 0 | |
| """ | |
| nonzero = self._r_Nr_non_zero() | |
| if not nonzero: | |
| return [], [] | |
| return zip(*sorted(nonzero.items())) | |
| def find_best_fit(self, r, nr): | |
| """ | |
| Use simple linear regression to tune parameters self._slope and | |
| self._intercept in the log-log space based on count and Nr(count) | |
| (Work in log space to avoid floating point underflow.) | |
| """ | |
| # For higher sample frequencies the data points becomes horizontal | |
| # along line Nr=1. To create a more evident linear model in log-log | |
| # space, we average positive Nr values with the surrounding zero | |
| # values. (Church and Gale, 1991) | |
| if not r or not nr: | |
| # Empty r or nr? | |
| return | |
| zr = [] | |
| for j in range(len(r)): | |
| i = r[j - 1] if j > 0 else 0 | |
| k = 2 * r[j] - i if j == len(r) - 1 else r[j + 1] | |
| zr_ = 2.0 * nr[j] / (k - i) | |
| zr.append(zr_) | |
| log_r = [math.log(i) for i in r] | |
| log_zr = [math.log(i) for i in zr] | |
| xy_cov = x_var = 0.0 | |
| x_mean = sum(log_r) / len(log_r) | |
| y_mean = sum(log_zr) / len(log_zr) | |
| for (x, y) in zip(log_r, log_zr): | |
| xy_cov += (x - x_mean) * (y - y_mean) | |
| x_var += (x - x_mean) ** 2 | |
| self._slope = xy_cov / x_var if x_var != 0 else 0.0 | |
| if self._slope >= -1: | |
| warnings.warn( | |
| "SimpleGoodTuring did not find a proper best fit " | |
| "line for smoothing probabilities of occurrences. " | |
| "The probability estimates are likely to be " | |
| "unreliable." | |
| ) | |
| self._intercept = y_mean - self._slope * x_mean | |
| def _switch(self, r, nr): | |
| """ | |
| Calculate the r frontier where we must switch from Nr to Sr | |
| when estimating E[Nr]. | |
| """ | |
| for i, r_ in enumerate(r): | |
| if len(r) == i + 1 or r[i + 1] != r_ + 1: | |
| # We are at the end of r, or there is a gap in r | |
| self._switch_at = r_ | |
| break | |
| Sr = self.smoothedNr | |
| smooth_r_star = (r_ + 1) * Sr(r_ + 1) / Sr(r_) | |
| unsmooth_r_star = (r_ + 1) * nr[i + 1] / nr[i] | |
| std = math.sqrt(self._variance(r_, nr[i], nr[i + 1])) | |
| if abs(unsmooth_r_star - smooth_r_star) <= 1.96 * std: | |
| self._switch_at = r_ | |
| break | |
| def _variance(self, r, nr, nr_1): | |
| r = float(r) | |
| nr = float(nr) | |
| nr_1 = float(nr_1) | |
| return (r + 1.0) ** 2 * (nr_1 / nr**2) * (1.0 + nr_1 / nr) | |
| def _renormalize(self, r, nr): | |
| """ | |
| It is necessary to renormalize all the probability estimates to | |
| ensure a proper probability distribution results. This can be done | |
| by keeping the estimate of the probability mass for unseen items as | |
| N(1)/N and renormalizing all the estimates for previously seen items | |
| (as Gale and Sampson (1995) propose). (See M&S P.213, 1999) | |
| """ | |
| prob_cov = 0.0 | |
| for r_, nr_ in zip(r, nr): | |
| prob_cov += nr_ * self._prob_measure(r_) | |
| if prob_cov: | |
| self._renormal = (1 - self._prob_measure(0)) / prob_cov | |
| def smoothedNr(self, r): | |
| """ | |
| Return the number of samples with count r. | |
| :param r: The amount of frequency. | |
| :type r: int | |
| :rtype: float | |
| """ | |
| # Nr = a*r^b (with b < -1 to give the appropriate hyperbolic | |
| # relationship) | |
| # Estimate a and b by simple linear regression technique on | |
| # the logarithmic form of the equation: log Nr = a + b*log(r) | |
| return math.exp(self._intercept + self._slope * math.log(r)) | |
| def prob(self, sample): | |
| """ | |
| Return the sample's probability. | |
| :param sample: sample of the event | |
| :type sample: str | |
| :rtype: float | |
| """ | |
| count = self._freqdist[sample] | |
| p = self._prob_measure(count) | |
| if count == 0: | |
| if self._bins == self._freqdist.B(): | |
| p = 0.0 | |
| else: | |
| p = p / (self._bins - self._freqdist.B()) | |
| else: | |
| p = p * self._renormal | |
| return p | |
| def _prob_measure(self, count): | |
| if count == 0 and self._freqdist.N() == 0: | |
| return 1.0 | |
| elif count == 0 and self._freqdist.N() != 0: | |
| return self._freqdist.Nr(1) / self._freqdist.N() | |
| if self._switch_at > count: | |
| Er_1 = self._freqdist.Nr(count + 1) | |
| Er = self._freqdist.Nr(count) | |
| else: | |
| Er_1 = self.smoothedNr(count + 1) | |
| Er = self.smoothedNr(count) | |
| r_star = (count + 1) * Er_1 / Er | |
| return r_star / self._freqdist.N() | |
| def check(self): | |
| prob_sum = 0.0 | |
| for i in range(0, len(self._Nr)): | |
| prob_sum += self._Nr[i] * self._prob_measure(i) / self._renormal | |
| print("Probability Sum:", prob_sum) | |
| # assert prob_sum != 1.0, "probability sum should be one!" | |
| def discount(self): | |
| """ | |
| This function returns the total mass of probability transfers from the | |
| seen samples to the unseen samples. | |
| """ | |
| return self.smoothedNr(1) / self._freqdist.N() | |
| def max(self): | |
| return self._freqdist.max() | |
| def samples(self): | |
| return self._freqdist.keys() | |
| def freqdist(self): | |
| return self._freqdist | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ProbDist``. | |
| :rtype: str | |
| """ | |
| return "<SimpleGoodTuringProbDist based on %d samples>" % self._freqdist.N() | |
| class MutableProbDist(ProbDistI): | |
| """ | |
| An mutable probdist where the probabilities may be easily modified. This | |
| simply copies an existing probdist, storing the probability values in a | |
| mutable dictionary and providing an update method. | |
| """ | |
| def __init__(self, prob_dist, samples, store_logs=True): | |
| """ | |
| Creates the mutable probdist based on the given prob_dist and using | |
| the list of samples given. These values are stored as log | |
| probabilities if the store_logs flag is set. | |
| :param prob_dist: the distribution from which to garner the | |
| probabilities | |
| :type prob_dist: ProbDist | |
| :param samples: the complete set of samples | |
| :type samples: sequence of any | |
| :param store_logs: whether to store the probabilities as logarithms | |
| :type store_logs: bool | |
| """ | |
| self._samples = samples | |
| self._sample_dict = {samples[i]: i for i in range(len(samples))} | |
| self._data = array.array("d", [0.0]) * len(samples) | |
| for i in range(len(samples)): | |
| if store_logs: | |
| self._data[i] = prob_dist.logprob(samples[i]) | |
| else: | |
| self._data[i] = prob_dist.prob(samples[i]) | |
| self._logs = store_logs | |
| def max(self): | |
| # inherit documentation | |
| return max((p, v) for (v, p) in self._sample_dict.items())[1] | |
| def samples(self): | |
| # inherit documentation | |
| return self._samples | |
| def prob(self, sample): | |
| # inherit documentation | |
| i = self._sample_dict.get(sample) | |
| if i is None: | |
| return 0.0 | |
| return 2 ** (self._data[i]) if self._logs else self._data[i] | |
| def logprob(self, sample): | |
| # inherit documentation | |
| i = self._sample_dict.get(sample) | |
| if i is None: | |
| return float("-inf") | |
| return self._data[i] if self._logs else math.log(self._data[i], 2) | |
| def update(self, sample, prob, log=True): | |
| """ | |
| Update the probability for the given sample. This may cause the object | |
| to stop being the valid probability distribution - the user must | |
| ensure that they update the sample probabilities such that all samples | |
| have probabilities between 0 and 1 and that all probabilities sum to | |
| one. | |
| :param sample: the sample for which to update the probability | |
| :type sample: any | |
| :param prob: the new probability | |
| :type prob: float | |
| :param log: is the probability already logged | |
| :type log: bool | |
| """ | |
| i = self._sample_dict.get(sample) | |
| assert i is not None | |
| if self._logs: | |
| self._data[i] = prob if log else math.log(prob, 2) | |
| else: | |
| self._data[i] = 2 ** (prob) if log else prob | |
| ##///////////////////////////////////////////////////// | |
| ## Kneser-Ney Probability Distribution | |
| ##////////////////////////////////////////////////////// | |
| # This method for calculating probabilities was introduced in 1995 by Reinhard | |
| # Kneser and Hermann Ney. It was meant to improve the accuracy of language | |
| # models that use backing-off to deal with sparse data. The authors propose two | |
| # ways of doing so: a marginal distribution constraint on the back-off | |
| # distribution and a leave-one-out distribution. For a start, the first one is | |
| # implemented as a class below. | |
| # | |
| # The idea behind a back-off n-gram model is that we have a series of | |
| # frequency distributions for our n-grams so that in case we have not seen a | |
| # given n-gram during training (and as a result have a 0 probability for it) we | |
| # can 'back off' (hence the name!) and try testing whether we've seen the | |
| # n-1-gram part of the n-gram in training. | |
| # | |
| # The novelty of Kneser and Ney's approach was that they decided to fiddle | |
| # around with the way this latter, backed off probability was being calculated | |
| # whereas their peers seemed to focus on the primary probability. | |
| # | |
| # The implementation below uses one of the techniques described in their paper | |
| # titled "Improved backing-off for n-gram language modeling." In the same paper | |
| # another technique is introduced to attempt to smooth the back-off | |
| # distribution as well as the primary one. There is also a much-cited | |
| # modification of this method proposed by Chen and Goodman. | |
| # | |
| # In order for the implementation of Kneser-Ney to be more efficient, some | |
| # changes have been made to the original algorithm. Namely, the calculation of | |
| # the normalizing function gamma has been significantly simplified and | |
| # combined slightly differently with beta. None of these changes affect the | |
| # nature of the algorithm, but instead aim to cut out unnecessary calculations | |
| # and take advantage of storing and retrieving information in dictionaries | |
| # where possible. | |
| class KneserNeyProbDist(ProbDistI): | |
| """ | |
| Kneser-Ney estimate of a probability distribution. This is a version of | |
| back-off that counts how likely an n-gram is provided the n-1-gram had | |
| been seen in training. Extends the ProbDistI interface, requires a trigram | |
| FreqDist instance to train on. Optionally, a different from default discount | |
| value can be specified. The default discount is set to 0.75. | |
| """ | |
| def __init__(self, freqdist, bins=None, discount=0.75): | |
| """ | |
| :param freqdist: The trigram frequency distribution upon which to base | |
| the estimation | |
| :type freqdist: FreqDist | |
| :param bins: Included for compatibility with nltk.tag.hmm | |
| :type bins: int or float | |
| :param discount: The discount applied when retrieving counts of | |
| trigrams | |
| :type discount: float (preferred, but can be set to int) | |
| """ | |
| if not bins: | |
| self._bins = freqdist.B() | |
| else: | |
| self._bins = bins | |
| self._D = discount | |
| # cache for probability calculation | |
| self._cache = {} | |
| # internal bigram and trigram frequency distributions | |
| self._bigrams = defaultdict(int) | |
| self._trigrams = freqdist | |
| # helper dictionaries used to calculate probabilities | |
| self._wordtypes_after = defaultdict(float) | |
| self._trigrams_contain = defaultdict(float) | |
| self._wordtypes_before = defaultdict(float) | |
| for w0, w1, w2 in freqdist: | |
| self._bigrams[(w0, w1)] += freqdist[(w0, w1, w2)] | |
| self._wordtypes_after[(w0, w1)] += 1 | |
| self._trigrams_contain[w1] += 1 | |
| self._wordtypes_before[(w1, w2)] += 1 | |
| def prob(self, trigram): | |
| # sample must be a triple | |
| if len(trigram) != 3: | |
| raise ValueError("Expected an iterable with 3 members.") | |
| trigram = tuple(trigram) | |
| w0, w1, w2 = trigram | |
| if trigram in self._cache: | |
| return self._cache[trigram] | |
| else: | |
| # if the sample trigram was seen during training | |
| if trigram in self._trigrams: | |
| prob = (self._trigrams[trigram] - self.discount()) / self._bigrams[ | |
| (w0, w1) | |
| ] | |
| # else if the 'rougher' environment was seen during training | |
| elif (w0, w1) in self._bigrams and (w1, w2) in self._wordtypes_before: | |
| aftr = self._wordtypes_after[(w0, w1)] | |
| bfr = self._wordtypes_before[(w1, w2)] | |
| # the probability left over from alphas | |
| leftover_prob = (aftr * self.discount()) / self._bigrams[(w0, w1)] | |
| # the beta (including normalization) | |
| beta = bfr / (self._trigrams_contain[w1] - aftr) | |
| prob = leftover_prob * beta | |
| # else the sample was completely unseen during training | |
| else: | |
| prob = 0.0 | |
| self._cache[trigram] = prob | |
| return prob | |
| def discount(self): | |
| """ | |
| Return the value by which counts are discounted. By default set to 0.75. | |
| :rtype: float | |
| """ | |
| return self._D | |
| def set_discount(self, discount): | |
| """ | |
| Set the value by which counts are discounted to the value of discount. | |
| :param discount: the new value to discount counts by | |
| :type discount: float (preferred, but int possible) | |
| :rtype: None | |
| """ | |
| self._D = discount | |
| def samples(self): | |
| return self._trigrams.keys() | |
| def max(self): | |
| return self._trigrams.max() | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ProbDist | |
| :rtype: str | |
| """ | |
| return f"<KneserNeyProbDist based on {self._trigrams.N()} trigrams" | |
| ##////////////////////////////////////////////////////// | |
| ## Probability Distribution Operations | |
| ##////////////////////////////////////////////////////// | |
| def log_likelihood(test_pdist, actual_pdist): | |
| if not isinstance(test_pdist, ProbDistI) or not isinstance(actual_pdist, ProbDistI): | |
| raise ValueError("expected a ProbDist.") | |
| # Is this right? | |
| return sum( | |
| actual_pdist.prob(s) * math.log(test_pdist.prob(s), 2) for s in actual_pdist | |
| ) | |
| def entropy(pdist): | |
| probs = (pdist.prob(s) for s in pdist.samples()) | |
| return -sum(p * math.log(p, 2) for p in probs) | |
| ##////////////////////////////////////////////////////// | |
| ## Conditional Distributions | |
| ##////////////////////////////////////////////////////// | |
| class ConditionalFreqDist(defaultdict): | |
| """ | |
| A collection of frequency distributions for a single experiment | |
| run under different conditions. Conditional frequency | |
| distributions are used to record the number of times each sample | |
| occurred, given the condition under which the experiment was run. | |
| For example, a conditional frequency distribution could be used to | |
| record the frequency of each word (type) in a document, given its | |
| length. Formally, a conditional frequency distribution can be | |
| defined as a function that maps from each condition to the | |
| FreqDist for the experiment under that condition. | |
| Conditional frequency distributions are typically constructed by | |
| repeatedly running an experiment under a variety of conditions, | |
| and incrementing the sample outcome counts for the appropriate | |
| conditions. For example, the following code will produce a | |
| conditional frequency distribution that encodes how often each | |
| word type occurs, given the length of that word type: | |
| >>> from nltk.probability import ConditionalFreqDist | |
| >>> from nltk.tokenize import word_tokenize | |
| >>> sent = "the the the dog dog some other words that we do not care about" | |
| >>> cfdist = ConditionalFreqDist() | |
| >>> for word in word_tokenize(sent): | |
| ... condition = len(word) | |
| ... cfdist[condition][word] += 1 | |
| An equivalent way to do this is with the initializer: | |
| >>> cfdist = ConditionalFreqDist((len(word), word) for word in word_tokenize(sent)) | |
| The frequency distribution for each condition is accessed using | |
| the indexing operator: | |
| >>> cfdist[3] | |
| FreqDist({'the': 3, 'dog': 2, 'not': 1}) | |
| >>> cfdist[3].freq('the') | |
| 0.5 | |
| >>> cfdist[3]['dog'] | |
| 2 | |
| When the indexing operator is used to access the frequency | |
| distribution for a condition that has not been accessed before, | |
| ``ConditionalFreqDist`` creates a new empty FreqDist for that | |
| condition. | |
| """ | |
| def __init__(self, cond_samples=None): | |
| """ | |
| Construct a new empty conditional frequency distribution. In | |
| particular, the count for every sample, under every condition, | |
| is zero. | |
| :param cond_samples: The samples to initialize the conditional | |
| frequency distribution with | |
| :type cond_samples: Sequence of (condition, sample) tuples | |
| """ | |
| defaultdict.__init__(self, FreqDist) | |
| if cond_samples: | |
| for (cond, sample) in cond_samples: | |
| self[cond][sample] += 1 | |
| def __reduce__(self): | |
| kv_pairs = ((cond, self[cond]) for cond in self.conditions()) | |
| return (self.__class__, (), None, None, kv_pairs) | |
| def conditions(self): | |
| """ | |
| Return a list of the conditions that have been accessed for | |
| this ``ConditionalFreqDist``. Use the indexing operator to | |
| access the frequency distribution for a given condition. | |
| Note that the frequency distributions for some conditions | |
| may contain zero sample outcomes. | |
| :rtype: list | |
| """ | |
| return list(self.keys()) | |
| def N(self): | |
| """ | |
| Return the total number of sample outcomes that have been | |
| recorded by this ``ConditionalFreqDist``. | |
| :rtype: int | |
| """ | |
| return sum(fdist.N() for fdist in self.values()) | |
| def plot( | |
| self, | |
| *args, | |
| samples=None, | |
| title="", | |
| cumulative=False, | |
| percents=False, | |
| conditions=None, | |
| show=True, | |
| **kwargs, | |
| ): | |
| """ | |
| Plot the given samples from the conditional frequency distribution. | |
| For a cumulative plot, specify cumulative=True. Additional ``*args`` and | |
| ``**kwargs`` are passed to matplotlib's plot function. | |
| (Requires Matplotlib to be installed.) | |
| :param samples: The samples to plot | |
| :type samples: list | |
| :param title: The title for the graph | |
| :type title: str | |
| :param cumulative: Whether the plot is cumulative. (default = False) | |
| :type cumulative: bool | |
| :param percents: Whether the plot uses percents instead of counts. (default = False) | |
| :type percents: bool | |
| :param conditions: The conditions to plot (default is all) | |
| :type conditions: list | |
| :param show: Whether to show the plot, or only return the ax. | |
| :type show: bool | |
| """ | |
| try: | |
| import matplotlib.pyplot as plt # import statement fix | |
| except ImportError as e: | |
| raise ValueError( | |
| "The plot function requires matplotlib to be installed." | |
| "See https://matplotlib.org/" | |
| ) from e | |
| if not conditions: | |
| conditions = self.conditions() | |
| else: | |
| conditions = [c for c in conditions if c in self] | |
| if not samples: | |
| samples = sorted({v for c in conditions for v in self[c]}) | |
| if "linewidth" not in kwargs: | |
| kwargs["linewidth"] = 2 | |
| ax = plt.gca() | |
| if conditions: | |
| freqs = [] | |
| for condition in conditions: | |
| if cumulative: | |
| # freqs should be a list of list where each sub list will be a frequency of a condition | |
| freq = list(self[condition]._cumulative_frequencies(samples)) | |
| else: | |
| freq = [self[condition][sample] for sample in samples] | |
| if percents: | |
| freq = [f / self[condition].N() * 100 for f in freq] | |
| freqs.append(freq) | |
| if cumulative: | |
| ylabel = "Cumulative " | |
| legend_loc = "lower right" | |
| else: | |
| ylabel = "" | |
| legend_loc = "upper right" | |
| if percents: | |
| ylabel += "Percents" | |
| else: | |
| ylabel += "Counts" | |
| i = 0 | |
| for freq in freqs: | |
| kwargs["label"] = conditions[i] # label for each condition | |
| i += 1 | |
| ax.plot(freq, *args, **kwargs) | |
| ax.legend(loc=legend_loc) | |
| ax.grid(True, color="silver") | |
| ax.set_xticks(range(len(samples))) | |
| ax.set_xticklabels([str(s) for s in samples], rotation=90) | |
| if title: | |
| ax.set_title(title) | |
| ax.set_xlabel("Samples") | |
| ax.set_ylabel(ylabel) | |
| if show: | |
| plt.show() | |
| return ax | |
| def tabulate(self, *args, **kwargs): | |
| """ | |
| Tabulate the given samples from the conditional frequency distribution. | |
| :param samples: The samples to plot | |
| :type samples: list | |
| :param conditions: The conditions to plot (default is all) | |
| :type conditions: list | |
| :param cumulative: A flag to specify whether the freqs are cumulative (default = False) | |
| :type title: bool | |
| """ | |
| cumulative = _get_kwarg(kwargs, "cumulative", False) | |
| conditions = _get_kwarg(kwargs, "conditions", sorted(self.conditions())) | |
| samples = _get_kwarg( | |
| kwargs, | |
| "samples", | |
| sorted({v for c in conditions if c in self for v in self[c]}), | |
| ) # this computation could be wasted | |
| width = max(len("%s" % s) for s in samples) | |
| freqs = dict() | |
| for c in conditions: | |
| if cumulative: | |
| freqs[c] = list(self[c]._cumulative_frequencies(samples)) | |
| else: | |
| freqs[c] = [self[c][sample] for sample in samples] | |
| width = max(width, max(len("%d" % f) for f in freqs[c])) | |
| condition_size = max(len("%s" % c) for c in conditions) | |
| print(" " * condition_size, end=" ") | |
| for s in samples: | |
| print("%*s" % (width, s), end=" ") | |
| print() | |
| for c in conditions: | |
| print("%*s" % (condition_size, c), end=" ") | |
| for f in freqs[c]: | |
| print("%*d" % (width, f), end=" ") | |
| print() | |
| # Mathematical operators | |
| def __add__(self, other): | |
| """ | |
| Add counts from two ConditionalFreqDists. | |
| """ | |
| if not isinstance(other, ConditionalFreqDist): | |
| return NotImplemented | |
| result = self.copy() | |
| for cond in other.conditions(): | |
| result[cond] += other[cond] | |
| return result | |
| def __sub__(self, other): | |
| """ | |
| Subtract count, but keep only results with positive counts. | |
| """ | |
| if not isinstance(other, ConditionalFreqDist): | |
| return NotImplemented | |
| result = self.copy() | |
| for cond in other.conditions(): | |
| result[cond] -= other[cond] | |
| if not result[cond]: | |
| del result[cond] | |
| return result | |
| def __or__(self, other): | |
| """ | |
| Union is the maximum of value in either of the input counters. | |
| """ | |
| if not isinstance(other, ConditionalFreqDist): | |
| return NotImplemented | |
| result = self.copy() | |
| for cond in other.conditions(): | |
| result[cond] |= other[cond] | |
| return result | |
| def __and__(self, other): | |
| """ | |
| Intersection is the minimum of corresponding counts. | |
| """ | |
| if not isinstance(other, ConditionalFreqDist): | |
| return NotImplemented | |
| result = ConditionalFreqDist() | |
| for cond in self.conditions(): | |
| newfreqdist = self[cond] & other[cond] | |
| if newfreqdist: | |
| result[cond] = newfreqdist | |
| return result | |
| # @total_ordering doesn't work here, since the class inherits from a builtin class | |
| def __le__(self, other): | |
| if not isinstance(other, ConditionalFreqDist): | |
| raise_unorderable_types("<=", self, other) | |
| return set(self.conditions()).issubset(other.conditions()) and all( | |
| self[c] <= other[c] for c in self.conditions() | |
| ) | |
| def __lt__(self, other): | |
| if not isinstance(other, ConditionalFreqDist): | |
| raise_unorderable_types("<", self, other) | |
| return self <= other and self != other | |
| def __ge__(self, other): | |
| if not isinstance(other, ConditionalFreqDist): | |
| raise_unorderable_types(">=", self, other) | |
| return other <= self | |
| def __gt__(self, other): | |
| if not isinstance(other, ConditionalFreqDist): | |
| raise_unorderable_types(">", self, other) | |
| return other < self | |
| def deepcopy(self): | |
| from copy import deepcopy | |
| return deepcopy(self) | |
| copy = deepcopy | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ConditionalFreqDist``. | |
| :rtype: str | |
| """ | |
| return "<ConditionalFreqDist with %d conditions>" % len(self) | |
| class ConditionalProbDistI(dict, metaclass=ABCMeta): | |
| """ | |
| A collection of probability distributions for a single experiment | |
| run under different conditions. Conditional probability | |
| distributions are used to estimate the likelihood of each sample, | |
| given the condition under which the experiment was run. For | |
| example, a conditional probability distribution could be used to | |
| estimate the probability of each word type in a document, given | |
| the length of the word type. Formally, a conditional probability | |
| distribution can be defined as a function that maps from each | |
| condition to the ``ProbDist`` for the experiment under that | |
| condition. | |
| """ | |
| def __init__(self): | |
| """ | |
| Classes inheriting from ConditionalProbDistI should implement __init__. | |
| """ | |
| def conditions(self): | |
| """ | |
| Return a list of the conditions that are represented by | |
| this ``ConditionalProbDist``. Use the indexing operator to | |
| access the probability distribution for a given condition. | |
| :rtype: list | |
| """ | |
| return list(self.keys()) | |
| def __repr__(self): | |
| """ | |
| Return a string representation of this ``ConditionalProbDist``. | |
| :rtype: str | |
| """ | |
| return "<%s with %d conditions>" % (type(self).__name__, len(self)) | |
| class ConditionalProbDist(ConditionalProbDistI): | |
| """ | |
| A conditional probability distribution modeling the experiments | |
| that were used to generate a conditional frequency distribution. | |
| A ConditionalProbDist is constructed from a | |
| ``ConditionalFreqDist`` and a ``ProbDist`` factory: | |
| - The ``ConditionalFreqDist`` specifies the frequency | |
| distribution for each condition. | |
| - The ``ProbDist`` factory is a function that takes a | |
| condition's frequency distribution, and returns its | |
| probability distribution. A ``ProbDist`` class's name (such as | |
| ``MLEProbDist`` or ``HeldoutProbDist``) can be used to specify | |
| that class's constructor. | |
| The first argument to the ``ProbDist`` factory is the frequency | |
| distribution that it should model; and the remaining arguments are | |
| specified by the ``factory_args`` parameter to the | |
| ``ConditionalProbDist`` constructor. For example, the following | |
| code constructs a ``ConditionalProbDist``, where the probability | |
| distribution for each condition is an ``ELEProbDist`` with 10 bins: | |
| >>> from nltk.corpus import brown | |
| >>> from nltk.probability import ConditionalFreqDist | |
| >>> from nltk.probability import ConditionalProbDist, ELEProbDist | |
| >>> cfdist = ConditionalFreqDist(brown.tagged_words()[:5000]) | |
| >>> cpdist = ConditionalProbDist(cfdist, ELEProbDist, 10) | |
| >>> cpdist['passed'].max() | |
| 'VBD' | |
| >>> cpdist['passed'].prob('VBD') #doctest: +ELLIPSIS | |
| 0.423... | |
| """ | |
| def __init__(self, cfdist, probdist_factory, *factory_args, **factory_kw_args): | |
| """ | |
| Construct a new conditional probability distribution, based on | |
| the given conditional frequency distribution and ``ProbDist`` | |
| factory. | |
| :type cfdist: ConditionalFreqDist | |
| :param cfdist: The ``ConditionalFreqDist`` specifying the | |
| frequency distribution for each condition. | |
| :type probdist_factory: class or function | |
| :param probdist_factory: The function or class that maps | |
| a condition's frequency distribution to its probability | |
| distribution. The function is called with the frequency | |
| distribution as its first argument, | |
| ``factory_args`` as its remaining arguments, and | |
| ``factory_kw_args`` as keyword arguments. | |
| :type factory_args: (any) | |
| :param factory_args: Extra arguments for ``probdist_factory``. | |
| These arguments are usually used to specify extra | |
| properties for the probability distributions of individual | |
| conditions, such as the number of bins they contain. | |
| :type factory_kw_args: (any) | |
| :param factory_kw_args: Extra keyword arguments for ``probdist_factory``. | |
| """ | |
| self._probdist_factory = probdist_factory | |
| self._factory_args = factory_args | |
| self._factory_kw_args = factory_kw_args | |
| for condition in cfdist: | |
| self[condition] = probdist_factory( | |
| cfdist[condition], *factory_args, **factory_kw_args | |
| ) | |
| def __missing__(self, key): | |
| self[key] = self._probdist_factory( | |
| FreqDist(), *self._factory_args, **self._factory_kw_args | |
| ) | |
| return self[key] | |
| class DictionaryConditionalProbDist(ConditionalProbDistI): | |
| """ | |
| An alternative ConditionalProbDist that simply wraps a dictionary of | |
| ProbDists rather than creating these from FreqDists. | |
| """ | |
| def __init__(self, probdist_dict): | |
| """ | |
| :param probdist_dict: a dictionary containing the probdists indexed | |
| by the conditions | |
| :type probdist_dict: dict any -> probdist | |
| """ | |
| self.update(probdist_dict) | |
| def __missing__(self, key): | |
| self[key] = DictionaryProbDist() | |
| return self[key] | |
| ##////////////////////////////////////////////////////// | |
| ## Adding in log-space. | |
| ##////////////////////////////////////////////////////// | |
| # If the difference is bigger than this, then just take the bigger one: | |
| _ADD_LOGS_MAX_DIFF = math.log(1e-30, 2) | |
| def add_logs(logx, logy): | |
| """ | |
| Given two numbers ``logx`` = *log(x)* and ``logy`` = *log(y)*, return | |
| *log(x+y)*. Conceptually, this is the same as returning | |
| ``log(2**(logx)+2**(logy))``, but the actual implementation | |
| avoids overflow errors that could result from direct computation. | |
| """ | |
| if logx < logy + _ADD_LOGS_MAX_DIFF: | |
| return logy | |
| if logy < logx + _ADD_LOGS_MAX_DIFF: | |
| return logx | |
| base = min(logx, logy) | |
| return base + math.log(2 ** (logx - base) + 2 ** (logy - base), 2) | |
| def sum_logs(logs): | |
| return reduce(add_logs, logs[1:], logs[0]) if len(logs) != 0 else _NINF | |
| ##////////////////////////////////////////////////////// | |
| ## Probabilistic Mix-in | |
| ##////////////////////////////////////////////////////// | |
| class ProbabilisticMixIn: | |
| """ | |
| A mix-in class to associate probabilities with other classes | |
| (trees, rules, etc.). To use the ``ProbabilisticMixIn`` class, | |
| define a new class that derives from an existing class and from | |
| ProbabilisticMixIn. You will need to define a new constructor for | |
| the new class, which explicitly calls the constructors of both its | |
| parent classes. For example: | |
| >>> from nltk.probability import ProbabilisticMixIn | |
| >>> class A: | |
| ... def __init__(self, x, y): self.data = (x,y) | |
| ... | |
| >>> class ProbabilisticA(A, ProbabilisticMixIn): | |
| ... def __init__(self, x, y, **prob_kwarg): | |
| ... A.__init__(self, x, y) | |
| ... ProbabilisticMixIn.__init__(self, **prob_kwarg) | |
| See the documentation for the ProbabilisticMixIn | |
| ``constructor<__init__>`` for information about the arguments it | |
| expects. | |
| You should generally also redefine the string representation | |
| methods, the comparison methods, and the hashing method. | |
| """ | |
| def __init__(self, **kwargs): | |
| """ | |
| Initialize this object's probability. This initializer should | |
| be called by subclass constructors. ``prob`` should generally be | |
| the first argument for those constructors. | |
| :param prob: The probability associated with the object. | |
| :type prob: float | |
| :param logprob: The log of the probability associated with | |
| the object. | |
| :type logprob: float | |
| """ | |
| if "prob" in kwargs: | |
| if "logprob" in kwargs: | |
| raise TypeError("Must specify either prob or logprob " "(not both)") | |
| else: | |
| ProbabilisticMixIn.set_prob(self, kwargs["prob"]) | |
| elif "logprob" in kwargs: | |
| ProbabilisticMixIn.set_logprob(self, kwargs["logprob"]) | |
| else: | |
| self.__prob = self.__logprob = None | |
| def set_prob(self, prob): | |
| """ | |
| Set the probability associated with this object to ``prob``. | |
| :param prob: The new probability | |
| :type prob: float | |
| """ | |
| self.__prob = prob | |
| self.__logprob = None | |
| def set_logprob(self, logprob): | |
| """ | |
| Set the log probability associated with this object to | |
| ``logprob``. I.e., set the probability associated with this | |
| object to ``2**(logprob)``. | |
| :param logprob: The new log probability | |
| :type logprob: float | |
| """ | |
| self.__logprob = logprob | |
| self.__prob = None | |
| def prob(self): | |
| """ | |
| Return the probability associated with this object. | |
| :rtype: float | |
| """ | |
| if self.__prob is None: | |
| if self.__logprob is None: | |
| return None | |
| self.__prob = 2 ** (self.__logprob) | |
| return self.__prob | |
| def logprob(self): | |
| """ | |
| Return ``log(p)``, where ``p`` is the probability associated | |
| with this object. | |
| :rtype: float | |
| """ | |
| if self.__logprob is None: | |
| if self.__prob is None: | |
| return None | |
| self.__logprob = math.log(self.__prob, 2) | |
| return self.__logprob | |
| class ImmutableProbabilisticMixIn(ProbabilisticMixIn): | |
| def set_prob(self, prob): | |
| raise ValueError("%s is immutable" % self.__class__.__name__) | |
| def set_logprob(self, prob): | |
| raise ValueError("%s is immutable" % self.__class__.__name__) | |
| ## Helper function for processing keyword arguments | |
| def _get_kwarg(kwargs, key, default): | |
| if key in kwargs: | |
| arg = kwargs[key] | |
| del kwargs[key] | |
| else: | |
| arg = default | |
| return arg | |
| ##////////////////////////////////////////////////////// | |
| ## Demonstration | |
| ##////////////////////////////////////////////////////// | |
| def _create_rand_fdist(numsamples, numoutcomes): | |
| """ | |
| Create a new frequency distribution, with random samples. The | |
| samples are numbers from 1 to ``numsamples``, and are generated by | |
| summing two numbers, each of which has a uniform distribution. | |
| """ | |
| fdist = FreqDist() | |
| for x in range(numoutcomes): | |
| y = random.randint(1, (1 + numsamples) // 2) + random.randint( | |
| 0, numsamples // 2 | |
| ) | |
| fdist[y] += 1 | |
| return fdist | |
| def _create_sum_pdist(numsamples): | |
| """ | |
| Return the true probability distribution for the experiment | |
| ``_create_rand_fdist(numsamples, x)``. | |
| """ | |
| fdist = FreqDist() | |
| for x in range(1, (1 + numsamples) // 2 + 1): | |
| for y in range(0, numsamples // 2 + 1): | |
| fdist[x + y] += 1 | |
| return MLEProbDist(fdist) | |
| def demo(numsamples=6, numoutcomes=500): | |
| """ | |
| A demonstration of frequency distributions and probability | |
| distributions. This demonstration creates three frequency | |
| distributions with, and uses them to sample a random process with | |
| ``numsamples`` samples. Each frequency distribution is sampled | |
| ``numoutcomes`` times. These three frequency distributions are | |
| then used to build six probability distributions. Finally, the | |
| probability estimates of these distributions are compared to the | |
| actual probability of each sample. | |
| :type numsamples: int | |
| :param numsamples: The number of samples to use in each demo | |
| frequency distributions. | |
| :type numoutcomes: int | |
| :param numoutcomes: The total number of outcomes for each | |
| demo frequency distribution. These outcomes are divided into | |
| ``numsamples`` bins. | |
| :rtype: None | |
| """ | |
| # Randomly sample a stochastic process three times. | |
| fdist1 = _create_rand_fdist(numsamples, numoutcomes) | |
| fdist2 = _create_rand_fdist(numsamples, numoutcomes) | |
| fdist3 = _create_rand_fdist(numsamples, numoutcomes) | |
| # Use our samples to create probability distributions. | |
| pdists = [ | |
| MLEProbDist(fdist1), | |
| LidstoneProbDist(fdist1, 0.5, numsamples), | |
| HeldoutProbDist(fdist1, fdist2, numsamples), | |
| HeldoutProbDist(fdist2, fdist1, numsamples), | |
| CrossValidationProbDist([fdist1, fdist2, fdist3], numsamples), | |
| SimpleGoodTuringProbDist(fdist1), | |
| SimpleGoodTuringProbDist(fdist1, 7), | |
| _create_sum_pdist(numsamples), | |
| ] | |
| # Find the probability of each sample. | |
| vals = [] | |
| for n in range(1, numsamples + 1): | |
| vals.append(tuple([n, fdist1.freq(n)] + [pdist.prob(n) for pdist in pdists])) | |
| # Print the results in a formatted table. | |
| print( | |
| "%d samples (1-%d); %d outcomes were sampled for each FreqDist" | |
| % (numsamples, numsamples, numoutcomes) | |
| ) | |
| print("=" * 9 * (len(pdists) + 2)) | |
| FORMATSTR = " FreqDist " + "%8s " * (len(pdists) - 1) + "| Actual" | |
| print(FORMATSTR % tuple(repr(pdist)[1:9] for pdist in pdists[:-1])) | |
| print("-" * 9 * (len(pdists) + 2)) | |
| FORMATSTR = "%3d %8.6f " + "%8.6f " * (len(pdists) - 1) + "| %8.6f" | |
| for val in vals: | |
| print(FORMATSTR % val) | |
| # Print the totals for each column (should all be 1.0) | |
| zvals = list(zip(*vals)) | |
| sums = [sum(val) for val in zvals[1:]] | |
| print("-" * 9 * (len(pdists) + 2)) | |
| FORMATSTR = "Total " + "%8.6f " * (len(pdists)) + "| %8.6f" | |
| print(FORMATSTR % tuple(sums)) | |
| print("=" * 9 * (len(pdists) + 2)) | |
| # Display the distributions themselves, if they're short enough. | |
| if len("%s" % fdist1) < 70: | |
| print(" fdist1: %s" % fdist1) | |
| print(" fdist2: %s" % fdist2) | |
| print(" fdist3: %s" % fdist3) | |
| print() | |
| print("Generating:") | |
| for pdist in pdists: | |
| fdist = FreqDist(pdist.generate() for i in range(5000)) | |
| print("{:>20} {}".format(pdist.__class__.__name__[:20], ("%s" % fdist)[:55])) | |
| print() | |
| def gt_demo(): | |
| from nltk import corpus | |
| emma_words = corpus.gutenberg.words("austen-emma.txt") | |
| fd = FreqDist(emma_words) | |
| sgt = SimpleGoodTuringProbDist(fd) | |
| print("{:>18} {:>8} {:>14}".format("word", "frequency", "SimpleGoodTuring")) | |
| fd_keys_sorted = ( | |
| key for key, value in sorted(fd.items(), key=lambda item: item[1], reverse=True) | |
| ) | |
| for key in fd_keys_sorted: | |
| print("%18s %8d %14e" % (key, fd[key], sgt.prob(key))) | |
| if __name__ == "__main__": | |
| demo(6, 10) | |
| demo(5, 5000) | |
| gt_demo() | |
| __all__ = [ | |
| "ConditionalFreqDist", | |
| "ConditionalProbDist", | |
| "ConditionalProbDistI", | |
| "CrossValidationProbDist", | |
| "DictionaryConditionalProbDist", | |
| "DictionaryProbDist", | |
| "ELEProbDist", | |
| "FreqDist", | |
| "SimpleGoodTuringProbDist", | |
| "HeldoutProbDist", | |
| "ImmutableProbabilisticMixIn", | |
| "LaplaceProbDist", | |
| "LidstoneProbDist", | |
| "MLEProbDist", | |
| "MutableProbDist", | |
| "KneserNeyProbDist", | |
| "ProbDistI", | |
| "ProbabilisticMixIn", | |
| "UniformProbDist", | |
| "WittenBellProbDist", | |
| "add_logs", | |
| "log_likelihood", | |
| "sum_logs", | |
| "entropy", | |
| ] | |