peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/sklearn
/linear_model
/_perceptron.py
| # Author: Mathieu Blondel | |
| # License: BSD 3 clause | |
| from numbers import Real | |
| from ..utils._param_validation import Interval, StrOptions | |
| from ._stochastic_gradient import BaseSGDClassifier | |
| class Perceptron(BaseSGDClassifier): | |
| """Linear perceptron classifier. | |
| The implementation is a wrapper around :class:`~sklearn.linear_model.SGDClassifier` | |
| by fixing the `loss` and `learning_rate` parameters as:: | |
| SGDClassifier(loss="perceptron", learning_rate="constant") | |
| Other available parameters are described below and are forwarded to | |
| :class:`~sklearn.linear_model.SGDClassifier`. | |
| Read more in the :ref:`User Guide <perceptron>`. | |
| Parameters | |
| ---------- | |
| penalty : {'l2','l1','elasticnet'}, default=None | |
| The penalty (aka regularization term) to be used. | |
| alpha : float, default=0.0001 | |
| Constant that multiplies the regularization term if regularization is | |
| used. | |
| l1_ratio : float, default=0.15 | |
| The Elastic Net mixing parameter, with `0 <= l1_ratio <= 1`. | |
| `l1_ratio=0` corresponds to L2 penalty, `l1_ratio=1` to L1. | |
| Only used if `penalty='elasticnet'`. | |
| .. versionadded:: 0.24 | |
| fit_intercept : bool, default=True | |
| Whether the intercept should be estimated or not. If False, the | |
| data is assumed to be already centered. | |
| max_iter : int, default=1000 | |
| The maximum number of passes over the training data (aka epochs). | |
| It only impacts the behavior in the ``fit`` method, and not the | |
| :meth:`partial_fit` method. | |
| .. versionadded:: 0.19 | |
| tol : float or None, default=1e-3 | |
| The stopping criterion. If it is not None, the iterations will stop | |
| when (loss > previous_loss - tol). | |
| .. versionadded:: 0.19 | |
| shuffle : bool, default=True | |
| Whether or not the training data should be shuffled after each epoch. | |
| verbose : int, default=0 | |
| The verbosity level. | |
| eta0 : float, default=1 | |
| Constant by which the updates are multiplied. | |
| n_jobs : int, default=None | |
| The number of CPUs to use to do the OVA (One Versus All, for | |
| multi-class problems) computation. | |
| ``None`` means 1 unless in a :obj:`joblib.parallel_backend` context. | |
| ``-1`` means using all processors. See :term:`Glossary <n_jobs>` | |
| for more details. | |
| random_state : int, RandomState instance or None, default=0 | |
| Used to shuffle the training data, when ``shuffle`` is set to | |
| ``True``. Pass an int for reproducible output across multiple | |
| function calls. | |
| See :term:`Glossary <random_state>`. | |
| early_stopping : bool, default=False | |
| Whether to use early stopping to terminate training when validation | |
| score is not improving. If set to True, it will automatically set aside | |
| a stratified fraction of training data as validation and terminate | |
| training when validation score is not improving by at least `tol` for | |
| `n_iter_no_change` consecutive epochs. | |
| .. versionadded:: 0.20 | |
| validation_fraction : float, default=0.1 | |
| The proportion of training data to set aside as validation set for | |
| early stopping. Must be between 0 and 1. | |
| Only used if early_stopping is True. | |
| .. versionadded:: 0.20 | |
| n_iter_no_change : int, default=5 | |
| Number of iterations with no improvement to wait before early stopping. | |
| .. versionadded:: 0.20 | |
| class_weight : dict, {class_label: weight} or "balanced", default=None | |
| Preset for the class_weight fit parameter. | |
| Weights associated with classes. If not given, all classes | |
| are supposed to have weight one. | |
| The "balanced" mode uses the values of y to automatically adjust | |
| weights inversely proportional to class frequencies in the input data | |
| as ``n_samples / (n_classes * np.bincount(y))``. | |
| warm_start : bool, default=False | |
| When set to True, reuse the solution of the previous call to fit as | |
| initialization, otherwise, just erase the previous solution. See | |
| :term:`the Glossary <warm_start>`. | |
| Attributes | |
| ---------- | |
| classes_ : ndarray of shape (n_classes,) | |
| The unique classes labels. | |
| coef_ : ndarray of shape (1, n_features) if n_classes == 2 else \ | |
| (n_classes, n_features) | |
| Weights assigned to the features. | |
| intercept_ : ndarray of shape (1,) if n_classes == 2 else (n_classes,) | |
| Constants in decision function. | |
| loss_function_ : concrete LossFunction | |
| The function that determines the loss, or difference between the | |
| output of the algorithm and the target values. | |
| n_features_in_ : int | |
| Number of features seen during :term:`fit`. | |
| .. versionadded:: 0.24 | |
| feature_names_in_ : ndarray of shape (`n_features_in_`,) | |
| Names of features seen during :term:`fit`. Defined only when `X` | |
| has feature names that are all strings. | |
| .. versionadded:: 1.0 | |
| n_iter_ : int | |
| The actual number of iterations to reach the stopping criterion. | |
| For multiclass fits, it is the maximum over every binary fit. | |
| t_ : int | |
| Number of weight updates performed during training. | |
| Same as ``(n_iter_ * n_samples + 1)``. | |
| See Also | |
| -------- | |
| sklearn.linear_model.SGDClassifier : Linear classifiers | |
| (SVM, logistic regression, etc.) with SGD training. | |
| Notes | |
| ----- | |
| ``Perceptron`` is a classification algorithm which shares the same | |
| underlying implementation with ``SGDClassifier``. In fact, | |
| ``Perceptron()`` is equivalent to `SGDClassifier(loss="perceptron", | |
| eta0=1, learning_rate="constant", penalty=None)`. | |
| References | |
| ---------- | |
| https://en.wikipedia.org/wiki/Perceptron and references therein. | |
| Examples | |
| -------- | |
| >>> from sklearn.datasets import load_digits | |
| >>> from sklearn.linear_model import Perceptron | |
| >>> X, y = load_digits(return_X_y=True) | |
| >>> clf = Perceptron(tol=1e-3, random_state=0) | |
| >>> clf.fit(X, y) | |
| Perceptron() | |
| >>> clf.score(X, y) | |
| 0.939... | |
| """ | |
| _parameter_constraints: dict = {**BaseSGDClassifier._parameter_constraints} | |
| _parameter_constraints.pop("loss") | |
| _parameter_constraints.pop("average") | |
| _parameter_constraints.update( | |
| { | |
| "penalty": [StrOptions({"l2", "l1", "elasticnet"}), None], | |
| "alpha": [Interval(Real, 0, None, closed="left")], | |
| "l1_ratio": [Interval(Real, 0, 1, closed="both")], | |
| "eta0": [Interval(Real, 0, None, closed="left")], | |
| } | |
| ) | |
| def __init__( | |
| self, | |
| *, | |
| penalty=None, | |
| alpha=0.0001, | |
| l1_ratio=0.15, | |
| fit_intercept=True, | |
| max_iter=1000, | |
| tol=1e-3, | |
| shuffle=True, | |
| verbose=0, | |
| eta0=1.0, | |
| n_jobs=None, | |
| random_state=0, | |
| early_stopping=False, | |
| validation_fraction=0.1, | |
| n_iter_no_change=5, | |
| class_weight=None, | |
| warm_start=False, | |
| ): | |
| super().__init__( | |
| loss="perceptron", | |
| penalty=penalty, | |
| alpha=alpha, | |
| l1_ratio=l1_ratio, | |
| fit_intercept=fit_intercept, | |
| max_iter=max_iter, | |
| tol=tol, | |
| shuffle=shuffle, | |
| verbose=verbose, | |
| random_state=random_state, | |
| learning_rate="constant", | |
| eta0=eta0, | |
| early_stopping=early_stopping, | |
| validation_fraction=validation_fraction, | |
| n_iter_no_change=n_iter_no_change, | |
| power_t=0.5, | |
| warm_start=warm_start, | |
| class_weight=class_weight, | |
| n_jobs=n_jobs, | |
| ) | |