peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/scipy
/optimize
/_linprog_ip.py
| """Interior-point method for linear programming | |
| The *interior-point* method uses the primal-dual path following algorithm | |
| outlined in [1]_. This algorithm supports sparse constraint matrices and | |
| is typically faster than the simplex methods, especially for large, sparse | |
| problems. Note, however, that the solution returned may be slightly less | |
| accurate than those of the simplex methods and will not, in general, | |
| correspond with a vertex of the polytope defined by the constraints. | |
| .. versionadded:: 1.0.0 | |
| References | |
| ---------- | |
| .. [1] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| # Author: Matt Haberland | |
| import numpy as np | |
| import scipy as sp | |
| import scipy.sparse as sps | |
| from warnings import warn | |
| from scipy.linalg import LinAlgError | |
| from ._optimize import OptimizeWarning, OptimizeResult, _check_unknown_options | |
| from ._linprog_util import _postsolve | |
| has_umfpack = True | |
| has_cholmod = True | |
| try: | |
| import sksparse # noqa: F401 | |
| from sksparse.cholmod import cholesky as cholmod # noqa: F401 | |
| from sksparse.cholmod import analyze as cholmod_analyze | |
| except ImportError: | |
| has_cholmod = False | |
| try: | |
| import scikits.umfpack # test whether to use factorized # noqa: F401 | |
| except ImportError: | |
| has_umfpack = False | |
| def _get_solver(M, sparse=False, lstsq=False, sym_pos=True, | |
| cholesky=True, permc_spec='MMD_AT_PLUS_A'): | |
| """ | |
| Given solver options, return a handle to the appropriate linear system | |
| solver. | |
| Parameters | |
| ---------- | |
| M : 2-D array | |
| As defined in [4] Equation 8.31 | |
| sparse : bool (default = False) | |
| True if the system to be solved is sparse. This is typically set | |
| True when the original ``A_ub`` and ``A_eq`` arrays are sparse. | |
| lstsq : bool (default = False) | |
| True if the system is ill-conditioned and/or (nearly) singular and | |
| thus a more robust least-squares solver is desired. This is sometimes | |
| needed as the solution is approached. | |
| sym_pos : bool (default = True) | |
| True if the system matrix is symmetric positive definite | |
| Sometimes this needs to be set false as the solution is approached, | |
| even when the system should be symmetric positive definite, due to | |
| numerical difficulties. | |
| cholesky : bool (default = True) | |
| True if the system is to be solved by Cholesky, rather than LU, | |
| decomposition. This is typically faster unless the problem is very | |
| small or prone to numerical difficulties. | |
| permc_spec : str (default = 'MMD_AT_PLUS_A') | |
| Sparsity preservation strategy used by SuperLU. Acceptable values are: | |
| - ``NATURAL``: natural ordering. | |
| - ``MMD_ATA``: minimum degree ordering on the structure of A^T A. | |
| - ``MMD_AT_PLUS_A``: minimum degree ordering on the structure of A^T+A. | |
| - ``COLAMD``: approximate minimum degree column ordering. | |
| See SuperLU documentation. | |
| Returns | |
| ------- | |
| solve : function | |
| Handle to the appropriate solver function | |
| """ | |
| try: | |
| if sparse: | |
| if lstsq: | |
| def solve(r, sym_pos=False): | |
| return sps.linalg.lsqr(M, r)[0] | |
| elif cholesky: | |
| try: | |
| # Will raise an exception in the first call, | |
| # or when the matrix changes due to a new problem | |
| _get_solver.cholmod_factor.cholesky_inplace(M) | |
| except Exception: | |
| _get_solver.cholmod_factor = cholmod_analyze(M) | |
| _get_solver.cholmod_factor.cholesky_inplace(M) | |
| solve = _get_solver.cholmod_factor | |
| else: | |
| if has_umfpack and sym_pos: | |
| solve = sps.linalg.factorized(M) | |
| else: # factorized doesn't pass permc_spec | |
| solve = sps.linalg.splu(M, permc_spec=permc_spec).solve | |
| else: | |
| if lstsq: # sometimes necessary as solution is approached | |
| def solve(r): | |
| return sp.linalg.lstsq(M, r)[0] | |
| elif cholesky: | |
| L = sp.linalg.cho_factor(M) | |
| def solve(r): | |
| return sp.linalg.cho_solve(L, r) | |
| else: | |
| # this seems to cache the matrix factorization, so solving | |
| # with multiple right hand sides is much faster | |
| def solve(r, sym_pos=sym_pos): | |
| if sym_pos: | |
| return sp.linalg.solve(M, r, assume_a="pos") | |
| else: | |
| return sp.linalg.solve(M, r) | |
| # There are many things that can go wrong here, and it's hard to say | |
| # what all of them are. It doesn't really matter: if the matrix can't be | |
| # factorized, return None. get_solver will be called again with different | |
| # inputs, and a new routine will try to factorize the matrix. | |
| except KeyboardInterrupt: | |
| raise | |
| except Exception: | |
| return None | |
| return solve | |
| def _get_delta(A, b, c, x, y, z, tau, kappa, gamma, eta, sparse=False, | |
| lstsq=False, sym_pos=True, cholesky=True, pc=True, ip=False, | |
| permc_spec='MMD_AT_PLUS_A'): | |
| """ | |
| Given standard form problem defined by ``A``, ``b``, and ``c``; | |
| current variable estimates ``x``, ``y``, ``z``, ``tau``, and ``kappa``; | |
| algorithmic parameters ``gamma and ``eta; | |
| and options ``sparse``, ``lstsq``, ``sym_pos``, ``cholesky``, ``pc`` | |
| (predictor-corrector), and ``ip`` (initial point improvement), | |
| get the search direction for increments to the variable estimates. | |
| Parameters | |
| ---------- | |
| As defined in [4], except: | |
| sparse : bool | |
| True if the system to be solved is sparse. This is typically set | |
| True when the original ``A_ub`` and ``A_eq`` arrays are sparse. | |
| lstsq : bool | |
| True if the system is ill-conditioned and/or (nearly) singular and | |
| thus a more robust least-squares solver is desired. This is sometimes | |
| needed as the solution is approached. | |
| sym_pos : bool | |
| True if the system matrix is symmetric positive definite | |
| Sometimes this needs to be set false as the solution is approached, | |
| even when the system should be symmetric positive definite, due to | |
| numerical difficulties. | |
| cholesky : bool | |
| True if the system is to be solved by Cholesky, rather than LU, | |
| decomposition. This is typically faster unless the problem is very | |
| small or prone to numerical difficulties. | |
| pc : bool | |
| True if the predictor-corrector method of Mehrota is to be used. This | |
| is almost always (if not always) beneficial. Even though it requires | |
| the solution of an additional linear system, the factorization | |
| is typically (implicitly) reused so solution is efficient, and the | |
| number of algorithm iterations is typically reduced. | |
| ip : bool | |
| True if the improved initial point suggestion due to [4] section 4.3 | |
| is desired. It's unclear whether this is beneficial. | |
| permc_spec : str (default = 'MMD_AT_PLUS_A') | |
| (Has effect only with ``sparse = True``, ``lstsq = False``, ``sym_pos = | |
| True``.) A matrix is factorized in each iteration of the algorithm. | |
| This option specifies how to permute the columns of the matrix for | |
| sparsity preservation. Acceptable values are: | |
| - ``NATURAL``: natural ordering. | |
| - ``MMD_ATA``: minimum degree ordering on the structure of A^T A. | |
| - ``MMD_AT_PLUS_A``: minimum degree ordering on the structure of A^T+A. | |
| - ``COLAMD``: approximate minimum degree column ordering. | |
| This option can impact the convergence of the | |
| interior point algorithm; test different values to determine which | |
| performs best for your problem. For more information, refer to | |
| ``scipy.sparse.linalg.splu``. | |
| Returns | |
| ------- | |
| Search directions as defined in [4] | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| if A.shape[0] == 0: | |
| # If there are no constraints, some solvers fail (understandably) | |
| # rather than returning empty solution. This gets the job done. | |
| sparse, lstsq, sym_pos, cholesky = False, False, True, False | |
| n_x = len(x) | |
| # [4] Equation 8.8 | |
| r_P = b * tau - A.dot(x) | |
| r_D = c * tau - A.T.dot(y) - z | |
| r_G = c.dot(x) - b.transpose().dot(y) + kappa | |
| mu = (x.dot(z) + tau * kappa) / (n_x + 1) | |
| # Assemble M from [4] Equation 8.31 | |
| Dinv = x / z | |
| if sparse: | |
| M = A.dot(sps.diags(Dinv, 0, format="csc").dot(A.T)) | |
| else: | |
| M = A.dot(Dinv.reshape(-1, 1) * A.T) | |
| solve = _get_solver(M, sparse, lstsq, sym_pos, cholesky, permc_spec) | |
| # pc: "predictor-corrector" [4] Section 4.1 | |
| # In development this option could be turned off | |
| # but it always seems to improve performance substantially | |
| n_corrections = 1 if pc else 0 | |
| i = 0 | |
| alpha, d_x, d_z, d_tau, d_kappa = 0, 0, 0, 0, 0 | |
| while i <= n_corrections: | |
| # Reference [4] Eq. 8.6 | |
| rhatp = eta(gamma) * r_P | |
| rhatd = eta(gamma) * r_D | |
| rhatg = eta(gamma) * r_G | |
| # Reference [4] Eq. 8.7 | |
| rhatxs = gamma * mu - x * z | |
| rhattk = gamma * mu - tau * kappa | |
| if i == 1: | |
| if ip: # if the correction is to get "initial point" | |
| # Reference [4] Eq. 8.23 | |
| rhatxs = ((1 - alpha) * gamma * mu - | |
| x * z - alpha**2 * d_x * d_z) | |
| rhattk = ((1 - alpha) * gamma * mu - | |
| tau * kappa - | |
| alpha**2 * d_tau * d_kappa) | |
| else: # if the correction is for "predictor-corrector" | |
| # Reference [4] Eq. 8.13 | |
| rhatxs -= d_x * d_z | |
| rhattk -= d_tau * d_kappa | |
| # sometimes numerical difficulties arise as the solution is approached | |
| # this loop tries to solve the equations using a sequence of functions | |
| # for solve. For dense systems, the order is: | |
| # 1. scipy.linalg.cho_factor/scipy.linalg.cho_solve, | |
| # 2. scipy.linalg.solve w/ sym_pos = True, | |
| # 3. scipy.linalg.solve w/ sym_pos = False, and if all else fails | |
| # 4. scipy.linalg.lstsq | |
| # For sparse systems, the order is: | |
| # 1. sksparse.cholmod.cholesky (if available) | |
| # 2. scipy.sparse.linalg.factorized (if umfpack available) | |
| # 3. scipy.sparse.linalg.splu | |
| # 4. scipy.sparse.linalg.lsqr | |
| solved = False | |
| while not solved: | |
| try: | |
| # [4] Equation 8.28 | |
| p, q = _sym_solve(Dinv, A, c, b, solve) | |
| # [4] Equation 8.29 | |
| u, v = _sym_solve(Dinv, A, rhatd - | |
| (1 / x) * rhatxs, rhatp, solve) | |
| if np.any(np.isnan(p)) or np.any(np.isnan(q)): | |
| raise LinAlgError | |
| solved = True | |
| except (LinAlgError, ValueError, TypeError) as e: | |
| # Usually this doesn't happen. If it does, it happens when | |
| # there are redundant constraints or when approaching the | |
| # solution. If so, change solver. | |
| if cholesky: | |
| cholesky = False | |
| warn( | |
| "Solving system with option 'cholesky':True " | |
| "failed. It is normal for this to happen " | |
| "occasionally, especially as the solution is " | |
| "approached. However, if you see this frequently, " | |
| "consider setting option 'cholesky' to False.", | |
| OptimizeWarning, stacklevel=5) | |
| elif sym_pos: | |
| sym_pos = False | |
| warn( | |
| "Solving system with option 'sym_pos':True " | |
| "failed. It is normal for this to happen " | |
| "occasionally, especially as the solution is " | |
| "approached. However, if you see this frequently, " | |
| "consider setting option 'sym_pos' to False.", | |
| OptimizeWarning, stacklevel=5) | |
| elif not lstsq: | |
| lstsq = True | |
| warn( | |
| "Solving system with option 'sym_pos':False " | |
| "failed. This may happen occasionally, " | |
| "especially as the solution is " | |
| "approached. However, if you see this frequently, " | |
| "your problem may be numerically challenging. " | |
| "If you cannot improve the formulation, consider " | |
| "setting 'lstsq' to True. Consider also setting " | |
| "`presolve` to True, if it is not already.", | |
| OptimizeWarning, stacklevel=5) | |
| else: | |
| raise e | |
| solve = _get_solver(M, sparse, lstsq, sym_pos, | |
| cholesky, permc_spec) | |
| # [4] Results after 8.29 | |
| d_tau = ((rhatg + 1 / tau * rhattk - (-c.dot(u) + b.dot(v))) / | |
| (1 / tau * kappa + (-c.dot(p) + b.dot(q)))) | |
| d_x = u + p * d_tau | |
| d_y = v + q * d_tau | |
| # [4] Relations between after 8.25 and 8.26 | |
| d_z = (1 / x) * (rhatxs - z * d_x) | |
| d_kappa = 1 / tau * (rhattk - kappa * d_tau) | |
| # [4] 8.12 and "Let alpha be the maximal possible step..." before 8.23 | |
| alpha = _get_step(x, d_x, z, d_z, tau, d_tau, kappa, d_kappa, 1) | |
| if ip: # initial point - see [4] 4.4 | |
| gamma = 10 | |
| else: # predictor-corrector, [4] definition after 8.12 | |
| beta1 = 0.1 # [4] pg. 220 (Table 8.1) | |
| gamma = (1 - alpha)**2 * min(beta1, (1 - alpha)) | |
| i += 1 | |
| return d_x, d_y, d_z, d_tau, d_kappa | |
| def _sym_solve(Dinv, A, r1, r2, solve): | |
| """ | |
| An implementation of [4] equation 8.31 and 8.32 | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| # [4] 8.31 | |
| r = r2 + A.dot(Dinv * r1) | |
| v = solve(r) | |
| # [4] 8.32 | |
| u = Dinv * (A.T.dot(v) - r1) | |
| return u, v | |
| def _get_step(x, d_x, z, d_z, tau, d_tau, kappa, d_kappa, alpha0): | |
| """ | |
| An implementation of [4] equation 8.21 | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| # [4] 4.3 Equation 8.21, ignoring 8.20 requirement | |
| # same step is taken in primal and dual spaces | |
| # alpha0 is basically beta3 from [4] Table 8.1, but instead of beta3 | |
| # the value 1 is used in Mehrota corrector and initial point correction | |
| i_x = d_x < 0 | |
| i_z = d_z < 0 | |
| alpha_x = alpha0 * np.min(x[i_x] / -d_x[i_x]) if np.any(i_x) else 1 | |
| alpha_tau = alpha0 * tau / -d_tau if d_tau < 0 else 1 | |
| alpha_z = alpha0 * np.min(z[i_z] / -d_z[i_z]) if np.any(i_z) else 1 | |
| alpha_kappa = alpha0 * kappa / -d_kappa if d_kappa < 0 else 1 | |
| alpha = np.min([1, alpha_x, alpha_tau, alpha_z, alpha_kappa]) | |
| return alpha | |
| def _get_message(status): | |
| """ | |
| Given problem status code, return a more detailed message. | |
| Parameters | |
| ---------- | |
| status : int | |
| An integer representing the exit status of the optimization:: | |
| 0 : Optimization terminated successfully | |
| 1 : Iteration limit reached | |
| 2 : Problem appears to be infeasible | |
| 3 : Problem appears to be unbounded | |
| 4 : Serious numerical difficulties encountered | |
| Returns | |
| ------- | |
| message : str | |
| A string descriptor of the exit status of the optimization. | |
| """ | |
| messages = ( | |
| ["Optimization terminated successfully.", | |
| "The iteration limit was reached before the algorithm converged.", | |
| "The algorithm terminated successfully and determined that the " | |
| "problem is infeasible.", | |
| "The algorithm terminated successfully and determined that the " | |
| "problem is unbounded.", | |
| "Numerical difficulties were encountered before the problem " | |
| "converged. Please check your problem formulation for errors, " | |
| "independence of linear equality constraints, and reasonable " | |
| "scaling and matrix condition numbers. If you continue to " | |
| "encounter this error, please submit a bug report." | |
| ]) | |
| return messages[status] | |
| def _do_step(x, y, z, tau, kappa, d_x, d_y, d_z, d_tau, d_kappa, alpha): | |
| """ | |
| An implementation of [4] Equation 8.9 | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| x = x + alpha * d_x | |
| tau = tau + alpha * d_tau | |
| z = z + alpha * d_z | |
| kappa = kappa + alpha * d_kappa | |
| y = y + alpha * d_y | |
| return x, y, z, tau, kappa | |
| def _get_blind_start(shape): | |
| """ | |
| Return the starting point from [4] 4.4 | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| m, n = shape | |
| x0 = np.ones(n) | |
| y0 = np.zeros(m) | |
| z0 = np.ones(n) | |
| tau0 = 1 | |
| kappa0 = 1 | |
| return x0, y0, z0, tau0, kappa0 | |
| def _indicators(A, b, c, c0, x, y, z, tau, kappa): | |
| """ | |
| Implementation of several equations from [4] used as indicators of | |
| the status of optimization. | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| # residuals for termination are relative to initial values | |
| x0, y0, z0, tau0, kappa0 = _get_blind_start(A.shape) | |
| # See [4], Section 4 - The Homogeneous Algorithm, Equation 8.8 | |
| def r_p(x, tau): | |
| return b * tau - A.dot(x) | |
| def r_d(y, z, tau): | |
| return c * tau - A.T.dot(y) - z | |
| def r_g(x, y, kappa): | |
| return kappa + c.dot(x) - b.dot(y) | |
| # np.dot unpacks if they are arrays of size one | |
| def mu(x, tau, z, kappa): | |
| return (x.dot(z) + np.dot(tau, kappa)) / (len(x) + 1) | |
| obj = c.dot(x / tau) + c0 | |
| def norm(a): | |
| return np.linalg.norm(a) | |
| # See [4], Section 4.5 - The Stopping Criteria | |
| r_p0 = r_p(x0, tau0) | |
| r_d0 = r_d(y0, z0, tau0) | |
| r_g0 = r_g(x0, y0, kappa0) | |
| mu_0 = mu(x0, tau0, z0, kappa0) | |
| rho_A = norm(c.T.dot(x) - b.T.dot(y)) / (tau + norm(b.T.dot(y))) | |
| rho_p = norm(r_p(x, tau)) / max(1, norm(r_p0)) | |
| rho_d = norm(r_d(y, z, tau)) / max(1, norm(r_d0)) | |
| rho_g = norm(r_g(x, y, kappa)) / max(1, norm(r_g0)) | |
| rho_mu = mu(x, tau, z, kappa) / mu_0 | |
| return rho_p, rho_d, rho_A, rho_g, rho_mu, obj | |
| def _display_iter(rho_p, rho_d, rho_g, alpha, rho_mu, obj, header=False): | |
| """ | |
| Print indicators of optimization status to the console. | |
| Parameters | |
| ---------- | |
| rho_p : float | |
| The (normalized) primal feasibility, see [4] 4.5 | |
| rho_d : float | |
| The (normalized) dual feasibility, see [4] 4.5 | |
| rho_g : float | |
| The (normalized) duality gap, see [4] 4.5 | |
| alpha : float | |
| The step size, see [4] 4.3 | |
| rho_mu : float | |
| The (normalized) path parameter, see [4] 4.5 | |
| obj : float | |
| The objective function value of the current iterate | |
| header : bool | |
| True if a header is to be printed | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| """ | |
| if header: | |
| print("Primal Feasibility ", | |
| "Dual Feasibility ", | |
| "Duality Gap ", | |
| "Step ", | |
| "Path Parameter ", | |
| "Objective ") | |
| # no clue why this works | |
| fmt = '{0:<20.13}{1:<20.13}{2:<20.13}{3:<17.13}{4:<20.13}{5:<20.13}' | |
| print(fmt.format( | |
| float(rho_p), | |
| float(rho_d), | |
| float(rho_g), | |
| alpha if isinstance(alpha, str) else float(alpha), | |
| float(rho_mu), | |
| float(obj))) | |
| def _ip_hsd(A, b, c, c0, alpha0, beta, maxiter, disp, tol, sparse, lstsq, | |
| sym_pos, cholesky, pc, ip, permc_spec, callback, postsolve_args): | |
| r""" | |
| Solve a linear programming problem in standard form: | |
| Minimize:: | |
| c @ x | |
| Subject to:: | |
| A @ x == b | |
| x >= 0 | |
| using the interior point method of [4]. | |
| Parameters | |
| ---------- | |
| A : 2-D array | |
| 2-D array such that ``A @ x``, gives the values of the equality | |
| constraints at ``x``. | |
| b : 1-D array | |
| 1-D array of values representing the RHS of each equality constraint | |
| (row) in ``A`` (for standard form problem). | |
| c : 1-D array | |
| Coefficients of the linear objective function to be minimized (for | |
| standard form problem). | |
| c0 : float | |
| Constant term in objective function due to fixed (and eliminated) | |
| variables. (Purely for display.) | |
| alpha0 : float | |
| The maximal step size for Mehrota's predictor-corrector search | |
| direction; see :math:`\beta_3`of [4] Table 8.1 | |
| beta : float | |
| The desired reduction of the path parameter :math:`\mu` (see [6]_) | |
| maxiter : int | |
| The maximum number of iterations of the algorithm. | |
| disp : bool | |
| Set to ``True`` if indicators of optimization status are to be printed | |
| to the console each iteration. | |
| tol : float | |
| Termination tolerance; see [4]_ Section 4.5. | |
| sparse : bool | |
| Set to ``True`` if the problem is to be treated as sparse. However, | |
| the inputs ``A_eq`` and ``A_ub`` should nonetheless be provided as | |
| (dense) arrays rather than sparse matrices. | |
| lstsq : bool | |
| Set to ``True`` if the problem is expected to be very poorly | |
| conditioned. This should always be left as ``False`` unless severe | |
| numerical difficulties are frequently encountered, and a better option | |
| would be to improve the formulation of the problem. | |
| sym_pos : bool | |
| Leave ``True`` if the problem is expected to yield a well conditioned | |
| symmetric positive definite normal equation matrix (almost always). | |
| cholesky : bool | |
| Set to ``True`` if the normal equations are to be solved by explicit | |
| Cholesky decomposition followed by explicit forward/backward | |
| substitution. This is typically faster for moderate, dense problems | |
| that are numerically well-behaved. | |
| pc : bool | |
| Leave ``True`` if the predictor-corrector method of Mehrota is to be | |
| used. This is almost always (if not always) beneficial. | |
| ip : bool | |
| Set to ``True`` if the improved initial point suggestion due to [4]_ | |
| Section 4.3 is desired. It's unclear whether this is beneficial. | |
| permc_spec : str (default = 'MMD_AT_PLUS_A') | |
| (Has effect only with ``sparse = True``, ``lstsq = False``, ``sym_pos = | |
| True``.) A matrix is factorized in each iteration of the algorithm. | |
| This option specifies how to permute the columns of the matrix for | |
| sparsity preservation. Acceptable values are: | |
| - ``NATURAL``: natural ordering. | |
| - ``MMD_ATA``: minimum degree ordering on the structure of A^T A. | |
| - ``MMD_AT_PLUS_A``: minimum degree ordering on the structure of A^T+A. | |
| - ``COLAMD``: approximate minimum degree column ordering. | |
| This option can impact the convergence of the | |
| interior point algorithm; test different values to determine which | |
| performs best for your problem. For more information, refer to | |
| ``scipy.sparse.linalg.splu``. | |
| callback : callable, optional | |
| If a callback function is provided, it will be called within each | |
| iteration of the algorithm. The callback function must accept a single | |
| `scipy.optimize.OptimizeResult` consisting of the following fields: | |
| x : 1-D array | |
| Current solution vector | |
| fun : float | |
| Current value of the objective function | |
| success : bool | |
| True only when an algorithm has completed successfully, | |
| so this is always False as the callback function is called | |
| only while the algorithm is still iterating. | |
| slack : 1-D array | |
| The values of the slack variables. Each slack variable | |
| corresponds to an inequality constraint. If the slack is zero, | |
| the corresponding constraint is active. | |
| con : 1-D array | |
| The (nominally zero) residuals of the equality constraints, | |
| that is, ``b - A_eq @ x`` | |
| phase : int | |
| The phase of the algorithm being executed. This is always | |
| 1 for the interior-point method because it has only one phase. | |
| status : int | |
| For revised simplex, this is always 0 because if a different | |
| status is detected, the algorithm terminates. | |
| nit : int | |
| The number of iterations performed. | |
| message : str | |
| A string descriptor of the exit status of the optimization. | |
| postsolve_args : tuple | |
| Data needed by _postsolve to convert the solution to the standard-form | |
| problem into the solution to the original problem. | |
| Returns | |
| ------- | |
| x_hat : float | |
| Solution vector (for standard form problem). | |
| status : int | |
| An integer representing the exit status of the optimization:: | |
| 0 : Optimization terminated successfully | |
| 1 : Iteration limit reached | |
| 2 : Problem appears to be infeasible | |
| 3 : Problem appears to be unbounded | |
| 4 : Serious numerical difficulties encountered | |
| message : str | |
| A string descriptor of the exit status of the optimization. | |
| iteration : int | |
| The number of iterations taken to solve the problem | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| .. [6] Freund, Robert M. "Primal-Dual Interior-Point Methods for Linear | |
| Programming based on Newton's Method." Unpublished Course Notes, | |
| March 2004. Available 2/25/2017 at: | |
| https://ocw.mit.edu/courses/sloan-school-of-management/15-084j-nonlinear-programming-spring-2004/lecture-notes/lec14_int_pt_mthd.pdf | |
| """ | |
| iteration = 0 | |
| # default initial point | |
| x, y, z, tau, kappa = _get_blind_start(A.shape) | |
| # first iteration is special improvement of initial point | |
| ip = ip if pc else False | |
| # [4] 4.5 | |
| rho_p, rho_d, rho_A, rho_g, rho_mu, obj = _indicators( | |
| A, b, c, c0, x, y, z, tau, kappa) | |
| go = rho_p > tol or rho_d > tol or rho_A > tol # we might get lucky : ) | |
| if disp: | |
| _display_iter(rho_p, rho_d, rho_g, "-", rho_mu, obj, header=True) | |
| if callback is not None: | |
| x_o, fun, slack, con = _postsolve(x/tau, postsolve_args) | |
| res = OptimizeResult({'x': x_o, 'fun': fun, 'slack': slack, | |
| 'con': con, 'nit': iteration, 'phase': 1, | |
| 'complete': False, 'status': 0, | |
| 'message': "", 'success': False}) | |
| callback(res) | |
| status = 0 | |
| message = "Optimization terminated successfully." | |
| if sparse: | |
| A = sps.csc_matrix(A) | |
| while go: | |
| iteration += 1 | |
| if ip: # initial point | |
| # [4] Section 4.4 | |
| gamma = 1 | |
| def eta(g): | |
| return 1 | |
| else: | |
| # gamma = 0 in predictor step according to [4] 4.1 | |
| # if predictor/corrector is off, use mean of complementarity [6] | |
| # 5.1 / [4] Below Figure 10-4 | |
| gamma = 0 if pc else beta * np.mean(z * x) | |
| # [4] Section 4.1 | |
| def eta(g=gamma): | |
| return 1 - g | |
| try: | |
| # Solve [4] 8.6 and 8.7/8.13/8.23 | |
| d_x, d_y, d_z, d_tau, d_kappa = _get_delta( | |
| A, b, c, x, y, z, tau, kappa, gamma, eta, | |
| sparse, lstsq, sym_pos, cholesky, pc, ip, permc_spec) | |
| if ip: # initial point | |
| # [4] 4.4 | |
| # Formula after 8.23 takes a full step regardless if this will | |
| # take it negative | |
| alpha = 1.0 | |
| x, y, z, tau, kappa = _do_step( | |
| x, y, z, tau, kappa, d_x, d_y, | |
| d_z, d_tau, d_kappa, alpha) | |
| x[x < 1] = 1 | |
| z[z < 1] = 1 | |
| tau = max(1, tau) | |
| kappa = max(1, kappa) | |
| ip = False # done with initial point | |
| else: | |
| # [4] Section 4.3 | |
| alpha = _get_step(x, d_x, z, d_z, tau, | |
| d_tau, kappa, d_kappa, alpha0) | |
| # [4] Equation 8.9 | |
| x, y, z, tau, kappa = _do_step( | |
| x, y, z, tau, kappa, d_x, d_y, d_z, d_tau, d_kappa, alpha) | |
| except (LinAlgError, FloatingPointError, | |
| ValueError, ZeroDivisionError): | |
| # this can happen when sparse solver is used and presolve | |
| # is turned off. Also observed ValueError in AppVeyor Python 3.6 | |
| # Win32 build (PR #8676). I've never seen it otherwise. | |
| status = 4 | |
| message = _get_message(status) | |
| break | |
| # [4] 4.5 | |
| rho_p, rho_d, rho_A, rho_g, rho_mu, obj = _indicators( | |
| A, b, c, c0, x, y, z, tau, kappa) | |
| go = rho_p > tol or rho_d > tol or rho_A > tol | |
| if disp: | |
| _display_iter(rho_p, rho_d, rho_g, alpha, rho_mu, obj) | |
| if callback is not None: | |
| x_o, fun, slack, con = _postsolve(x/tau, postsolve_args) | |
| res = OptimizeResult({'x': x_o, 'fun': fun, 'slack': slack, | |
| 'con': con, 'nit': iteration, 'phase': 1, | |
| 'complete': False, 'status': 0, | |
| 'message': "", 'success': False}) | |
| callback(res) | |
| # [4] 4.5 | |
| inf1 = (rho_p < tol and rho_d < tol and rho_g < tol and tau < tol * | |
| max(1, kappa)) | |
| inf2 = rho_mu < tol and tau < tol * min(1, kappa) | |
| if inf1 or inf2: | |
| # [4] Lemma 8.4 / Theorem 8.3 | |
| if b.transpose().dot(y) > tol: | |
| status = 2 | |
| else: # elif c.T.dot(x) < tol: ? Probably not necessary. | |
| status = 3 | |
| message = _get_message(status) | |
| break | |
| elif iteration >= maxiter: | |
| status = 1 | |
| message = _get_message(status) | |
| break | |
| x_hat = x / tau | |
| # [4] Statement after Theorem 8.2 | |
| return x_hat, status, message, iteration | |
| def _linprog_ip(c, c0, A, b, callback, postsolve_args, maxiter=1000, tol=1e-8, | |
| disp=False, alpha0=.99995, beta=0.1, sparse=False, lstsq=False, | |
| sym_pos=True, cholesky=None, pc=True, ip=False, | |
| permc_spec='MMD_AT_PLUS_A', **unknown_options): | |
| r""" | |
| Minimize a linear objective function subject to linear | |
| equality and non-negativity constraints using the interior point method | |
| of [4]_. Linear programming is intended to solve problems | |
| of the following form: | |
| Minimize:: | |
| c @ x | |
| Subject to:: | |
| A @ x == b | |
| x >= 0 | |
| User-facing documentation is in _linprog_doc.py. | |
| Parameters | |
| ---------- | |
| c : 1-D array | |
| Coefficients of the linear objective function to be minimized. | |
| c0 : float | |
| Constant term in objective function due to fixed (and eliminated) | |
| variables. (Purely for display.) | |
| A : 2-D array | |
| 2-D array such that ``A @ x``, gives the values of the equality | |
| constraints at ``x``. | |
| b : 1-D array | |
| 1-D array of values representing the right hand side of each equality | |
| constraint (row) in ``A``. | |
| callback : callable, optional | |
| Callback function to be executed once per iteration. | |
| postsolve_args : tuple | |
| Data needed by _postsolve to convert the solution to the standard-form | |
| problem into the solution to the original problem. | |
| Options | |
| ------- | |
| maxiter : int (default = 1000) | |
| The maximum number of iterations of the algorithm. | |
| tol : float (default = 1e-8) | |
| Termination tolerance to be used for all termination criteria; | |
| see [4]_ Section 4.5. | |
| disp : bool (default = False) | |
| Set to ``True`` if indicators of optimization status are to be printed | |
| to the console each iteration. | |
| alpha0 : float (default = 0.99995) | |
| The maximal step size for Mehrota's predictor-corrector search | |
| direction; see :math:`\beta_{3}` of [4]_ Table 8.1. | |
| beta : float (default = 0.1) | |
| The desired reduction of the path parameter :math:`\mu` (see [6]_) | |
| when Mehrota's predictor-corrector is not in use (uncommon). | |
| sparse : bool (default = False) | |
| Set to ``True`` if the problem is to be treated as sparse after | |
| presolve. If either ``A_eq`` or ``A_ub`` is a sparse matrix, | |
| this option will automatically be set ``True``, and the problem | |
| will be treated as sparse even during presolve. If your constraint | |
| matrices contain mostly zeros and the problem is not very small (less | |
| than about 100 constraints or variables), consider setting ``True`` | |
| or providing ``A_eq`` and ``A_ub`` as sparse matrices. | |
| lstsq : bool (default = False) | |
| Set to ``True`` if the problem is expected to be very poorly | |
| conditioned. This should always be left ``False`` unless severe | |
| numerical difficulties are encountered. Leave this at the default | |
| unless you receive a warning message suggesting otherwise. | |
| sym_pos : bool (default = True) | |
| Leave ``True`` if the problem is expected to yield a well conditioned | |
| symmetric positive definite normal equation matrix | |
| (almost always). Leave this at the default unless you receive | |
| a warning message suggesting otherwise. | |
| cholesky : bool (default = True) | |
| Set to ``True`` if the normal equations are to be solved by explicit | |
| Cholesky decomposition followed by explicit forward/backward | |
| substitution. This is typically faster for problems | |
| that are numerically well-behaved. | |
| pc : bool (default = True) | |
| Leave ``True`` if the predictor-corrector method of Mehrota is to be | |
| used. This is almost always (if not always) beneficial. | |
| ip : bool (default = False) | |
| Set to ``True`` if the improved initial point suggestion due to [4]_ | |
| Section 4.3 is desired. Whether this is beneficial or not | |
| depends on the problem. | |
| permc_spec : str (default = 'MMD_AT_PLUS_A') | |
| (Has effect only with ``sparse = True``, ``lstsq = False``, ``sym_pos = | |
| True``, and no SuiteSparse.) | |
| A matrix is factorized in each iteration of the algorithm. | |
| This option specifies how to permute the columns of the matrix for | |
| sparsity preservation. Acceptable values are: | |
| - ``NATURAL``: natural ordering. | |
| - ``MMD_ATA``: minimum degree ordering on the structure of A^T A. | |
| - ``MMD_AT_PLUS_A``: minimum degree ordering on the structure of A^T+A. | |
| - ``COLAMD``: approximate minimum degree column ordering. | |
| This option can impact the convergence of the | |
| interior point algorithm; test different values to determine which | |
| performs best for your problem. For more information, refer to | |
| ``scipy.sparse.linalg.splu``. | |
| unknown_options : dict | |
| Optional arguments not used by this particular solver. If | |
| `unknown_options` is non-empty a warning is issued listing all | |
| unused options. | |
| Returns | |
| ------- | |
| x : 1-D array | |
| Solution vector. | |
| status : int | |
| An integer representing the exit status of the optimization:: | |
| 0 : Optimization terminated successfully | |
| 1 : Iteration limit reached | |
| 2 : Problem appears to be infeasible | |
| 3 : Problem appears to be unbounded | |
| 4 : Serious numerical difficulties encountered | |
| message : str | |
| A string descriptor of the exit status of the optimization. | |
| iteration : int | |
| The number of iterations taken to solve the problem. | |
| Notes | |
| ----- | |
| This method implements the algorithm outlined in [4]_ with ideas from [8]_ | |
| and a structure inspired by the simpler methods of [6]_. | |
| The primal-dual path following method begins with initial 'guesses' of | |
| the primal and dual variables of the standard form problem and iteratively | |
| attempts to solve the (nonlinear) Karush-Kuhn-Tucker conditions for the | |
| problem with a gradually reduced logarithmic barrier term added to the | |
| objective. This particular implementation uses a homogeneous self-dual | |
| formulation, which provides certificates of infeasibility or unboundedness | |
| where applicable. | |
| The default initial point for the primal and dual variables is that | |
| defined in [4]_ Section 4.4 Equation 8.22. Optionally (by setting initial | |
| point option ``ip=True``), an alternate (potentially improved) starting | |
| point can be calculated according to the additional recommendations of | |
| [4]_ Section 4.4. | |
| A search direction is calculated using the predictor-corrector method | |
| (single correction) proposed by Mehrota and detailed in [4]_ Section 4.1. | |
| (A potential improvement would be to implement the method of multiple | |
| corrections described in [4]_ Section 4.2.) In practice, this is | |
| accomplished by solving the normal equations, [4]_ Section 5.1 Equations | |
| 8.31 and 8.32, derived from the Newton equations [4]_ Section 5 Equations | |
| 8.25 (compare to [4]_ Section 4 Equations 8.6-8.8). The advantage of | |
| solving the normal equations rather than 8.25 directly is that the | |
| matrices involved are symmetric positive definite, so Cholesky | |
| decomposition can be used rather than the more expensive LU factorization. | |
| With default options, the solver used to perform the factorization depends | |
| on third-party software availability and the conditioning of the problem. | |
| For dense problems, solvers are tried in the following order: | |
| 1. ``scipy.linalg.cho_factor`` | |
| 2. ``scipy.linalg.solve`` with option ``sym_pos=True`` | |
| 3. ``scipy.linalg.solve`` with option ``sym_pos=False`` | |
| 4. ``scipy.linalg.lstsq`` | |
| For sparse problems: | |
| 1. ``sksparse.cholmod.cholesky`` (if scikit-sparse and SuiteSparse are installed) | |
| 2. ``scipy.sparse.linalg.factorized`` | |
| (if scikit-umfpack and SuiteSparse are installed) | |
| 3. ``scipy.sparse.linalg.splu`` (which uses SuperLU distributed with SciPy) | |
| 4. ``scipy.sparse.linalg.lsqr`` | |
| If the solver fails for any reason, successively more robust (but slower) | |
| solvers are attempted in the order indicated. Attempting, failing, and | |
| re-starting factorization can be time consuming, so if the problem is | |
| numerically challenging, options can be set to bypass solvers that are | |
| failing. Setting ``cholesky=False`` skips to solver 2, | |
| ``sym_pos=False`` skips to solver 3, and ``lstsq=True`` skips | |
| to solver 4 for both sparse and dense problems. | |
| Potential improvements for combatting issues associated with dense | |
| columns in otherwise sparse problems are outlined in [4]_ Section 5.3 and | |
| [10]_ Section 4.1-4.2; the latter also discusses the alleviation of | |
| accuracy issues associated with the substitution approach to free | |
| variables. | |
| After calculating the search direction, the maximum possible step size | |
| that does not activate the non-negativity constraints is calculated, and | |
| the smaller of this step size and unity is applied (as in [4]_ Section | |
| 4.1.) [4]_ Section 4.3 suggests improvements for choosing the step size. | |
| The new point is tested according to the termination conditions of [4]_ | |
| Section 4.5. The same tolerance, which can be set using the ``tol`` option, | |
| is used for all checks. (A potential improvement would be to expose | |
| the different tolerances to be set independently.) If optimality, | |
| unboundedness, or infeasibility is detected, the solve procedure | |
| terminates; otherwise it repeats. | |
| The expected problem formulation differs between the top level ``linprog`` | |
| module and the method specific solvers. The method specific solvers expect a | |
| problem in standard form: | |
| Minimize:: | |
| c @ x | |
| Subject to:: | |
| A @ x == b | |
| x >= 0 | |
| Whereas the top level ``linprog`` module expects a problem of form: | |
| Minimize:: | |
| c @ x | |
| Subject to:: | |
| A_ub @ x <= b_ub | |
| A_eq @ x == b_eq | |
| lb <= x <= ub | |
| where ``lb = 0`` and ``ub = None`` unless set in ``bounds``. | |
| The original problem contains equality, upper-bound and variable constraints | |
| whereas the method specific solver requires equality constraints and | |
| variable non-negativity. | |
| ``linprog`` module converts the original problem to standard form by | |
| converting the simple bounds to upper bound constraints, introducing | |
| non-negative slack variables for inequality constraints, and expressing | |
| unbounded variables as the difference between two non-negative variables. | |
| References | |
| ---------- | |
| .. [4] Andersen, Erling D., and Knud D. Andersen. "The MOSEK interior point | |
| optimizer for linear programming: an implementation of the | |
| homogeneous algorithm." High performance optimization. Springer US, | |
| 2000. 197-232. | |
| .. [6] Freund, Robert M. "Primal-Dual Interior-Point Methods for Linear | |
| Programming based on Newton's Method." Unpublished Course Notes, | |
| March 2004. Available 2/25/2017 at | |
| https://ocw.mit.edu/courses/sloan-school-of-management/15-084j-nonlinear-programming-spring-2004/lecture-notes/lec14_int_pt_mthd.pdf | |
| .. [8] Andersen, Erling D., and Knud D. Andersen. "Presolving in linear | |
| programming." Mathematical Programming 71.2 (1995): 221-245. | |
| .. [9] Bertsimas, Dimitris, and J. Tsitsiklis. "Introduction to linear | |
| programming." Athena Scientific 1 (1997): 997. | |
| .. [10] Andersen, Erling D., et al. Implementation of interior point methods | |
| for large scale linear programming. HEC/Universite de Geneve, 1996. | |
| """ | |
| _check_unknown_options(unknown_options) | |
| # These should be warnings, not errors | |
| if (cholesky or cholesky is None) and sparse and not has_cholmod: | |
| if cholesky: | |
| warn("Sparse cholesky is only available with scikit-sparse. " | |
| "Setting `cholesky = False`", | |
| OptimizeWarning, stacklevel=3) | |
| cholesky = False | |
| if sparse and lstsq: | |
| warn("Option combination 'sparse':True and 'lstsq':True " | |
| "is not recommended.", | |
| OptimizeWarning, stacklevel=3) | |
| if lstsq and cholesky: | |
| warn("Invalid option combination 'lstsq':True " | |
| "and 'cholesky':True; option 'cholesky' has no effect when " | |
| "'lstsq' is set True.", | |
| OptimizeWarning, stacklevel=3) | |
| valid_permc_spec = ('NATURAL', 'MMD_ATA', 'MMD_AT_PLUS_A', 'COLAMD') | |
| if permc_spec.upper() not in valid_permc_spec: | |
| warn("Invalid permc_spec option: '" + str(permc_spec) + "'. " | |
| "Acceptable values are 'NATURAL', 'MMD_ATA', 'MMD_AT_PLUS_A', " | |
| "and 'COLAMD'. Reverting to default.", | |
| OptimizeWarning, stacklevel=3) | |
| permc_spec = 'MMD_AT_PLUS_A' | |
| # This can be an error | |
| if not sym_pos and cholesky: | |
| raise ValueError( | |
| "Invalid option combination 'sym_pos':False " | |
| "and 'cholesky':True: Cholesky decomposition is only possible " | |
| "for symmetric positive definite matrices.") | |
| cholesky = cholesky or (cholesky is None and sym_pos and not lstsq) | |
| x, status, message, iteration = _ip_hsd(A, b, c, c0, alpha0, beta, | |
| maxiter, disp, tol, sparse, | |
| lstsq, sym_pos, cholesky, | |
| pc, ip, permc_spec, callback, | |
| postsolve_args) | |
| return x, status, message, iteration | |