Nsopy
Methods for Non-Smooth Convex Optimization (NSO), written in Python
Install / Use
/learn @robin-vjc/NsopyREADME
nsopy - non-smooth optimization in Python
nsopy is a Python library implementing a set of first order methods to solve non-smooth, constrained convex optimization models.
Installation
pip install nsopy
Usage
Quick Example
We seek to minimize a function obtained by taking the max over a set of affine functions.
The feasible set considered is the set of non-negative real numbers, i.e., <img src="https://latex.codecogs.com/svg.latex?\mathbb{X}&space;=&space;\mathbb{R}^+" title="\mathbb{X} = \mathbb{R}^+"/>
for which the projection operation is straightforward.
It is straightforward to see that the optimum is at x* = 2.25; we can solve this optimization problem numerically as follows:
import numpy as np
def oracle(x_k):
# evaluation of the f_i components at x_k
fi_x_k = [-2*x_k + 2, -1.0/3*x_k + 1, x_k - 2]
f_x_k = max(fi_x_k) # function value at x_k
diff_fi = [-2, -1.0/3.0, 1] # gradients of the components
max_i = fi_x_k.index(f_x_k)
# subgradient at x_k is the gradient of the active function component; cast as (1x1 dimensional) np.array
diff_f_xk = np.array([diff_fi[max_i], ])
return 0, f_x_k, diff_f_xk
def projection_function(x_k):
if x_k is 0:
return np.array([0,])
else:
return np.maximum(x_k, 0)
Instantiation of method and logger, solve and print
from nsopy.methods.subgradient import SubgradientMethod
from nsopy.loggers import GenericMethodLogger
method = SubgradientMethod(oracle, projection_function, stepsize_0=0.1, stepsize_rule='constant', sense='min')
logger = GenericMethodLogger(method)
for iteration in range(200):
method.step()
Result:
>>> print(logger.x_k_iterates[-5:])
[2.1999999999999904, 2.216666666666657, 2.2333333333333236, 2.2499999999999902, 2.266666666666657]
Available Methods
- Standard Subgradient Method
SubgradientMethod(oracle, projection_function, dimension=0, stepsize_0=1.0, stepsize_rule='1/k', sense='min')
Stepsize rules valiable: stepsize_rule: ['constant', '1/k', '1/sqrt(k)']
- Quasi-Monotone Methods
Implementation of double simple averaging, and triple averaging methods from Nesterov's paper on quasi-monotone methods.
SGMDoubleSimpleAveraging(oracle, projection_function, dimension=0, gamma=1.0, sense='min')
SGMTripleAveraging(oracle, projection_function, dimension=0, variant=1, gamma=1.0, sense='min'):
Variants of SGMTripleAveraging available: variant: [1, 2]
- Universal Gradient Methods
Implementation of Nesterov's universal gradient methods, primal, dual and fast versions.
UniversalPGM(oracle, projection_function, dimension=0, epsilon=1.0, averaging=False, sense='min')
UniversalDGM(oracle, projection_function, dimension=0, epsilon=1.0, averaging=False, sense='min'):
UniversalFGM(oracle, projection_function, dimension=0, epsilon=1.0, averaging=False, sense='min'):
- Cutting Planes Method
Warning: this method requires gurobipy; if you are an academic, you can get a free license here.
CuttingPlanesMethod(oracle, projection_function, dimension=0, epsilon=0.01, search_box_min=-10, search_box_max=10, sense='min')
The parameter epsilon is the absolute required suboptimality level |f_k - f*| used as a stopping criterion. Note that a search box needs to be specified.
- Bundle Method
Warning: this method requires gurobipy; if you are an academic, you can get a free license here.
Implementation of a basic variant of the bundle method.
BundleMethod(oracle, projection_function, dimension=0, epsilon=0.01, mu=0.5, sense='min'):
Important Remarks
-
Methods have to either be instantiated with the appropriate dimension argument, or implement a special case for 0. The basic usage example above illustrates an oracle implementing such a special case. For this example, alternatively one could have instantiated the solution method with
dimension = 1. -
The first-order oracle must also provide a projection function; here is a list of cases for which the projection operation is computationally inexpensive.
-
Currently, all methods are implemented in Python. Numerical performance is not optimized, but they may be still useful for quick comparisons or for applications in which the main computational burden is in evaluating the first order oracle.
Advanced Examples
- See analytical example for a more challenging optimization model.
- How to get approximate solutions to structured MILPs using Lagrangian duality.
We can also use these methods to decompose stochastic multistage mixed integer programs (preview), which in turn allows the computation of approximate solutions to these models on distributed environments (e.g., on cloud infrastructure).
Contributing
Contributions and pull requests are very much welcome. The TODO contains a number of tasks whose completion would be helpful.
Related Skills
node-connect
353.3kDiagnose OpenClaw node connection and pairing failures for Android, iOS, and macOS companion apps
frontend-design
111.7kCreate distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.
openai-whisper-api
353.3kTranscribe audio via OpenAI Audio Transcriptions API (Whisper).
qqbot-media
353.3kQQBot 富媒体收发能力。使用 <qqmedia> 标签,系统根据文件扩展名自动识别类型(图片/语音/视频/文件)。
