Published April 3, 2026 | Version v1
Publication Open

Linear Programming Formulation of Relu Function

Description

Relu function is a non-linear function. But if we formulate it as a hard constraint in linear programming, we can make it becomes a convex optimization problem as linear programming is convex. If we will to imagine Relu function as a linear inequality in linear programming, it will look something like inequalities. If we look at an example of feasible region in linear programming graphically, it will contains a few lines of inequalities where green strips mark the invalid regions. So in a BSnet Relu network, the Relu activation functions will act like inequalities and the BSnet can be represented by linear programming. The linear transform by the weights in a layer in BSnet will reorientate the inequality hyperplanes in the linear programming.

Files

Linear_Programming_Relu.pdf

Files (260.8 kB)

Name Size Download all
md5:6ed9b68194a37e46e85cd3531d1e1856
260.8 kB Preview Download