[{"authors":null,"categories":null,"content":"Linear Regions I talked a fair bit about this idea of a binary vector that results from embedding a ReLU activated neural network into a MIP. This is denoted as $\\mathscr{Z}(x)$ throughout the presentation. But where does that come from?\nSince the ReLU function is just the max between 0 and its input, we can model this using big-M constraints, $out \\ge 0, out \\ge in, out \\le in + Mz, out \\le M(1-z)$.\nExtending to Joint and Multiple Chance Constraints I only really cover the case in which we have a single chance constraint, but I mentioned that it extends to joint and multiple chance constraints. How does this happen?\nMultiple Chance Constraints To have multiple chance constraints, you just treat each as its own constraint. Since there is no overlap, you can add another set of constraints that correspond to each chance constraint that you want to consider.\nFor neural networks, we are able to use multiple outputs, where each output node corresponds to a different chance constraint. Then, you add the feasibility constraints individually on each of these outputs, corresponding to each chance constraint that you have. For example $\\mathscr{N}_1(x)$ could be the probability that $x$ violates the first chance constraint, so then you could add a constraint $\\mathscr{N}_1(x) \\le \\varepsilon_1$ for chance constraint 1 (where $\\mathscr{N}_1(x)$ is the first element of the vector produced by $\\mathscr{N}(x)$), $\\mathscr{N}_2(x) \\le \\varepsilon_2$ for chance constraint 2, etc.\nJoint Chance Constraints I’d strongly recommend reading more formal literature if you are more interested in join chance constraints in the SAA case. The very rough explanation of it is that there are multiple components to the chance constraint, so one “scenario” of a chance constraint really means that 2 or 3 or however many components of the joint chance constraint are all generated, and then all need to be satisfied for the “scenario” to be satisfied. We can abstract most of this away since for the neural network, we just need to be able to get a probability of the joint chance constraint being violated.\nIn-Sample vs Out-Of-Sample This idea is somewhat more common in machine learning than it is stochastic optimization. Essentially, the goal is to check how well our solution that we get generalizes to the underlying distribution. One way that we can go about this is by taking another set of samples from the underlying distribution, and checking its error. For our purpose, we generate the same number of scenarios to generate a solution from the MIP (In-Sample Scenarios) as we do to calculate the violation on a separate dataset from the underlying distribution (Out-of-Sample Scenarios). This can help to be some measure of how well our solution generalizes to the underlying distribution\n","date":1773160571,"expirydate":-62135596800,"kind":"page","lang":"en","lastmod":1773160571,"objectID":"c2d1628857873ae94e293ea6ca90530a","permalink":"https://maddiecolbert.com/post/iosc/","publishdate":"2026-03-10T11:36:11-05:00","relpermalink":"/post/iosc/","section":"post","summary":"Linear Regions I talked a fair bit about this idea of a binary vector that results from embedding a ReLU activated neural network into a MIP. This is denoted as $\\mathscr{Z}(x)$ throughout the presentation. But where does that come from?\nSince the ReLU function is just the max between 0 and its input, we can model this using big-M constraints, $out \\ge 0, out \\ge in, out \\le in + Mz, out \\le M(1-z)$.\n","tags":null,"title":"IOSC","type":"post"},{"authors":null,"categories":null,"content":"The following is the general MIP formulation of a ReLU function $$ out \\ge in, out \\ge 0$$\n","date":1773084452,"expirydate":-62135596800,"kind":"page","lang":"en","lastmod":1773084452,"objectID":"d086daca481abbc1486b8c2b6473e9d7","permalink":"https://maddiecolbert.com/post/linear_regions/","publishdate":"2026-03-09T14:27:32-05:00","relpermalink":"/post/linear_regions/","section":"post","summary":"The following is the general MIP formulation of a ReLU function $$ out \\ge in, out \\ge 0$$\n","tags":null,"title":"Linear_regions","type":"post"},{"authors":null,"categories":null,"content":"This is a first post This is my first post using hugo\n","date":1773073612,"expirydate":-62135596800,"kind":"page","lang":"en","lastmod":1773073612,"objectID":"ef6f966c97cdd4a9383ff0282697f419","permalink":"https://maddiecolbert.com/post/first_post/","publishdate":"2026-03-09T11:26:52-05:00","relpermalink":"/post/first_post/","section":"post","summary":"This is a first post This is my first post using hugo\n","tags":null,"title":"First_post","type":"post"}]