Poster
Constrained Online Convex Optimization with Polyak Feasibility Steps
Spencer Hutchinson · Mahnoosh Alizadeh
West Exhibition Hall B2-B3 #W-910
When machine-learning algorithms are deployed in the real world, they must respect system constraints that specify what they may and may not do. For instance, a control algorithm in an autonomous vehicle must never steer the car into an obstacle. We study learning under such constraints through the highly general framework of online convex optimization (OCO), which covers a broad range of adaptive learning tasks.Our key contribution is an OCO algorithm that always stays within these safety limits while relying only on local information and limited computation. It does so by incorporating Polyak feasibility steps into every update—small, principled adjustments that pull each decision back toward the safe region whenever it drifts too close to the boundary. These adjustments let the algorithm learn almost as efficiently as in an unconstrained setting, making it relevant to a wide range of safety-critical applications.