arXiv Analytics

Sign in

arXiv:2309.00203 [cs.LG]AbstractReferencesReviewsResources

Data-Driven Projection for Reducing Dimensionality of Linear Programs: Generalization Bound and Learning Methods

Shinsaku Sakaue, Taihei Oki

Published 2023-09-01Version 1

This paper studies a simple data-driven approach to high-dimensional linear programs (LPs). Given data of past $n$-dimensional LPs, we learn an $n\times k$ \textit{projection matrix} ($n > k$), which reduces the dimensionality from $n$ to $k$. Then, we address future LP instances by solving $k$-dimensional LPs and recovering $n$-dimensional solutions by multiplying the projection matrix. This idea is compatible with any user-preferred LP solvers, hence a versatile approach to faster LP solving. One natural question is: how much data is sufficient to ensure the recovered solutions' quality? We address this question based on the idea of \textit{data-driven algorithm design}, which relates the amount of data sufficient for generalization guarantees to the \textit{pseudo-dimension} of performance metrics. We present an $\tilde{\mathrm{O}}(nk^2)$ upper bound on the pseudo-dimension ($\tilde{\mathrm{O}}$ compresses logarithmic factors) and complement it by an $\Omega(nk)$ lower bound, hence tight up to an $\tilde{\mathrm{O}}(k)$ factor. On the practical side, we study two natural methods for learning projection matrices: PCA- and gradient-based methods. While the former is simple and efficient, the latter sometimes leads to better solution quality. Experiments confirm that learned projection matrices are beneficial for reducing the time for solving LPs while maintaining high solution quality.

Related articles: Most relevant | Search more
arXiv:2308.04011 [cs.LG] (Published 2023-08-08)
Generalization bound for estimating causal effects from observational network data
arXiv:2404.19456 [cs.LG] (Published 2024-04-30)
Imitation Learning: A Survey of Learning Methods, Environments and Metrics
arXiv:1810.02180 [cs.LG] (Published 2018-10-04)
Improved generalization bounds for robust learning