Many (most?) column subset selection criteria are NP hard
By: Ilse C. F. Ipsen, Arvind K. Saibaba
Potential Business Impact:
Finds the most important parts of big numbers.
We consider a variety of criteria for selecting k representative columns from a real matrix A with rank(A)>=k. The criteria include the following optimization problems: absolute volume and S-optimality maximization; norm and condition minimization in the two-norm, Frobenius norm and Schatten p-norms for p>2; stable rank maximization; and the new criterion of relative volume maximization. We show that these criteria are NP hard and do not admit polynomial time approximation schemes (PTAS). To formulate the optimization problems as decision problems, we derive optimal values for the subset selection criteria, as well as expressions for partitioned pseudo-inverses.
Similar Papers
Subset selection for matrices in spectral norm
Numerical Analysis
Finds best data columns for faster computer math.
Pareto Optimization with Robust Evaluation for Noisy Subset Selection
Neural and Evolutionary Computing
Finds best groups of things even with bad information.
Optimality and computational barriers in variable selection under dependence
Statistics Theory
Finds the best way to pick important data.