Score: 1

Network Cross-Validation and Model Selection via Subsampling

Published: April 9, 2025 | arXiv ID: 2504.06903v1

By: Sayan Chakrabarty, Srijan Sengupta, Yuguo Chen

Potential Business Impact:

Tests computer networks better and faster.

Business Areas:
Content Delivery Network Content and Publishing

Complex and larger networks are becoming increasingly prevalent in scientific applications in various domains. Although a number of models and methods exist for such networks, cross-validation on networks remains challenging due to the unique structure of network data. In this paper, we propose a general cross-validation procedure called NETCROP (NETwork CRoss-Validation using Overlapping Partitions). The key idea is to divide the original network into multiple subnetworks with a shared overlap part, producing training sets consisting of the subnetworks and a test set with the node pairs between the subnetworks. This train-test split provides the basis for a network cross-validation procedure that can be applied on a wide range of model selection and parameter tuning problems for networks. The method is computationally efficient for large networks as it uses smaller subnetworks for the training step. We provide methodological details and theoretical guarantees for several model selection and parameter tuning tasks using NETCROP. Numerical results demonstrate that NETCROP performs accurate cross-validation on a diverse set of network model selection and parameter tuning problems. The results also indicate that NETCROP is computationally much faster while being often more accurate than the existing methods for network cross-validation.

Repos / Data Links

Page Count
76 pages

Category
Statistics:
Methodology