Score: 1

LLM-Guided Exemplar Selection for Few-Shot Wearable-Sensor Human Activity Recognition

Published: December 26, 2025 | arXiv ID: 2512.22385v1

By: Elsen Ronando, Sozo Inoue

Potential Business Impact:

Helps computers learn from fewer examples.

Business Areas:
Image Recognition Data and Analytics, Software

In this paper, we propose an LLM-Guided Exemplar Selection framework to address a key limitation in state-of-the-art Human Activity Recognition (HAR) methods: their reliance on large labeled datasets and purely geometric exemplar selection, which often fail to distinguish similar weara-ble sensor activities such as walking, walking upstairs, and walking downstairs. Our method incorporates semantic reasoning via an LLM-generated knowledge prior that captures feature importance, inter-class confusability, and exemplar budget multipliers, and uses it to guide exemplar scoring and selection. These priors are combined with margin-based validation cues, PageRank centrality, hubness penalization, and facility-location optimization to obtain a compact and informative set of exemplars. Evaluated on the UCI-HAR dataset under strict few-shot conditions, the framework achieves a macro F1-score of 88.78%, outperforming classical approaches such as random sampling, herding, and $k$-center. The results show that LLM-derived semantic priors, when integrated with structural and geometric cues, provide a stronger foundation for selecting representative sensor exemplars in few-shot wearable-sensor HAR.

Country of Origin
🇯🇵 Japan

Page Count
37 pages

Category
Computer Science:
Computation and Language