Demajh logo Demajh, Inc.

Keyword Spotting with Hyper-Matched Filters for Small Footprint Devices: what it means for business leaders

Open-vocabulary keyword spotting for tiny devices: a hypernetwork turns text prompts into matched-filter weights that guide a compact Perceiver detector, coupled with Whisper-tiny or Conformer-tiny encoders, delivering state-of-the-art accuracy and strong out-of-domain robustness.

1. What the method is

A lightweight, open-vocabulary keyword-spotting approach that runs on small-footprint hardware while matching or surpassing larger systems. The model combines a speech encoder (Whisper-tiny or Conformer-tiny), a target keyword encoder that reads the keyword text and emits a compact set of convolution weights, and a Perceiver-based detection network. Those generated weights act like a keyword-specific matched filter, steering the detector toward the term of interest in streaming audio. Crucially, the keyword encoder can be used offline to pre-generate filter weights, so devices only carry the speech encoder and detector at runtime. Parameter counts on device range from roughly 4.2 M to ~10 M while retaining state-of-the-art detection quality across languages and acoustic conditions.

2. Why the method was developed

Organizations want wake words and custom commands that can be added instantly—without retraining heavy ASR models or shipping privacy-sensitive audio to the cloud. Prior open-vocabulary KWS methods either degrade badly when shrunk for embedded targets or remain too large, energy-hungry, and latency-prone for phones, wearables, or smart speakers. Query-by-example systems are brittle to recording conditions; transcription-first pipelines are overkill. This work aims to prove you can keep models tiny, keep the vocabulary open, and still deliver state-of-the-art accuracy and robust generalization, including to second-language speech, by turning the text keyword into a tuned matched filter that guides a compact attention stack.

3. Who should care

Product owners for voice assistants, OEMs building on-device wake-word or command spotting, call-center QA vendors needing fast term detection, and teams shipping regulated or offline experiences where cloud ASR is unacceptable. Platform leads consolidating model catalogs across devices can use one small engine for many dynamic keywords. Operations leaders targeting longer battery life, lower BOM costs, and predictable latency—especially in multilingual products—benefit from an approach that keeps most weights on device small while generating new keyword filters off device.

4. How the method works

Audio is encoded with Whisper-tiny (~7.6 M parameters) or a ~3.7 M-parameter Conformer-tiny to produce frame embeddings. The target keyword (as characters) is fed to a small hypernetwork that outputs weights for a depth-wise convolution—interpretable as a keyword-specific matched filter. That convolution modulates a Perceiver module’s cross-attention, producing a compact latent that a small head scores as “present/absent.” Training uses Binary Cross-Entropy; the Conformer encoder can be pre-trained with CTC on public speech corpora, then the detector is trained for open-vocabulary KWS with curated negative sampling (nearest negatives, character swaps). Only the detector and, when desired, the encoder are fine-tuned for downstream datasets; keyword filters can be generated offline and cached.

5. How it was evaluated

Experiments span multilingual VoxPopuli for open-vocabulary KWS, LibriPhrase (Easy/Hard) for phrase spotting, and out-of-domain sets such as Speech Commands V1 (10 keywords), FLEURS low-resource languages, and Wildcat Diapix dialogues with both native (L1) and non-native (L2) English speakers. Metrics include AUC, F1, Equal Error Rate (EER), and FRR at 5% FAR, alongside strict accounting of on-device parameters that excludes the offline keyword-encoder. Baselines cover CED, EMKWS, CMCD, and a compact AdaKWS-tiny reimplementation. Perceiver depth (1–5 layers) is swept to map the size–accuracy trade-off.

6. How it performed

Conformer-tiny variants delivered the best accuracy–size balance. On VoxPopuli, mid-depth Perceiver models edged compact baselines while using far fewer on-device parameters; performance remained strong across languages, with predictable dips in very low-resource settings. On LibriPhrase, AUC approached 99.9 and EER hovered near ~1% for the best settings, surpassing prior small-footprint open-vocabulary approaches. Out-of-domain tests (FLEURS, Speech Commands, Wildcat Diapix) showed robust transfer, including to L2 speech, and the smallest ~4.2 M-parameter model matched or beat larger baselines in several scenarios—useful for edge deployments where every milliwatt matters. (Source: arXiv 2508.04857, 2025)

← Back to dossier index