Abstract:
Presence of speech and motion artifacts has been shown to impact the performance of
wearable sensor systems used for automatic detection of food intake. This work presents a novel
wearable device which can detect food intake even when the user is physically active and/or talking.
The device consists of a piezoelectric strain sensor placed on the temporalis muscle, an accelerometer,
and a data acquisition module connected to the temple of eyeglasses. Data from 10 participants was
collected while they performed activities including quiet sitting, talking, eating while sitting, eating
while walking, and walking. Piezoelectric strain sensor and accelerometer signals were divided into
non-overlapping epochs of 3 s; four features were computed for each signal. To differentiate between
eating and not eating, as well as between sedentary postures and physical activity, two multiclass
classification approaches are presented. The first approach used a single classifier with sensor
fusion and the second approach used two-stage classification. The best results were achieved when
two separate linear support vector machine (SVM) classifiers were trained for food intake and
activity detection, and their results were combined using a decision tree (two-stage classification) to
determine the final class. This approach resulted in an average F1-score of 99.85% and area under
the curve (AUC) of 0.99 for multiclass classification. With its ability to differentiate between food
intake and activity level, this device may potentially be used for tracking both energy intake and
energy expenditure.