Abstract
Wearable sensors (e.g., lifelogging cameras) represent very useful tools to monitor people's daily habits and lifestyle. Wearable cameras are able to continuously capture different moments of the day of their wearers, their environment, and interactions with objects, people, and places reflecting their personal lifestyle. The food places where people eat, drink, and buy food, such as restaurants, bars, and supermarkets, can directly affect their daily dietary intake and behavior. Consequently, developing an automated monitoring system based on analyzing a person's food habits from daily recorded egocentric photo-streams of the food places can provide valuable means for people to improve their eating habits. This can be done by generating a detailed report of the time spent in specific food places by classifying the captured food place images to different groups. In this paper, we propose a self-attention mechanism with multi-scale atrous convolutional networks to generate discriminative features from image streams to recognize a predetermined set of food place categories. We apply our model on an egocentric food place dataset called 'EgoFoodPlaces' that comprises of 43 392 images captured by 16 individuals using a lifelogging camera. The proposed model achieved an overall classification accuracy of 80% on the 'EgoFoodPlaces' dataset, respectively, outperforming the baseline methods, such as VGG16, ResNet50, and InceptionV3.
Original language | English |
---|---|
Article number | 8671710 |
Pages (from-to) | 39069-39082 |
Number of pages | 14 |
Journal | IEEE Access |
Volume | 7 |
DOIs | |
Publication status | Published - 2019 |
Externally published | Yes |
Bibliographical note
Funding Information:This work was supported in part by the program Marti Franques under the agreement between Universitat Rovira Virgili and Fundacio Catalunya La Pedrera under Project TIN2015-66951-C2, Project SGR 1742, and Project CERCA, in part by the Nestore Horizon2020 SC1-PM-15-2017 under Grant 769643, in part by the EIT Validithi, in part by the ICREA Academia 2014, and in part by the NVIDIA Corporation.
Publisher Copyright:
© 2013 IEEE.