Md. Mostafa Kamal Sarker
Recognizing food places in egocentric photo-streams using multi-scale atrous convolutional networks and self-attention mechanism.
Sarker, Md. Mostafa Kamal; Rashwan, Hatem A.; Akram, Farhan; Talavera, Estefania; Banu, Syeda Furruka; Radeva, Petia; Puig, Domenec
Authors
Hatem A. Rashwan
Farhan Akram
Estefania Talavera
Syeda Furruka Banu
Petia Radeva
Domenec Puig
Abstract
Wearable sensors (e.g., lifelogging cameras) represent very useful tools to monitor people's daily habits and lifestyle. Wearable cameras are able to continuously capture different moments of the day of their wearers, their environment, and interactions with objects, people, and places reflecting their personal lifestyle. The food places where people eat, drink, and buy food, such as restaurants, bars, and supermarkets, can directly affect their daily dietary intake and behavior. Consequently, developing an automated monitoring system based on analyzing a person's food habits from daily recorded egocentric photo-streams of the food places can provide valuable means for people to improve their eating habits. This can be done by generating a detailed report of the time spent in specific food places by classifying the captured food place images to different groups. In this paper, we propose a self-attention mechanism with multi-scale atrous convolutional networks to generate discriminative features from image streams to recognize a predetermined set of food place categories. We apply our model on an egocentric food place dataset called 'EgoFoodPlaces' that comprises of 43 392 images captured by 16 individuals using a lifelogging camera. The proposed model achieved an overall classification accuracy of 80% on the 'EgoFoodPlaces' dataset, respectively, outperforming the baseline methods, such as VGG16, ResNet50, and InceptionV3.
Citation
SARKER, M.M.K., RASHWAN, H.A., AKRAM, F., TALAVERA, E., BANU, S.F., RADEVA, P. and PUIG, D. 2019. Recognizing food places in egocentric photo-streams using multi-scale atrous convolutional networks and self-attention mechanism. IEEE access [online], 7, pages 39069-39082. Available from: https://doi.org/10.1109/ACCESS.2019.2902225
Journal Article Type | Article |
---|---|
Acceptance Date | Feb 18, 2019 |
Online Publication Date | Mar 20, 2019 |
Publication Date | Dec 31, 2019 |
Deposit Date | Dec 4, 2021 |
Publicly Available Date | Mar 29, 2022 |
Journal | IEEE Access |
Electronic ISSN | 2169-3536 |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Peer Reviewed | Peer Reviewed |
Volume | 7 |
Pages | 39069-39082 |
DOI | https://doi.org/10.1109/ACCESS.2019.2902225 |
Keywords | Food places recognition; Scene classification; Self-attention model; Atrous convolutional networks; Egocentric photo-streams; Visual lifelogging |
Public URL | https://rgu-repository.worktribe.com/output/1542055 |
Files
SARKER 2019 Recognizing food places (VOR)
(3.1 Mb)
PDF
Publisher Licence URL
https://creativecommons.org/licenses/by-nc-nd/4.0/
Copyright Statement
© 2019 The Authors.
You might also like
AWEU-Net: an attention-aware weight excitation U-Net for lung nodule segmentation.
(2021)
Journal Article
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search