Adel Saleh
FinSeg: finger parts semantic segmentation using multi-scale feature maps aggregation of FCN.
Saleh, Adel; Rashwan, Hatem; Abdel-Nasser, Mohamed; Singh, Vivek; Abdulwahab, Saddam; Sarker, Md.; Garcia, Miguel; Puig, Domenec
Authors
Hatem Rashwan
Mohamed Abdel-Nasser
Vivek Singh
Saddam Abdulwahab
Md. Sarker
Miguel Garcia
Domenec Puig
Abstract
Image semantic segmentation is in the center of interest for computer vision researchers. Indeed, huge number of applications requires efficient segmentation performance, such as activity recognition, navigation, and human body parsing, etc. One of the important applications is gesture recognition that is the ability to understanding human hand gestures by detecting and counting finger parts in a video stream or in still images. Thus, accurate finger parts segmentation yields more accurate gesture recognition. Consequently, in this paper, we highlight two contributions as follows: First, we propose data-driven deep learning pooling policy based on multi-scale feature maps extraction at different scales (called FinSeg). A novel aggregation layer is introduced in this model, in which the features maps generated at each scale is weighted using a fully connected layer. Second, with the lack of realistic labeled finger parts datasets, we propose a labeled dataset for finger parts segmentation (FingerParts dataset). To the best of our knowledge, the proposed dataset is the first attempt to build a realistic dataset for finger parts semantic segmentation. The experimental results show that the proposed model yields an improvement of 5% compared to the standard FCN network.
Citation
SALEH, A., RASHWAN, H., ABDEL-NASSER, M., SINGH, V., ABDULWAHAB, S., SARKER, M., GARCIA, M. and PUIG, D. 2019. FinSeg: finger parts semantic segmentation using multi-scale feature maps aggregation of FCN. In Tremeau, A., Farinella, G.M. and Braz, J. (eds.). Proceedings of 14th international joint conferences on Computer vision, imaging and computer graphics theory and applications 2019 (VISIGRAPP 2019), 25-27 February 2019, Prague, Czech Republic. Setúbal, Portugal: SciTePress [online], 5, pages 77-84. Available from: https://doi.org/10.5220/0007382100770084
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | 14th international joint conferences on Computer vision, imaging and computer graphics theory and applications 2019 (VISIGRAPP 2019) |
Start Date | Feb 25, 2019 |
End Date | Feb 27, 2019 |
Acceptance Date | Nov 29, 2018 |
Online Publication Date | Feb 27, 2019 |
Publication Date | Dec 31, 2019 |
Deposit Date | Dec 4, 2021 |
Publicly Available Date | Jan 20, 2022 |
Publisher | SciTePress |
Peer Reviewed | Peer Reviewed |
Volume | 5 |
Pages | 77-84 |
ISBN | 9789897583544 |
DOI | https://doi.org/10.5220/0007382100770084 |
Keywords | Semantic segmentation; Fully convolutional network; Pixel-wise classification; Finger parts |
Public URL | https://rgu-repository.worktribe.com/output/1542064 |
Files
SALEH 2019 FinSeg (VOR)
(1.5 Mb)
PDF
Publisher Licence URL
https://creativecommons.org/licenses/by-nc-nd/4.0/
Copyright Statement
© 2019 by SCITEPRESS.
You might also like
AWEU-Net: an attention-aware weight excitation U-Net for lung nodule segmentation.
(2021)
Journal Article
Downloadable Citations
About OpenAIR@RGU
Administrator e-mail: publications@rgu.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search