Automated Facial Recognition in Older Photographs Using One-Shot Learning in Siamese Networks and Transfer Learning
Main Article Content
Abstract
A lot of historical information comes in various forms, such as old documents, papers, photographs, videos, audio, and even artefacts and sculptures. Photographs, audio, and videos are especially important because they effectively convey information. When we convert these into digital versions, it becomes easy to share, access online or offline, copy, move around, back up, and store in numerous places. However, a challenge with digital content is that it is often difficult to search due to the absence of readable text. Consequently, we cannot effectively analyse and utilise critical information. To make it useful, we manually look at pictures and add tags to create labels. While basic labels suffice for most searches, it becomes more complicated when dealing with a large number of photographs. Enhancements in search capabilities are needed to make the process easier, quicker, and more efficient. Fortunately, recent technological advances, such as artificial intelligence, provide us with facilities to simplify this process. This paper explores how artificial intelligence can streamline this process, enhancing search efficiency and enabling automatic identification and tagging of individuals in photos, thus facilitating easier access and analysis of digital archives.
It is anticipated that manual tagging efforts could be reduced by approximately 80%, and the searchability of photographs could be enhanced by about 84%.
Downloads
Metrics
Article Details
This work is licensed under a Creative Commons Attribution 4.0 International License.
You are free to:
- Share — copy and redistribute the material in any medium or format for any purpose, even commercially.
- Adapt — remix, transform, and build upon the material for any purpose, even commercially.
- The licensor cannot revoke these freedoms as long as you follow the license terms.
Under the following terms:
- Attribution — You must give appropriate credit , provide a link to the license, and indicate if changes were made . You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use.
- No additional restrictions — You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits.
Notices:
You do not have to comply with the license for elements of the material in the public domain or where your use is permitted by an applicable exception or limitation .
No warranties are given. The license may not give you all of the permissions necessary for your intended use. For example, other rights such as publicity, privacy, or moral rights may limit how you use the material.
References
Alex Krizhevsky et al., “Imagenet classification with deep convolutional neural networks,” Journal of Machine Learning Research, 2012.
Karen Simonyan et al., “Very Deep Convolutional Networks for Large-Scale Image Recognition,” arXiv preprint arXiv:1409.1556, 2014.
Shaoqing Ren et al., “Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), 2015.
Joseph Redmon et al., “YOLO9000: Better, Faster, Stronger,” arXiv preprint arXiv:1612.08242, 2016.
Gregory Koch et al., “Siamese Neural Networks for One-shot Image Recognition,” Advances in Neural Information Processing Systems (NeurIPS), 2015.
Kaiming He et al., “Deep Residual Learning for Image Recognition,” Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 2016.
Kaiming He et al., “Mask R-CNN,” Proceedings of the IEEE international conference on computer vision (ICCV), 2017.
Christian Szegedy et al., “Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning,” Proceedings of the AAAI Conference on Artificial Intelligence, 2016.
Gao Huang et al., “DenseNet: Densely Connected Convolutional Networks,” Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR), 2017.
Jifeng Dai et al., “R-FCN: Object Detection via Region-based Fully Convolutional Networks,” Advances in Neural Information Processing Systems (NeurIPS), 2016.
Andrew G. Howard et al., “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications,” arXiv preprint arXiv:1704.04861, 2017.
Yunpeng Chen et al., “Dual Path Networks,” Proceedings of the IEEE international conference on computer vision (ICCV), 2017.
Forrest N. Iandola et al., “SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size,” arXiv preprint arXiv:1602.07360, 2016.
Hyeonwoo Noh et al., “Learning Deconvolution Network for Semantic Segmentation,” Proceedings of the IEEE international conference on computer vision (ICCV), 2015.
Liangming Pan et al., “Learning to Compare: Siamese Network for Knowledge Graph Embedding,” arXiv preprint arXiv:1702.03814, 2017.