Abstract
In this paper, we study a deep learning (DL)-based multimodal technology for military, surveillance, and defense applications based on a pixel-by-pixel classification of soldier’s image dataset. We explore the acquisition of images from a remote tactical-robot to a ground station, where the detection and tracking of soldiers can help the operator to take actions or automate the tactical-robot in battlefield. The soldier detection is achieved by training a convolutional neural network to learn the patterns of the soldier’s uniforms. Our CNN learns from the initial dataset and from the actions taken by the operator, as opposed to the old-fashioned and hard-coded image processing algorithms. Our system attains an accuracy of over 81% in distinguishing the specific soldier uniform and the background. These experimental results prove our hypothesis that dilated convolutions can increase the segmentation performance when compared with patch-based, and fully connected networks.
| Original language | English |
|---|---|
| Pages (from-to) | 476-487 |
| Number of pages | 12 |
| Journal | Applied Artificial Intelligence |
| Volume | 35 |
| Issue number | 6 |
| DOIs | |
| State | Published - 2021 |
| Externally published | Yes |
Fingerprint
Dive into the research topics of 'Pattern recognition of soldier uniforms with dilated convolutions and a modified encoder-decoder neural network architecture'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver