TY - JOUR
T1 - Handling Occlusions via Occlusion-aware Labels
AU - Amerehi, Fatemeh
AU - Healy, Patrick
N1 - Publisher Copyright:
© This is an open access article published by the IET under the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0/)
PY - 2024
Y1 - 2024
N2 - In real-world scenarios, many objects in view are often partially occluded, making the ability to handle occlusion essential for everyday activities. While human vision exhibits robustness to extreme occlusion, Convolutional Neural Networks struggle in this regard. Current regional masking strategies effectively improve generalization to occlusion. However, these methods typically eliminate informative pixels in training images by overlaying a patch of either random pixels or Gaussian noise, leading to a loss of image features. This limitation can be addressed by employing a more informative label. Rather than augmenting occluded images during training and assigning identical labels to both clean and occluded images, we differentiate between their labels. Essentially, we treat occlusion as a virtual class and assign it a virtual label. During training with occluded images, we merge the ground truth labels with these virtual labels, thereby informing the model about the presence of occlusion. Our findings indicate a 49.26% improvement in generalization for occlusion scenarios and an 8.22% enhancement on the common corruptions benchmark.
AB - In real-world scenarios, many objects in view are often partially occluded, making the ability to handle occlusion essential for everyday activities. While human vision exhibits robustness to extreme occlusion, Convolutional Neural Networks struggle in this regard. Current regional masking strategies effectively improve generalization to occlusion. However, these methods typically eliminate informative pixels in training images by overlaying a patch of either random pixels or Gaussian noise, leading to a loss of image features. This limitation can be addressed by employing a more informative label. Rather than augmenting occluded images during training and assigning identical labels to both clean and occluded images, we differentiate between their labels. Essentially, we treat occlusion as a virtual class and assign it a virtual label. During training with occluded images, we merge the ground truth labels with these virtual labels, thereby informing the model about the presence of occlusion. Our findings indicate a 49.26% improvement in generalization for occlusion scenarios and an 8.22% enhancement on the common corruptions benchmark.
KW - Augmentations
KW - Deep Neural Networks
KW - Generalization
KW - Occlusion
UR - http://www.scopus.com/inward/record.url?scp=85216746895&partnerID=8YFLogxK
U2 - 10.1049/icp.2024.3291
DO - 10.1049/icp.2024.3291
M3 - Conference article
AN - SCOPUS:85216746895
SN - 2732-4494
VL - 2024
SP - 103
EP - 109
JO - IET Conference Proceedings
JF - IET Conference Proceedings
IS - 10
T2 - 26th Irish Machine Vision and Image Processing Conference, IMVIP 2024
Y2 - 21 August 2024 through 23 August 2024
ER -