Frequency-Based Deep Occlusion Awareness Instance Segmentation

dc.contributor.author Guzel, Yasin
dc.contributor.author Aydin, Zafer
dc.contributor.author Talu, Muhammed Fatih
dc.date.accessioned 2026-03-23T14:49:38Z
dc.date.available 2026-03-23T14:49:38Z
dc.date.issued 2026
dc.description.abstract One major challenge faced by deep learning-based methods that detect target objects in the form of bounding boxes is object occlusion. High degrees of occlusion significantly diminish the accuracy of instance segmentation. Nonetheless, complex-valued Fourier descriptors can robustly represent object boundaries using minimal information. In this study, the impact of integrating Fourier descriptors-renowned for their strong representational capacity-with deep network models (UNet) that exhibit high generalization performance on instance segmentation accuracy was investigated. Within the scope of the research, nine network models were designed based on different strategies for utilizing frequency components. These variants fall into four strategy families: (i) UNet-style spectrum regression on fixed low-frequency windows (FUNet), (ii) magnitude-guided frequency selection/ROI construction (FUNet-Thr, FUNet-BBox), (iii) sequence models over tokenized FFT coefficients (BiLSTM Patch/Sorted), and (iv) encoder-only spectrum predictors with different depth/capacity (EncoderFFT1/2). To fairly evaluate the models' performance in segmenting objects subjected to disruptive factors (e.g., occlusion, blurring, noise), a specialized synthetic dataset was prepared. The task is formulated as single-target (single-instance), single-class segmentation. This dataset, automatically generated according to initial parameter values, contains images of objects moving at various speeds within a single frame. Among these models, the one termed FUNet, which relies on partial matching of central frequency components, achieved the highest segmentation accuracy despite the disruptive effects. Under the challenging Dataset 8 setting, the proposed FUNet achieved the highest overlap-based performance (Dice = 0.9329, IoU = 0.8842) among Attention U-Net, U-Net, and FourierNet, with statistically significant gains confirmed by paired per-image tests.
dc.identifier.doi 10.3390/math14050792
dc.identifier.issn 2227-7390
dc.identifier.uri https://hdl.handle.net/20.500.12573/5836
dc.identifier.uri https://doi.org/10.3390/math14050792
dc.language.iso en
dc.publisher MDPI
dc.rights info:eu-repo/semantics/openAccess
dc.subject Deep Learning
dc.subject Segmentation
dc.subject Frequency Domain
dc.subject Fourier Transform
dc.title Frequency-Based Deep Occlusion Awareness Instance Segmentation
dc.type Article
dspace.entity.type Publication
gdc.description.department Abdullah Gül University
gdc.description.departmenttemp [Guzel, Yasin] Suleyman Demirel Univ, Dept Educ Sci, TR-32200 Isparta, Turkiye; [Aydin, Zafer] Abdullah Gul Univ, Dept Elect & Comp Engn, TR-38080 Kayseri, Turkiye; [Talu, Muhammed Fatih] Inonu Univ, Dept Comp Engn, TR-44200 Malatya, Turkiye
gdc.description.issue 5
gdc.description.publicationcategory Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı
gdc.description.volume 14
gdc.description.woscitationindex Science Citation Index Expanded
gdc.identifier.wos WOS:001713739000001
gdc.index.type WoS
gdc.virtual.author Aydın, Zafer
relation.isAuthorOfPublication a26c06af-eae3-407c-a21a-128459fa4d2f
relation.isAuthorOfPublication.latestForDiscovery a26c06af-eae3-407c-a21a-128459fa4d2f
relation.isOrgUnitOfPublication 52f507ab-f278-4a1f-824c-44da2a86bd51
relation.isOrgUnitOfPublication ef13a800-4c99-4124-81e0-3e25b33c0c2b
relation.isOrgUnitOfPublication 665d3039-05f8-4a25-9a3c-b9550bffecef
relation.isOrgUnitOfPublication.latestForDiscovery 52f507ab-f278-4a1f-824c-44da2a86bd51

Files