Aygün Çakıroğlu, M.Kizilkaya Aydoǧan, E.Bolatturk, Ö.F.Aydoğan, S.Ismailoǧullari, S.Delice, Y.2026-02-212026-02-2120261520-9512https://doi.org/10.1007/s11325-026-03594-2https://hdl.handle.net/20.500.12573/5791Purpose: To explore the feasibility of using camera-derived, non-contact audio synchronized with PSG for clinically relevant sleep-apnea classification, and to benchmark compact deep models under a subject-aware design using a previously unstudied, real-world dataset. Methods: Thirty-two adults underwent simultaneous polysomnography (PSG) and camera-based non-contact audio recording. The synchronized audio segments were used to train and compare three compact deep-learning architectures (convolutional, attention-augmented, and transformer-based) under a subject-aware evaluation design that prevented identity leakage. Model performance and calibration were assessed at both segment and subject levels using standard statistical tests. Results: Subject-level evaluation was based on a very small, imbalanced test set of six subjects (one positive). Within this limited yet previously unstudied local dataset, the CNN_trans model achieved an apparent perfect ranking performance (AUC = 1.00; 95% CI 0.00–1.00), though this likely reflects the small, imbalanced test cohort, with recall = 1.00 and precision = 0.55. The wide confidence interval reflects substantial statistical uncertainty, and DeLong comparisons showed no significant AUC difference between CNN_trans and CNN_att (ΔAUC = − 0.042; p = 0.43). Conclusion: PSG-synchronized, non-contact audio supports accurate and well-calibrated sleep-apnea classification with compact deep models. This subject-aware evaluation suggests that contactless acoustic monitoring may have potential clinical relevance, motivating larger, multi-site validation. © The Author(s), under exclusive licence to Springer Nature Switzerland AG 2026.eninfo:eu-repo/semantics/closedAccessContactless (Non-Contact) AudioDeep LearningModel Calibration and ExplainabilityPolysomnographySleep ApneaNon-Contact Acoustic Screening for Sleep Apnea: A Subject-Aware Deep Learning ApproachArticle10.1007/s11325-026-03594-22-s2.0-105029749384