Abstract
This paper presents a framework designed to enhance the robustness of human-centric computer vision systems operating under severe hardware constraints. The core challenge addressed is the performance degradation of authentication and dermatological analysis platforms in unconstrained imaging scenarios, characterized by variable illumination, occlusion, and a critical scarcity of labeled multi-modal data, particularly in thermal and clinical spectral domains. Our contribution is a unified pipeline integrating deep generative models for cross-modal synthetic data augmentation with a hybrid convolutional architecture, refined by embedded classical optimization layers. The generative component employs a Spectral-Consistent Generative Adversarial Network (SC-GAN) to produce physiologically plausible thermal and clinical image pairs from limited RGB-Depth inputs, effectively mitigating data paucity and privacy concerns. The discriminative backbone is a multi-stream convolutional neural network optimized for heterogeneous input fusion. A pivotal innovation is the integration of an Alternating Direction Method of Multipliers (ADMM) layer within the network's decoding stage, formulated as a learnable optimization step for precise segmentation and depth refinement. This layer enforces spatial consistency and boundary accuracy within the gradient-based learning paradigm. Extensive experimental validation on a newly compiled multi-modal dataset demonstrates that our hybrid model significantly outperforms conventional deep learning baselines. We report a mean increase of 18.7% in segmentation Intersection-over-Union (IoU) for skin lesion boundaries under challenging lighting, a 22.3% reduction in biometric authentication error rates in occluded scenarios, and a 15.9% decrease in generalization error on unseen domains. The architecture maintains an inference latency of 47.3 ms on a mobile system-on-chip, ensuring realtime device level deployments.



![Author ORCID: We display the ORCID iD icon alongside authors names on our website to acknowledge that the ORCiD has been authenticated when entered by the user. To view the users ORCiD record click the icon. [opens in a new tab]](https://www.cambridge.org/engage/assets/public/coe/logo/orcid.png)