Automated Video Colorization Techniques for Enhanced Visual Realism and Computational Efficiency

Main Article Content

Zahoor M. Aydam
Nidhal K. El Abbadi

Abstract

Automatic video colorization remains a challenging computer vision task, particularly when ensuring semantic accuracy and temporal coherence across dynamic, multi-scene content. Existing methods often rely on a single fixed reference image, which fails to adapt to abrupt scene changes or variations in lighting and texture. This study presents a hybrid deep learning framework that dynamically selects multiple reference images per scene using adaptive thresholds derived from the Structural Similarity Index Measure (SSIM) and deep features extracted via a ResNet50 backbone with Generalized Mean Pooling (GeM). The framework integrates three specialized modules pre-processing, reference image processing, and attention-based colorization—operating in the Lab color space before conversion to RGB. Experimental evaluations on the YouTube-8M dataset demonstrate a PSNR of 37.89, SSIM of 0.998, and inference speed of 2.6 FPS with a compact 81 MB model (3.2M parameters). Compared to state-of-the-art methods, the proposed approach achieves superior color fidelity and temporal stability while maintaining efficiency, making it suitable for deployment in resource-constrained environments such as embedded vision and IoT systems.

Article Details

Section

Articles

How to Cite

[1]
Z. . M. Aydam and N. . K. El Abbadi , Trans., “Automated Video Colorization Techniques for Enhanced Visual Realism and Computational Efficiency”, Mesopotamian Journal of Computer Science, vol. 2025, pp. 258–287, Sep. 2025, doi: 10.58496/MJCSC/2025/017.

Similar Articles

You may also start an advanced similarity search for this article.