Please use this identifier to cite or link to this item: http://dspace.dtu.ac.in:8080/jspui/handle/repository/21787
Title: LIGHTWEIGHT NETWORKS FOR LOW-LIGHT IMAGE ENHANCEMENT BASED ON MULTI-OBJECTIVE KNOWLEDGE DISTILLATION
Authors: NARSHIHBHAI, JOSHI KULADIPKUMAR
Keywords: LIGHTWEIGHT NETWORKS
LOW-LIGHT IMAGE ENHANCEMENT
KNOWLEDGE DISTILLATION
Issue Date: May-2025
Series/Report no.: TD-7998;
Abstract: Low-light image enhancement is a critical computer vision challenge impacting visual quality and downstream tasks. While large deep learning models like MIRNet-v2 excel at restoring details and colors, their computational and memory demands hinder real-time or edge device deployment. Conversely, lightweight models such as Zero-DCE efficiently adjust illumination but may lack fine texture or structural recovery. This thesis introduces a novel framework for computationally efficient, high performing lightweight networks for low-light image enhancement using pro gressive multi-teacher knowledge distillation. We leverage the complementary expertise of MIRNet-v2 (˜26.6M parameters) for structural preservation and Zero DCE (˜0.1M parameters) for efficient illumination correction, training a compact student network to learn from both. Key innovations include: 1. Resolution-Progressive Training: The student network trains over 250 epochs, starting with 64x64 pixel images and progressively increasing resolution (to 128x128, then 256x256). This curriculum learning, with dynamic batch siz ing, ensures stable optimization and manages GPU memory. 2. Multi-Objective Hybrid Loss: A weighted 7-component loss guides stu dent learning, incorporating Charbonnier loss (Lrecon) for reconstruction, FFT-based L1 loss (Lfreq) for frequency fidelity, VGG perceptual loss (Lperc), MS-SSIM loss (Lmsssim) for structure, Sobel gradient L1 loss (Lgrad) for edges, a PatchGAN adversarial loss (Ladv-G), and contrastive distillation loss (Lcont) to mimic teacher features. 3. Lightweight Enhanced Student Architecture: The 9.99M parameter stu dent generator uses an efficient backbone of Recursive Residual Groups (RRGs) and Multi-scale Residual Blocks (MRBs). Optional Adaptive Fea ture Stretch (AFS) blocks for dynamic feature range expansion and Gradient Guided Convolution (GGC) blocks for edge awareness further augment its representational power. Experiments on the standard LOL (Low-Light) dataset show our model achiev ing a PSNR of 21.89 dB and an SSIM of 0.858. With total training parameters (including discriminator) of 23.6M, our approach significantly balances perfor mance and computational efficiency, offering a promising solution for deploying high-quality low-light enhancement in practical, resource-aware scenarios.
URI: http://dspace.dtu.ac.in:8080/jspui/handle/repository/21787
Appears in Collections:M.E./M.Tech. Computer Engineering

Files in This Item:
File Description SizeFormat 
JOSHI KULADIPKUMAR NARSHIHBHAI M.Tech.pdf3.26 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.