Please use this identifier to cite or link to this item: http://dspace.dtu.ac.in:8080/jspui/handle/repository/21671
Full metadata record
DC FieldValueLanguage
dc.contributor.authorYAMAN-
dc.contributor.authorSHONDHANI, DEEPANSHU-
dc.date.accessioned2025-06-12T05:12:18Z-
dc.date.available2025-06-12T05:12:18Z-
dc.date.issued2025-05-
dc.identifier.urihttp://dspace.dtu.ac.in:8080/jspui/handle/repository/21671-
dc.description.abstractEffective neural network weight initialization is crucial for successful training, yet standard methods often rely on assumptions violated by modern architectures and advanced activation functions like Swish. This dissertation details a study investigating the feasibility of using Reinforcement Learning (RL) to tune a scaling factor for He initialization when employing Swish activations. An RL agent explored different scaling factors, evaluating them by training a small convolutional neural network on CIFAR-10 for a 7-epoch proxy task. Over 200 episodes, the RL agent demonstrated learning, converging towards a specific range of scaling factors that optimized the 7-epoch validation accuracy. This preliminary investigation highlights the functionality of the RL framework for initialization tuning and underscores the importance of evaluating the fidelity of short-term proxy tasks in predicting longer-term training performance, informing subsequent research into more complex symbolic initialization discovery.en_US
dc.language.isoenen_US
dc.relation.ispartofseriesTD-7876;-
dc.subjectREINFORCEMENT LEARNINGen_US
dc.subjectWEIGHT INITIALIZATIONen_US
dc.subjectDEEP NEURAL NETWORKSen_US
dc.subjectACTIVATION FUNCTIONSen_US
dc.titleREINFORCEMENT LEARNING FOR WEIGHT INITIALIZATIONen_US
dc.typeThesisen_US
Appears in Collections:M Sc Applied Maths

Files in This Item:
File Description SizeFormat 
YAMAN & DEEPANSHU m.sC..pdf1.06 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.