Ajou University repository

Learning Neural Networks without Lazy Weights
  • Lee, Dong Gi ;
  • Cho, Junhee ;
  • Kim, Myungjun ;
  • Park, Sunghong ;
  • Shin, Hyunjung
Citations

SCOPUS

1

Citation Export

DC Field Value Language
dc.contributor.authorLee, Dong Gi-
dc.contributor.authorCho, Junhee-
dc.contributor.authorKim, Myungjun-
dc.contributor.authorPark, Sunghong-
dc.contributor.authorShin, Hyunjung-
dc.date.issued2022-01-01-
dc.identifier.urihttps://aurora.ajou.ac.kr/handle/2018.oak/36787-
dc.identifier.urihttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85127599667&origin=inward-
dc.description.abstractVarious approaches have been suggested for the regularization of neural networks, including the well-known Dropout and Dropconnect, which are simple and efficient to implement and therefore have been widely used. However, there is a risk of loss of well-trained weights when dropping nodes or weights randomly. In this paper, we propose a regularization method that preserves well-trained weights and removes poorly trained weights. This was motivated by the observation that the trained weights become further trained. We define these as eager weights whereas the opposite as lazy weights. On every weight update, the distribution of the changes in weight values is examined, and the lazy weights are removed layer-wise. The results demonstrate that the proposed method has a faster convergence rate, avoids overfitting, and outperforms competing methods on the classification of benchmark datasets.-
dc.description.sponsorshipACKNOWLEDGMENT The authors would like to gratefully acknowledge supported from the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2021R1A2C2003474), BK21 FOUR program of the National Research Foundation of Korea funded by the Ministry of Education (NRF5199991014091) and the Ajou University research fund.-
dc.language.isoeng-
dc.publisherInstitute of Electrical and Electronics Engineers Inc.-
dc.subject.meshDeep learning-
dc.subject.meshLayer-wise-
dc.subject.meshLearning neural networks-
dc.subject.meshNeural-networks-
dc.subject.meshOverfitting-
dc.subject.meshRegularisation-
dc.subject.meshRegularization methods-
dc.subject.meshSimple++-
dc.subject.meshWeight update-
dc.subject.meshWeight values-
dc.titleLearning Neural Networks without Lazy Weights-
dc.typeConference-
dc.citation.conferenceDate2022.1.17. ~ 2022.1.20.-
dc.citation.conferenceName2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022-
dc.citation.editionProceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022-
dc.citation.endPage87-
dc.citation.startPage82-
dc.citation.titleProceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022-
dc.identifier.bibliographicCitationProceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022, pp.82-87-
dc.identifier.doi10.1109/bigcomp54360.2022.00026-
dc.identifier.scopusid2-s2.0-85127599667-
dc.identifier.urlhttp://ieeexplore.ieee.org/xpl/mostRecentIssue.jsp?punumber=9736461-
dc.subject.keywordDeep learning-
dc.subject.keywordNeural networks-
dc.subject.keywordOverfitting-
dc.subject.keywordRegularization-
dc.type.otherConference Paper-
dc.description.isoafalse-
dc.subject.subareaArtificial Intelligence-
dc.subject.subareaComputer Science Applications-
dc.subject.subareaComputer Vision and Pattern Recognition-
dc.subject.subareaInformation Systems and Management-
dc.subject.subareaHealth Informatics-
Show simple item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Shin, HyunJung Image
Shin, HyunJung신현정
Department of Industrial Engineering
Read More

Total Views & Downloads

File Download

  • There are no files associated with this item.