Ajou University repository

MINIMUM WIDTH FOR UNIVERSAL APPROXIMATION USING RELU NETWORKS ON COMPACT DOMAIN
Citations

SCOPUS

0

Citation Export

DC Field Value Language
dc.contributor.authorKim, Namjun-
dc.contributor.authorMin, Chanho-
dc.contributor.authorPark, Sejun-
dc.date.issued2024-01-01-
dc.identifier.urihttps://aurora.ajou.ac.kr/handle/2018.oak/37103-
dc.identifier.urihttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85195527806&origin=inward-
dc.description.abstractIt has been shown that deep neural networks of a large enough width are universal approximators but they are not if the width is too small. There were several attempts to characterize the minimum width wmin enabling the universal approximation property; however, only a few of them found the exact values. In this work, we show that the minimum width for Lp approximation of Lp functions from [0, 1]dx to ℝdy is exactly max{dx, dy, 2} if an activation function is RELU-LIKE (e.g., RELU, GELU, SOFTPLUS). Compared to the known result for RELU networks, wmin = max{dx + 1, dy} when the domain is ℝdx, our result first shows that approximation on a compact domain requires smaller width than on Rdx. We next prove a lower bound on wmin for uniform approximation using general activation functions including RELU: wmin ≥ dy + 1 if dx < dy ≤ 2dx. Together with our first result, this shows a dichotomy between Lp and uniform approximations for general activation functions and input/output dimensions.-
dc.description.sponsorshipNK and SP were supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2019-0-00079, Artificial Intelligence Graduate School Program, Korea University) and Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2022R1F1A1076180).-
dc.language.isoeng-
dc.publisherInternational Conference on Learning Representations, ICLR-
dc.subject.meshActivation functions-
dc.subject.meshInput-output-
dc.subject.meshLow bound-
dc.subject.meshMin-max-
dc.subject.meshp-Approximation-
dc.subject.meshP-function-
dc.subject.meshUniform approximation-
dc.subject.meshUniversal approximation-
dc.subject.meshUniversal approximation properties-
dc.subject.meshUniversal approximators-
dc.titleMINIMUM WIDTH FOR UNIVERSAL APPROXIMATION USING RELU NETWORKS ON COMPACT DOMAIN-
dc.typeConference-
dc.citation.conferenceDate2024.5.7. ~ 2024.5.11.-
dc.citation.conferenceName12th International Conference on Learning Representations, ICLR 2024-
dc.citation.edition12th International Conference on Learning Representations, ICLR 2024-
dc.citation.title12th International Conference on Learning Representations, ICLR 2024-
dc.identifier.bibliographicCitation12th International Conference on Learning Representations, ICLR 2024-
dc.identifier.scopusid2-s2.0-85195527806-
dc.type.otherConference Paper-
dc.subject.subareaLanguage and Linguistics-
dc.subject.subareaComputer Science Applications-
dc.subject.subareaEducation-
dc.subject.subareaLinguistics and Language-
Show simple item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Min, Chan Ho Image
Min, Chan Ho민찬호
Department of Financial Engineering
Read More

Total Views & Downloads

File Download

  • There are no files associated with this item.