SCOPUS
0Citation Export
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kim, Namjun | - |
dc.contributor.author | Min, Chanho | - |
dc.contributor.author | Park, Sejun | - |
dc.date.issued | 2024-01-01 | - |
dc.identifier.uri | https://aurora.ajou.ac.kr/handle/2018.oak/37103 | - |
dc.identifier.uri | https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85195527806&origin=inward | - |
dc.description.abstract | It has been shown that deep neural networks of a large enough width are universal approximators but they are not if the width is too small. There were several attempts to characterize the minimum width wmin enabling the universal approximation property; however, only a few of them found the exact values. In this work, we show that the minimum width for Lp approximation of Lp functions from [0, 1]dx to ℝdy is exactly max{dx, dy, 2} if an activation function is RELU-LIKE (e.g., RELU, GELU, SOFTPLUS). Compared to the known result for RELU networks, wmin = max{dx + 1, dy} when the domain is ℝdx, our result first shows that approximation on a compact domain requires smaller width than on Rdx. We next prove a lower bound on wmin for uniform approximation using general activation functions including RELU: wmin ≥ dy + 1 if dx < dy ≤ 2dx. Together with our first result, this shows a dichotomy between Lp and uniform approximations for general activation functions and input/output dimensions. | - |
dc.description.sponsorship | NK and SP were supported by Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2019-0-00079, Artificial Intelligence Graduate School Program, Korea University) and Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (2022R1F1A1076180). | - |
dc.language.iso | eng | - |
dc.publisher | International Conference on Learning Representations, ICLR | - |
dc.subject.mesh | Activation functions | - |
dc.subject.mesh | Input-output | - |
dc.subject.mesh | Low bound | - |
dc.subject.mesh | Min-max | - |
dc.subject.mesh | p-Approximation | - |
dc.subject.mesh | P-function | - |
dc.subject.mesh | Uniform approximation | - |
dc.subject.mesh | Universal approximation | - |
dc.subject.mesh | Universal approximation properties | - |
dc.subject.mesh | Universal approximators | - |
dc.title | MINIMUM WIDTH FOR UNIVERSAL APPROXIMATION USING RELU NETWORKS ON COMPACT DOMAIN | - |
dc.type | Conference | - |
dc.citation.conferenceDate | 2024.5.7. ~ 2024.5.11. | - |
dc.citation.conferenceName | 12th International Conference on Learning Representations, ICLR 2024 | - |
dc.citation.edition | 12th International Conference on Learning Representations, ICLR 2024 | - |
dc.citation.title | 12th International Conference on Learning Representations, ICLR 2024 | - |
dc.identifier.bibliographicCitation | 12th International Conference on Learning Representations, ICLR 2024 | - |
dc.identifier.scopusid | 2-s2.0-85195527806 | - |
dc.type.other | Conference Paper | - |
dc.subject.subarea | Language and Linguistics | - |
dc.subject.subarea | Computer Science Applications | - |
dc.subject.subarea | Education | - |
dc.subject.subarea | Linguistics and Language | - |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.