Ajou University repository

Gramian Attention Heads are Strong yet Efficient Vision Learnersoa mark
Citations

SCOPUS

3

Citation Export

DC Field Value Language
dc.contributor.authorRyu, Jongbin-
dc.contributor.authorHan, Dongyoon-
dc.contributor.authorLim, Jongwoo-
dc.date.issued2023-01-01-
dc.identifier.issn1550-5499-
dc.identifier.urihttps://aurora.ajou.ac.kr/handle/2018.oak/36949-
dc.identifier.urihttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85188239686&origin=inward-
dc.description.abstractWe introduce a novel architecture design that enhances expressiveness by incorporating multiple head classifiers (i.e., classification heads) instead of relying on channel expansion or additional building blocks. Our approach employs attention-based aggregation, utilizing pairwise feature similarity to enhance multiple lightweight heads with minimal resource overhead. We compute the Gramian matrices to reinforce class tokens in an attention layer for each head. This enables the heads to learn more discriminative representations, enhancing their aggregation capabilities. Furthermore, we propose a learning algorithm that encourages heads to complement each other by reducing correlation for aggregation. Our models eventually surpass state-of-the-art CNNs and ViTs regarding the accuracy-throughput trade-off on ImageNet-1K and deliver remarkable performance across various downstream tasks, such as COCO object instance segmentation, ADE20k semantic segmentation, and fine-grained visual classification datasets. The effectiveness of our framework is substantiated by practical experimental results and further underpinned by generalization error bound. We release the code publicly at: https://github.com/Lab-LVM/imagenet-models.-
dc.language.isoeng-
dc.publisherInstitute of Electrical and Electronics Engineers Inc.-
dc.subject.meshArchitecture designs-
dc.subject.meshBuilding blockes-
dc.subject.meshChannel expansions-
dc.subject.meshGramians-
dc.subject.meshLearn+-
dc.subject.meshmatrix-
dc.subject.meshNovel architecture-
dc.subject.meshPerformance-
dc.subject.meshState of the art-
dc.subject.meshTrade off-
dc.titleGramian Attention Heads are Strong yet Efficient Vision Learners-
dc.typeConference-
dc.citation.conferenceDate2023.10.2. ~ 2023.10.6.-
dc.citation.conferenceName2023 IEEE/CVF International Conference on Computer Vision, ICCV 2023-
dc.citation.editionProceedings - 2023 IEEE/CVF International Conference on Computer Vision, ICCV 2023-
dc.citation.endPage5828-
dc.citation.startPage5818-
dc.citation.titleProceedings of the IEEE International Conference on Computer Vision-
dc.identifier.bibliographicCitationProceedings of the IEEE International Conference on Computer Vision, pp.5818-5828-
dc.identifier.doi10.1109/iccv51070.2023.00537-
dc.identifier.scopusid2-s2.0-85188239686-
dc.identifier.urlhttp://ieeexplore.ieee.org/xpl/conhome.jsp?punumber=1000149-
dc.type.otherConference Paper-
dc.description.isoatrue-
dc.subject.subareaSoftware-
dc.subject.subareaComputer Vision and Pattern Recognition-
Show simple item record

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Ryu, Jongbin Image
Ryu, Jongbin유종빈
Department of Software and Computer Engineering
Read More

Total Views & Downloads

File Download

  • There are no files associated with this item.