Please use this identifier to cite or link to this item: https://elib.vku.udn.vn/handle/123456789/4031
Full metadata record
DC FieldValueLanguage
dc.contributor.authorPham, Minh Tuan-
dc.contributor.authorNguyen, An Hung-
dc.date.accessioned2024-07-31T01:59:03Z-
dc.date.available2024-07-31T01:59:03Z-
dc.date.issued2024-07-
dc.identifier.isbn978-604-80-9774-5-
dc.identifier.urihttps://elib.vku.udn.vn/handle/123456789/4031-
dc.descriptionProceedings of the 13th International Conference on Information Technology and Its Applications (CITA 2024); pp: 174-185vi_VN
dc.description.abstractAbstract. Vision Transformer is a novel approach in artificial intelligence, focusing on image classification. Despite its potential, ViT's emphasis on global data processing presents accuracy challenges compared to local data processing methods like Convolutional Neural Networks (CNN). To address this, we propose two methods. The first integrates a portion of the Residual Network to replace token transformation layers, allowing for local data feature extraction and improved relationship learning between tokens. The second solution suggests transforming layers in the bottleneck component into types that process in the Quaternion hypercomplex domain, enhancing the multidimensional representation of data. Both solutions aim to leverage the strengths of CNN and ViT, thereby indirectly improving image classification accuracy.vi_VN
dc.language.isoenvi_VN
dc.publisherVietnam-Korea University of Information and Communication Technologyvi_VN
dc.relation.ispartofseriesCITA;-
dc.subjectImage classificationvi_VN
dc.subjectDeep learningvi_VN
dc.subjectVision Transformervi_VN
dc.subjectQuaternion Algebravi_VN
dc.subjectMultilayer Perceptron Algebravi_VN
dc.titleEnhancing Image Classification Capabilities in the Vision Transformer Network Model with Quaternion Algebravi_VN
dc.typeWorking Papervi_VN
Appears in Collections:CITA 2024 (Proceeding - Vol 2)

Files in This Item:

 Sign in to read



Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.