Loop and distillation: Attention weights fusion transformer for fine‐grained representation
Learning subtle discriminative feature representation plays a significant role in Fine-Grained Visual Categorisation (FGVC). The vision transformer (ViT) achieves promising performance in the traditional image classification filed due to its multi-head self-attention mechanism. Unfortunately, ViT ca...
Saved in:
Main Authors: | Sun, Fayou, Ngo, Hea Choon, Zuqiang, Meng, Sek, Yong Wee |
---|---|
Format: | Article |
Language: | English |
Published: |
John Wiley & Sons Ltd
2023
|
Online Access: | http://eprints.utem.edu.my/id/eprint/27758/2/0130221062024102412871.pdf http://eprints.utem.edu.my/id/eprint/27758/ https://ietresearch.onlinelibrary.wiley.com/doi/full/10.1049/cvi2.12181 https://doi.org/10.1049/cvi2.12181 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Adopting attention and cross-layer features for fine-grained representation
by: Sun, Fayou, et al.
Published: (2022) -
Adopting multiple vision transformer layers for fine-grained image representation
by: Sun, Fayou, et al.
Published: (2023) -
Clustering swap prediction for image-text pre-training
by: Fayou, Sun, et al.
Published: (2024) -
Improved attentive pairwise interaction (API-Net) for fine-grained image classification
by: Yet, Ong Zu, et al.
Published: (2021) -
Closed Loop Identification of Distillation Column
by: A/L P.Vellen, Nithianantham
Published: (2016)