dc.contributor.author |
Mahbub, Ridwan |
|
dc.contributor.author |
Anuva, Samiha Shafiq |
|
dc.contributor.author |
Khan, Ifrad Towhid |
|
dc.date.accessioned |
2024-01-18T06:36:55Z |
|
dc.date.available |
2024-01-18T06:36:55Z |
|
dc.date.issued |
2023-05-30 |
|
dc.identifier.uri |
http://hdl.handle.net/123456789/2059 |
|
dc.description |
Supervised by
Professor Dr. Md. Hasanul Kabir,
Co-Supervisors:
Mr. Shahriar Ivan, Lecturer,
Mr. Md. Zahidul Islam,
Department of Computer Science and Engineering(CSE),
Islamic University of Technology(IUT),
Board Bazar, Gazipur-1704, Bangladesh |
en_US |
dc.description.abstract |
To improve the representational power of convolutional neural networks, several
attention mechanisms have been introduced in recent years. These attention mecha-
nisms are calculated on input feature maps by enhancing some parts of the input data
and diminishing other parts of the input data as all parts of the input do not contain
important features for training. One exception can be seen where weights are used in
place of input feature maps and this approach is known as weight excitation. Since
the weights of a CNN get fine-tuned based on the input data, calculating attention
on weights can be an alternative to calculating attention on input feature maps. One
advantage of this method is that this doesn’t introduce any additional computational
cost at inference time. In this work, we explore different mechanism of weight ex-
citation on different types of architectures. We have conducted several experiments
to conclude whether weights can be used as an alternative to input feature maps for
computing attention and if this applies to all existing attention mechanisms for Con-
volutional Neural Networks. We also test other properties of weight excitation, like
the regularizing effect of weight excitation. |
en_US |
dc.language.iso |
en |
en_US |
dc.publisher |
Department of Computer Science and Engineering(CSE), Islamic University of Technology(IUT), Board Bazar, Gazipur-1704, Bangladesh |
en_US |
dc.title |
Improving Weight Excitation for ConvNets and MLP Mixer |
en_US |
dc.type |
Thesis |
en_US |