Show simple item record

dc.contributor.authorSelitskiy, Stanislav
dc.contributor.authorInoue, Chihiro
dc.contributor.authorSchetinin, Vitaly
dc.contributor.authorJakaite, Livija
dc.date.accessioned2024-02-02T10:24:05Z
dc.date.available2024-02-02T10:24:05Z
dc.date.issued2024-01-02
dc.identifier.citationSelitskiy S, Inoue C, Schetinin V, Jakaite L (2024) 'The batch primary components transformer and auto-plasticity learning linear units architecture: synthetic image generation case', 2023 Tenth International Conference on Social Networks Analysis, Management and Security (SNAMS) - Abu Dhabi, IEEE.en_US
dc.identifier.doi10.1109/SNAMS60348.2023.10375471
dc.identifier.urihttp://hdl.handle.net/10547/626147
dc.description.abstractContext tokenizing, which is popular in Large Language and Foundation Models (LLM, FM), leads to their excessive dimensionality inflation. Traditional Transformer models strive to reduce intractable excessive dimensionality at the among-token attention level, while we propose additional between-dimensions attention mechanism for dimensionality reduction. A novel Transformer-based architecture is presented, which aims at the individual dimension attention and, by doing so, performs the implicit relevant primary components' feature selection in artificial neural networks (ANN). As an additional mechanism allowing adaptive plasticity learning in ANN, a neuron-specific Learning Rectified Linear Unit layer is proposed for further feature selection via weight decay. The performance of the presented layers is tested on the encoder-decoder architecture applied for the synthetic image generation task for the benchmark MNIST data set.en_US
dc.language.isoenen_US
dc.publisherIEEEen_US
dc.relation.urlhttps://ieeexplore.ieee.org/document/10375471en_US
dc.subjecttransformeren_US
dc.subjectfeature selectionen_US
dc.subjectcosine distanceen_US
dc.subjectANN plasticityen_US
dc.subjectcatastrophic forgettingen_US
dc.subjectlearning ReLUen_US
dc.titleThe batch primary components transformer and auto-plasticity learning linear units architecture: synthetic image generation caseen_US
dc.typeConference papers, meetings and proceedingsen_US
dc.contributor.departmentUniversity of Bedfordshireen_US
dc.date.updated2024-02-02T10:21:19Z
dc.description.note


This item appears in the following Collection(s)

Show simple item record