Digital Repository

“CompressO” - Compression of Transformers using Pruning

Show simple item record

dc.contributor.author Weerasinghe, Chathuri
dc.date.accessioned 2022-12-16T09:36:37Z
dc.date.available 2022-12-16T09:36:37Z
dc.date.issued 2022
dc.identifier.citation Weerasinghe, Chathuri (2022) “CompressO” - Compression of Transformers using Pruning. BEng. Dissertation, Informatics Institute of Technology en_US
dc.identifier.issn 2017085
dc.identifier.uri http://dlib.iit.ac.lk/xmlui/handle/123456789/1142
dc.description.abstract "Providing meaningful representations of words has been considered one of the fundamental goals of Natural Language Processing (NLP) since its inception. With the advancement of technology, researchers have experimented with different approaches to achieve this goal. The introduction of word embeddings can be considered a significant improvement in this field of study which provides a vector representation of words. Models based on the Transformer architecture have reached state-of-the-art performance for numerous NLP tasks. However, due to the high dimensionality of these models making use of them has become computationally intensive. The focus of this research is to address the above issue by reducing the model size by compressing the Transformers architecture. Among various techniques used for Transformer model compression, Global Unstructured Pruning has been used in this project. " en_US
dc.language.iso en en_US
dc.subject Transformer Model Compression en_US
dc.subject Transformer architecture-based models en_US
dc.subject Neural network compression en_US
dc.subject Global Unstructured Pruning for Transformers en_US
dc.title “CompressO” - Compression of Transformers using Pruning en_US
dc.type Thesis en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search


Advanced Search

Browse

My Account