| dc.contributor.author | Dilshan, D.P | |
| dc.date.accessioned | 2022-03-11T06:56:40Z | |
| dc.date.available | 2022-03-11T06:56:40Z | |
| dc.date.issued | 2021 | |
| dc.identifier.citation | Dilshan, D.P (2021) PRUNELM: Compressing Bi-LSTM Architecture Based Pre-Trained Contextualized Word Embeddings . BSc. Dissertation Informatics Institute of Technology | en_US |
| dc.identifier.issn | 2017050 | |
| dc.identifier.uri | http://dlib.iit.ac.lk/xmlui/handle/123456789/915 | |
| dc.description.abstract | " Deep Learning has gained popularity in recent years due to its high accuracy in fields such as pattern recognition and computer vision. As a result, almost every task in Natural Language Processing, such as sentiment analysis, named entity recognition, language translation, and text classification, now uses deep learning architectures and methods to achieve their goals. There’s a special layer called the word embedding layer when it comes to Natural Language Processing with Deep Learning. The purpose of the word embedding layer is to give a vectorize representation to a word so that computers can do calculations on words as the human brain does. Due to highly dimensionality of these embedding layers, they require high memory requirements, hence these models are unable to use with memory constraint devices. To address this issue various research have been conducted on word embedding compression. After going through these research and identifying some valid gaps, PruneLM is focused to compress Bi LSTM based contextualized word embeddings. Unlike other compression systems, PruneLM provides a novel approach to compression. Pruning has been used as the compression method in the PruneLM. PruneLM provides two different compressions for a given model. A given Bi-LSTM based model can be compressed with or without retraining with the PruneLM and with the provided statistics dashboards, users can compare and view the performance of the compressed models." | en_US |
| dc.language.iso | en | en_US |
| dc.subject | Bi-LSTM Language Model | en_US |
| dc.subject | Model Pruning | en_US |
| dc.subject | Word Embedding Compression | en_US |
| dc.subject | Bi-LSTM based contextualized Word Embedding | en_US |
| dc.subject | Language Model Compression | en_US |
| dc.title | PRUNELM: Compressing Bi-LSTM Architecture Based Pre-Trained Contextualized Word Embeddings | en_US |
| dc.type | Thesis | en_US |