SlideShare a Scribd company logo
2
Most read
3
Most read
8
Most read
Comparative Analysis of Transformer based
Pre-Trained NLP Models
Authors : Saurav Singla & Ramachandra N
Abstract:
❖ Transformer based self-supervised pre-trained models have transformed the concept of Transfer learning
in Natural language processing (NLP) using Deep learning approach.
❖ In this project we analyze the performance of self-supervised models for Multi-class Sentiment analysis
on a Non benchmarking dataset.
❖ We used BERT, RoBERTa, & ALBERT models for this study.
❖ We fine-tuned these models on Sentiment analysis with a proposed architecture.
❖ We used f1-score & AUC (Area under ROC curve) score for evaluating model performance.
❖ We found the BERT model with proposed architecture performed well with the highest f1-score of 0.85
followed by RoBERTa (f1-score=0.80), & ALBERT (f1-score=0.78).
❖ This analysis reveals that the BERT model with proposed architecture is best for multi-class sentiment
task on a Non-benchmarking dataset.
Related work:
❖ A concise overview on several large pre-trained language models provided with state-of-the-art results
on benchmark datasets viz. GLUE, RACE, & SQuAD[5].
❖ Cristóbal et al. presented a benchmark comparison of various deep learning architectures such as
Convolutional Neural Networks (CNN) , Long short-term memory (LSTM) recurrent neural networks
and BERT with a Bi-LSTM for the sentiment analysis of drug reviews[6].
❖ Horsuwan et al. systematically compared four modern language models: ULMFiT, ELMo with biLSTM,
OpenAI GPT, and BERT across different dimensions including speed of pretraining and fine-tuning,
perplexity, downstream classification benchmarks, and performance in limited pre training data on Thai
Social Text Categorization[7].
❖ Carlos Aspillaga et al. have performed Stress test evaluation of Transformer based models (RoBERTa,
XLNet, & BERT) in Natural Language Inference (NLI) & Question Answering (QA) tasks with
adversarial-examples[8].
Methodology:
In this section we will explain briefly about the model architecture & dataset used in this task.
Dataset
❖ We used Covid19 tweets dataset, publicly available on Kaggle[10].
❖ The train dataset contains 41157 tweets & test dataset contains 3798 tweets.
❖ There are 5 classes in the sentiment variable viz. Extremely Negative(0), Extremely positive(1),
Negative(2), Neutral(3), & Positive(4).
We used the Pytorch framework for building deep learning models with the help of Hugging face transformers.
Methodology:
Model Architecture
We have proposed architectures for BERT, RoBERTa, & ALBERT models for this study.
BERT
❖ It is a bidirectional transformer, meaning that it uses both left & right contexts in all layers as in Fig 1.
❖ This stands for Bidirectional Encoding Representations from Transformers.
❖ The Fig 2 shows the BERT input representations which includes Token, Segment, & Position
embeddings.
❖ In practice, input embeddings also contain input/attention masks used to differentiate between actual
tokens & padded tokens.
Methodology: BERT
Fig 1. BERT architecture Fig 2. BERT input representation
❖ In our task, we fine tuned the BERT model on preprocessed tweets data using a dropout layer, a hidden
layer, a fully connected layer & a softmax layer for classification on top of BERT embeddings which is
shown in Fig 3.
❖ We have considered bert-base uncased pre-trained model for this task, which has 12 layers, 768 hidden
size, 110 M parameters.
Methodology: Proposed architectures
Fig 3. BERT Fig 4. RoBERTa Fig 5. ALBERT
Methodology: RoBERTa
❖ It is Robustly optimized BERT pre-training approach. This replicates the BERT model by tweaking
hyper parameter settings & increasing training data size.
❖ For this task, we have fine tuned the model on preprocessed tweets data using a dropout layer, a hidden
layer, a fully connected layer & a softmax layer on top of RoBERTa embeddings as shown in Fig 4.
❖ We have chosen distil roberta base pre-trained model, which has 6 layers, 768 hidden size,12 heads, 82
M parameters.
ALBERT
❖ A Light BERT introduced to overcome TPU/GPU limitations & longer training times.
❖ We have fine tuned this model on preprocessed tweets data using a dropout, a fully connected layer &
finally a softmax on top of ALBERT embeddings which is shown in Fig 5.
❖ We have selected albert-base-v2 pre-trained model for this task, which has 12 layers, 768 hidden size,
11 M parameters.
Results & Discussions:
Table 1 shows the Sentiment analysis results for all models & corresponding hyperparameters.
Table 1. Comparison between models
We have kept a constant learning rate (lr) of 2e-5 & sentiment length (Sent len) of 120 for all models by
varying batch size & drop out.
Model f1-score lr dropout batch Sent length
BERT 0.85 2e-5 0.35 8 120
RoBERT
a
0.80 2e-5 0.32 32 120
ALBERT 0.78 2e-5 0.35 8 120
Results & Discussions: BERT
Fig 6. Precision-Recall curve Fig 7. ROC curve
We got the best results for BERT at batch size of 8 & drop out of 0.35.
Results & Discussions: RoBERTa
Fig 8. Precision-Recall curve Fig 9. ROC curve
We achieved best results at batch size of 32 & drop out of 0.32.
Results & Discussions: ALBERT
Fig 8. Precision-Recall curve Fig 9. ROC curve
We got good model performance at batch size of 8 & drop out of 0.35.
Conclusions & Future work:
❖ In this paper, we have fine tuned Transformer based pre-trained models viz., BERT, RoBERTa, &
ALBERT with proposed method for Multiclass Sentiment analysis task on Covid19 tweets dataset.
❖ We obtained the best results for BERT with a high training time (batch size=8). RoBERTa model
achieves acceptable results with less training time (batch size=32). We got reasonable results for
ALBERT with high training time (batch size=8).
❖ From the accuracy point of view the BERT model is the best for Multiclass Sentiment classification on
our dataset following the RoBERTa & ALBERT model. If speed is the main consideration, we
recommend using RoBERTa due to its speed of pretraining and fine-tuning with acceptable results.
❖ This study was conducted at specific batch size & drop out for 5 epochs. So model performance may be
different beyond 5 epochs & for different batch size & drop out.
❖ This work can be carried out in future to investigate how these models perform for different batch sizes &
drop out values.
❖ This work would help to choose the best pre-trained models for Sentiment analysis based on accuracy &
speed.
References:
1. Ashish et al, “Attention is all you need”, 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA pp.5998-
6008, 2017.
2. Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, “BERT: Pre-training of deep bidirectional transformers for language
understanding”, In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational
Linguistics: Human Language Technologies, , Volume 1 (Long and Short Papers), Minneapolis, Minnesota, pp. 4171-4186.
3. Yinhan Liu et al, “Roberta: A robustly optimized bert pre training approaches”, 2019, arXiv preprint arXiv:1907.11692.
4. Zhenzhong Lan et al, “Albert: A lite bert for self-supervised learning of language representations”, 2019, arXiv preprint
arXiv:1909.11942.
5. Matthias Aßenmacher, Christian Heumann, “On the comparability of pre-trained language models”, CEUR Workshop Proceedings,
Vol.2624.
6. Cristóbal Colón-Ruiz, Isabel Segura-Bedmar, "Comparing deep learning architectures for sentiment analysis on drug reviews", Journal
of Biomedical Informatics, Volume 110, 2020, 103539, ISSN 1532-0464.
7. Thanapapas Horsuwan, Kasidis Kanwatchara, Peerapon Vateekul, Boonserm Kijsirikul, "A Comparative Study of Pretrained Language
Models on Thai Social Text Categorization", 2019, arXiv:1912.01580v1.
8. Carlos Aspillaga, Andres Carvallo, Vladimir Araujo,"Stress Test Evaluation of Transformer-based Models in Natural Language
Understanding Tasks", Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020), European Language
Resources Association (ELRA), Marseille , pp. 1882–1894, 2020.
9. Vishal Shirsat, Rajkumar Jagdale, Kanchan Shende, Sachin N. Deshmukh, Sunil Kawale, “Sentence Level Sentiment Analysis from
News Articles and Blogs using Machine Learning Techniques”, International Journal of Computer Sciences and Engineering, Vol.7,
Issue.5, 2019.
10. Avinash Kumar1 , Savita Sharma , Dinesh Singh, "Sentiment Analysis on Twitter Data using a Hybrid Approach", International Journal
of Computer Sciences and Engineering, Vol.-7, Issue-5, May 2019.
11. Dataset: https://www.kaggle.com/datatattle/covid-19-nlp-text-classification

More Related Content

PPTX
PPTX
BERT introduction
PDF
BERT Finetuning Webinar Presentation
PDF
Transformer Introduction (Seminar Material)
PPTX
Natural language processing and transformer models
PDF
An introduction to the Transformers architecture and BERT
PPTX
Introduction to Transformer Model
PDF
Deep learning for NLP and Transformer
BERT introduction
BERT Finetuning Webinar Presentation
Transformer Introduction (Seminar Material)
Natural language processing and transformer models
An introduction to the Transformers architecture and BERT
Introduction to Transformer Model
Deep learning for NLP and Transformer

What's hot (20)

PPTX
Thomas Wolf "Transfer learning in NLP"
PPTX
NLP State of the Art | BERT
PDF
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
PDF
Introduction to Transformers for NLP - Olga Petrova
PDF
Tutorial on Deep Generative Models
PDF
NLP using transformers
PPTX
INTRODUCTION TO NLP, RNN, LSTM, GRU
PPTX
Transformer Zoo
PPTX
PPTX
Bert.pptx
PPTX
Transformers AI PPT.pptx
PPTX
Presentation on Sentiment Analysis
PPTX
Fine-tuning BERT for Question Answering
PDF
BERT: Bidirectional Encoder Representations from Transformers
PPTX
social network analysis project twitter sentimental analysis
PDF
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
PPTX
Sentiment Analysis using Twitter Data
PPTX
Aspect Based Sentiment Analysis
PDF
Deep Learning for NLP (without Magic) - Richard Socher and Christopher Manning
Thomas Wolf "Transfer learning in NLP"
NLP State of the Art | BERT
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
Introduction to Transformers for NLP - Olga Petrova
Tutorial on Deep Generative Models
NLP using transformers
INTRODUCTION TO NLP, RNN, LSTM, GRU
Transformer Zoo
Bert.pptx
Transformers AI PPT.pptx
Presentation on Sentiment Analysis
Fine-tuning BERT for Question Answering
BERT: Bidirectional Encoder Representations from Transformers
social network analysis project twitter sentimental analysis
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Sentiment Analysis using Twitter Data
Aspect Based Sentiment Analysis
Deep Learning for NLP (without Magic) - Richard Socher and Christopher Manning
Ad

Similar to Comparative Analysis of Transformer Based Pre-Trained NLP Models (20)

PDF
Analysis of the evolution of advanced transformer-based language models: Expe...
PPTX
"Sentiment Analysis of Customer Reviews Using ALBERT Model to Understand and ...
PDF
Transforming Customer Experience through NLP and Sentiment Analysis
PDF
Transforming Customer Experience through NLP and Sentiment Analysis
PPTX
Kaggle Tweet Sentiment Extraction: 1st place solution
PPTX
Cumulative_BERT_Review_Presentation.pptx
PPT
SENTIMENT ANALYSIS ON SOCIAL MEDIA
PDF
A Intensified Approach On Enhanced Transformer Based Models Using Natural Lan...
PDF
BERT - Part 1 Learning Notes of Senthil Kumar
PPTX
[Paper review] BERT
PPTX
Sentiment Classification SVM&LogistiqueRegression.pptx
PDF
PDF
Transformer Seq2Sqe Models: Concepts, Trends & Limitations (DLI)
PDF
Portofolio Muhammad Afrizal Septiansyah 2024
PDF
Nlp research presentation
PPTX
BERT QnA System for Airplane Flight Manual
PDF
IMPROVED SENTIMENT ANALYSIS USING A CUSTOMIZED DISTILBERT NLP CONFIGURATION
PDF
2201.00598.pdf
PDF
Contextual Emotion Recognition Using Transformer-Based Models
PPTX
Predicting Tweet Sentiment
Analysis of the evolution of advanced transformer-based language models: Expe...
"Sentiment Analysis of Customer Reviews Using ALBERT Model to Understand and ...
Transforming Customer Experience through NLP and Sentiment Analysis
Transforming Customer Experience through NLP and Sentiment Analysis
Kaggle Tweet Sentiment Extraction: 1st place solution
Cumulative_BERT_Review_Presentation.pptx
SENTIMENT ANALYSIS ON SOCIAL MEDIA
A Intensified Approach On Enhanced Transformer Based Models Using Natural Lan...
BERT - Part 1 Learning Notes of Senthil Kumar
[Paper review] BERT
Sentiment Classification SVM&LogistiqueRegression.pptx
Transformer Seq2Sqe Models: Concepts, Trends & Limitations (DLI)
Portofolio Muhammad Afrizal Septiansyah 2024
Nlp research presentation
BERT QnA System for Airplane Flight Manual
IMPROVED SENTIMENT ANALYSIS USING A CUSTOMIZED DISTILBERT NLP CONFIGURATION
2201.00598.pdf
Contextual Emotion Recognition Using Transformer-Based Models
Predicting Tweet Sentiment
Ad

Recently uploaded (20)

PPTX
Database Infoormation System (DBIS).pptx
PPTX
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
PPT
ISS -ESG Data flows What is ESG and HowHow
PDF
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
PPT
Quality review (1)_presentation of this 21
PPTX
Qualitative Qantitative and Mixed Methods.pptx
PPTX
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
PPTX
Introduction-to-Cloud-ComputingFinal.pptx
PDF
.pdf is not working space design for the following data for the following dat...
PPTX
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
PPTX
IBA_Chapter_11_Slides_Final_Accessible.pptx
PPTX
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
PPTX
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PPTX
Acceptance and paychological effects of mandatory extra coach I classes.pptx
PPTX
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
PPTX
Business Ppt On Nestle.pptx huunnnhhgfvu
PPTX
STUDY DESIGN details- Lt Col Maksud (21).pptx
PDF
Foundation of Data Science unit number two notes
PDF
Mega Projects Data Mega Projects Data
Database Infoormation System (DBIS).pptx
The THESIS FINAL-DEFENSE-PRESENTATION.pptx
ISS -ESG Data flows What is ESG and HowHow
“Getting Started with Data Analytics Using R – Concepts, Tools & Case Studies”
Quality review (1)_presentation of this 21
Qualitative Qantitative and Mixed Methods.pptx
Microsoft-Fabric-Unifying-Analytics-for-the-Modern-Enterprise Solution.pptx
Introduction-to-Cloud-ComputingFinal.pptx
.pdf is not working space design for the following data for the following dat...
advance b rammar.pptxfdgdfgdfsgdfgsdgfdfgdfgsdfgdfgdfg
IBA_Chapter_11_Slides_Final_Accessible.pptx
DISORDERS OF THE LIVER, GALLBLADDER AND PANCREASE (1).pptx
AI Strategy room jwfjksfksfjsjsjsjsjfsjfsj
Data_Analytics_and_PowerBI_Presentation.pptx
Acceptance and paychological effects of mandatory extra coach I classes.pptx
ALIMENTARY AND BILIARY CONDITIONS 3-1.pptx
Business Ppt On Nestle.pptx huunnnhhgfvu
STUDY DESIGN details- Lt Col Maksud (21).pptx
Foundation of Data Science unit number two notes
Mega Projects Data Mega Projects Data

Comparative Analysis of Transformer Based Pre-Trained NLP Models

  • 1. Comparative Analysis of Transformer based Pre-Trained NLP Models Authors : Saurav Singla & Ramachandra N
  • 2. Abstract: ❖ Transformer based self-supervised pre-trained models have transformed the concept of Transfer learning in Natural language processing (NLP) using Deep learning approach. ❖ In this project we analyze the performance of self-supervised models for Multi-class Sentiment analysis on a Non benchmarking dataset. ❖ We used BERT, RoBERTa, & ALBERT models for this study. ❖ We fine-tuned these models on Sentiment analysis with a proposed architecture. ❖ We used f1-score & AUC (Area under ROC curve) score for evaluating model performance. ❖ We found the BERT model with proposed architecture performed well with the highest f1-score of 0.85 followed by RoBERTa (f1-score=0.80), & ALBERT (f1-score=0.78). ❖ This analysis reveals that the BERT model with proposed architecture is best for multi-class sentiment task on a Non-benchmarking dataset.
  • 3. Related work: ❖ A concise overview on several large pre-trained language models provided with state-of-the-art results on benchmark datasets viz. GLUE, RACE, & SQuAD[5]. ❖ Cristóbal et al. presented a benchmark comparison of various deep learning architectures such as Convolutional Neural Networks (CNN) , Long short-term memory (LSTM) recurrent neural networks and BERT with a Bi-LSTM for the sentiment analysis of drug reviews[6]. ❖ Horsuwan et al. systematically compared four modern language models: ULMFiT, ELMo with biLSTM, OpenAI GPT, and BERT across different dimensions including speed of pretraining and fine-tuning, perplexity, downstream classification benchmarks, and performance in limited pre training data on Thai Social Text Categorization[7]. ❖ Carlos Aspillaga et al. have performed Stress test evaluation of Transformer based models (RoBERTa, XLNet, & BERT) in Natural Language Inference (NLI) & Question Answering (QA) tasks with adversarial-examples[8].
  • 4. Methodology: In this section we will explain briefly about the model architecture & dataset used in this task. Dataset ❖ We used Covid19 tweets dataset, publicly available on Kaggle[10]. ❖ The train dataset contains 41157 tweets & test dataset contains 3798 tweets. ❖ There are 5 classes in the sentiment variable viz. Extremely Negative(0), Extremely positive(1), Negative(2), Neutral(3), & Positive(4). We used the Pytorch framework for building deep learning models with the help of Hugging face transformers.
  • 5. Methodology: Model Architecture We have proposed architectures for BERT, RoBERTa, & ALBERT models for this study. BERT ❖ It is a bidirectional transformer, meaning that it uses both left & right contexts in all layers as in Fig 1. ❖ This stands for Bidirectional Encoding Representations from Transformers. ❖ The Fig 2 shows the BERT input representations which includes Token, Segment, & Position embeddings. ❖ In practice, input embeddings also contain input/attention masks used to differentiate between actual tokens & padded tokens.
  • 6. Methodology: BERT Fig 1. BERT architecture Fig 2. BERT input representation ❖ In our task, we fine tuned the BERT model on preprocessed tweets data using a dropout layer, a hidden layer, a fully connected layer & a softmax layer for classification on top of BERT embeddings which is shown in Fig 3. ❖ We have considered bert-base uncased pre-trained model for this task, which has 12 layers, 768 hidden size, 110 M parameters.
  • 7. Methodology: Proposed architectures Fig 3. BERT Fig 4. RoBERTa Fig 5. ALBERT
  • 8. Methodology: RoBERTa ❖ It is Robustly optimized BERT pre-training approach. This replicates the BERT model by tweaking hyper parameter settings & increasing training data size. ❖ For this task, we have fine tuned the model on preprocessed tweets data using a dropout layer, a hidden layer, a fully connected layer & a softmax layer on top of RoBERTa embeddings as shown in Fig 4. ❖ We have chosen distil roberta base pre-trained model, which has 6 layers, 768 hidden size,12 heads, 82 M parameters. ALBERT ❖ A Light BERT introduced to overcome TPU/GPU limitations & longer training times. ❖ We have fine tuned this model on preprocessed tweets data using a dropout, a fully connected layer & finally a softmax on top of ALBERT embeddings which is shown in Fig 5. ❖ We have selected albert-base-v2 pre-trained model for this task, which has 12 layers, 768 hidden size, 11 M parameters.
  • 9. Results & Discussions: Table 1 shows the Sentiment analysis results for all models & corresponding hyperparameters. Table 1. Comparison between models We have kept a constant learning rate (lr) of 2e-5 & sentiment length (Sent len) of 120 for all models by varying batch size & drop out. Model f1-score lr dropout batch Sent length BERT 0.85 2e-5 0.35 8 120 RoBERT a 0.80 2e-5 0.32 32 120 ALBERT 0.78 2e-5 0.35 8 120
  • 10. Results & Discussions: BERT Fig 6. Precision-Recall curve Fig 7. ROC curve We got the best results for BERT at batch size of 8 & drop out of 0.35.
  • 11. Results & Discussions: RoBERTa Fig 8. Precision-Recall curve Fig 9. ROC curve We achieved best results at batch size of 32 & drop out of 0.32.
  • 12. Results & Discussions: ALBERT Fig 8. Precision-Recall curve Fig 9. ROC curve We got good model performance at batch size of 8 & drop out of 0.35.
  • 13. Conclusions & Future work: ❖ In this paper, we have fine tuned Transformer based pre-trained models viz., BERT, RoBERTa, & ALBERT with proposed method for Multiclass Sentiment analysis task on Covid19 tweets dataset. ❖ We obtained the best results for BERT with a high training time (batch size=8). RoBERTa model achieves acceptable results with less training time (batch size=32). We got reasonable results for ALBERT with high training time (batch size=8). ❖ From the accuracy point of view the BERT model is the best for Multiclass Sentiment classification on our dataset following the RoBERTa & ALBERT model. If speed is the main consideration, we recommend using RoBERTa due to its speed of pretraining and fine-tuning with acceptable results. ❖ This study was conducted at specific batch size & drop out for 5 epochs. So model performance may be different beyond 5 epochs & for different batch size & drop out. ❖ This work can be carried out in future to investigate how these models perform for different batch sizes & drop out values. ❖ This work would help to choose the best pre-trained models for Sentiment analysis based on accuracy & speed.
  • 14. References: 1. Ashish et al, “Attention is all you need”, 31st Conference on Neural Information Processing Systems, Long Beach, CA, USA pp.5998- 6008, 2017. 2. Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding”, In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, , Volume 1 (Long and Short Papers), Minneapolis, Minnesota, pp. 4171-4186. 3. Yinhan Liu et al, “Roberta: A robustly optimized bert pre training approaches”, 2019, arXiv preprint arXiv:1907.11692. 4. Zhenzhong Lan et al, “Albert: A lite bert for self-supervised learning of language representations”, 2019, arXiv preprint arXiv:1909.11942. 5. Matthias Aßenmacher, Christian Heumann, “On the comparability of pre-trained language models”, CEUR Workshop Proceedings, Vol.2624. 6. Cristóbal Colón-Ruiz, Isabel Segura-Bedmar, "Comparing deep learning architectures for sentiment analysis on drug reviews", Journal of Biomedical Informatics, Volume 110, 2020, 103539, ISSN 1532-0464. 7. Thanapapas Horsuwan, Kasidis Kanwatchara, Peerapon Vateekul, Boonserm Kijsirikul, "A Comparative Study of Pretrained Language Models on Thai Social Text Categorization", 2019, arXiv:1912.01580v1. 8. Carlos Aspillaga, Andres Carvallo, Vladimir Araujo,"Stress Test Evaluation of Transformer-based Models in Natural Language Understanding Tasks", Proceedings of the 12th Conference on Language Resources and Evaluation (LREC 2020), European Language Resources Association (ELRA), Marseille , pp. 1882–1894, 2020. 9. Vishal Shirsat, Rajkumar Jagdale, Kanchan Shende, Sachin N. Deshmukh, Sunil Kawale, “Sentence Level Sentiment Analysis from News Articles and Blogs using Machine Learning Techniques”, International Journal of Computer Sciences and Engineering, Vol.7, Issue.5, 2019. 10. Avinash Kumar1 , Savita Sharma , Dinesh Singh, "Sentiment Analysis on Twitter Data using a Hybrid Approach", International Journal of Computer Sciences and Engineering, Vol.-7, Issue-5, May 2019. 11. Dataset: https://www.kaggle.com/datatattle/covid-19-nlp-text-classification