1 Introduction

The coronavirus, also known as COVID-19, has generated a significant public health issue in the past two years. Coronavirus affected the lives of millions of people. Millions of infections and deaths have been reported since March 2020 [27]. The World Health Organization (WHO) reported 37,109,851 and 1,070,355 confirmed COVID-19 and death cases, respectively [43]. Due to the severity and hazards of COVID-19, the WHO issued emergency guidelines for all medical and public health system on 28 February 2020 [11]. Many people expressed their opinions about COVID-19 on social media such as YouTube, Twitter, and Facebook. The posted messages become the discussion topic among friends and family members with positive or negative response [25]. Social media information affects the life of people positively or negatively [13]. YouTube, Twitter, and Facebook are significant sources of ”social data”. Sentiment analysis plays a vital role to understand the human emotions by analyzing the human’s behaviour [22]. Nowadays, various authors applied the different techniques for the sentiment analysis and classification on social media text. However, a substantial quantity of data about COVID-19 is also available on social media. Historical social data can be utilized by the researchers to give better judgment and conclusion about coronavirus [32]. In this work, Coronavirus-related Hindi tweets posted between 15 March 2020 to 24 May 2021, are collected for sentiment analysis and classification. Many authors have shown the advantages of machine learning techniques for feature extraction and sentiment classification [3, 4, 14, 37]. The combination of lexicon and deep learning techniques can be used to analyze the human sentiments [2].

Motivation

The sentiment analysis of Hindi tweets about COVID-19 is done by few authors. In this work, sentiment analysis is performed to understand the Indian sentiments about COVID-19 using Hindi tweets. The various techniques of natural language processing (NLP) with meta-heuristic optimization and a hybrid deep learning model are utilized to analyze the public opinion.

1.1 Contribution of this work

The main contribution of present work is given as:

  • A Hindi COVID-19 annotated dataset is developed from Twitter data for sentiment analysis.

  • A unique Hindi stop word dictionary is prepared and applied for the experimental analysis.

  • The data pre-processing is performed by utilizing various techniques of NLP prior to the construction of labelled feature vectors.

  • The Grey wolf optimization algorithm is employed for selection of optimal features.

  • Finally, An ensemble deep learning model of convolution neural network (CNN) and long short-term memory (LSTM) is applied for sentiment classification as positive, neutral, or negative.

Organization of the paper

The remaining part of the paper is organized as: Section 2 discusses the existing work done by various authors. The proposed methodology is presented in Section 3. The outcome of the proposed model is discussed in Section 4, followed by conclusions in Section 5.

2 Literature review

Many authors investigated the sentiment analysis about COVID-19 tweets recently.

Singh et al. proposed a framework named as Covhindia for sentiment analysis [40]. It detects the emotional polarity of COVID-19 Hindi tweets posted between 23 March and 15 July 2020. The Covhindia framework achieved 88.9% accuracy. Chintalapud et al. investigated 3,090 Indian COVID-19 tweets during the lockdown period [12]. Authors have utilized the Bidirectional Encoder Representations from Transformers (BERT) model for data analysis and achieved 89% accuracy.

The long short-term memory model has been used by Chandra et al. for sentiment analysis [9]. Total 150,000 Indian COVID-19 tweets from March 2019 to September 2020 are examined by the authors. Global Twitter patterns about COVID-19 are examined by Lwin et al. [28]. A lexical approach with the ‘Crystal-Feel’ algorithm is utilized for text analysis in their work and classified the sentiment into four categories such as joy, sorrow, anger, and fear. It is concluded that the impact of negative feelings dominated the public mental health. Raamkumar et al. analyzed the public attitudes and responses about COVID-19 through social media channels [35]. The public information distribution processes is enhanced by the authors. Zhao et al. analyzed public sentiments by collecting public attention about COVID-19 events at two-month intervals using Sina Microblog of China [44]. Feelings of Indian citizens during the lockdown period is investigated by Barkur et al. [6]. They analyzed both negative and positive sentiments by using 2,400 tweets. Friendly and disputed Twitter keywords of COVID-19 is evaluated by Chen et al. [10]. A dictionary-based language tool is deployed for sentiment analysis in their work. They achieved highest F1-score of 0.9521% by using XLNet on 500,000 sample data. Jelodar et al. presented the LSTM model for sentiment classification of COVID-19 micro-blogs [21]. Kaur et al. investigated public attitudes towards the COVID- 19 by applying NLP and machine learning techniques on 2,058 tweets [24]. They obtained 24.0%, 32.1%, and 43.9% positive, negative, and neutral tweets, respectively. COVID-19 topic modelling and sentiment analysis is done by Prabhakar et al. by utilizing 18,000 tweets with National Research Council (NRC) sentiment lexicon [33]. Nemes et al. utilized a Recurrent Neural Network model to determine the emotional content of tweets as positive or negative [31]. Samuel et al. used Twitter data to discover public opinion on COVID-19 using Naïve Bayes classifier, Logistic regression, and linear regression techniques and obtained highest accuracy of 74% [38].

Gupta et al. evaluated different machine learning techniques, namely, logistic regression, Naive Bayes approach, Support Vector Machine(SVM), and Decision Tree, for sentiment analysis of Hindi tweets [19]. They have used the National Research Council (NRC) Emotion and Hindi Senti-WordNet Lexicon to identify the emotion of phrases. Finally, an integrated convolutional neural network (CNN) is presented for sentiment analysis of 23,767 Hindi tweets as positive, negative, or neutral with 85% accuracy. A subjective lexicon with a graph-based approach has been developed by Arora et al. for Hindi reviews classification and achieved the highest 74% accuracy [5]. Mittal et al. [30] generated an annotated corpus for the Hindi language and achieved 80% classification accuracy by using HindiSentiWordNet (HSWN). Nevertheless, most of the research work has been done by analyzing social media networks. In the recent study, Basile et al. examined the dramatic occurrence effect of social media post [7]. Rigorous research is conducted by the authors on the Reddit social network based wide range of topics and languages.

3 Framework of proposed methodology

The proposed methodology presents sentiment analysis and classification of COVID-19 Hindi tweets composed of three steps: (1) Data collection of Hindi tweets related to COVID-19, (ii) Data pre-processing and feature extraction using NLP followed by GWO based feature selection, and (iii) Sentiment classification as positive, negative, or neutral. The flow diagram of the proposed framework is presented in Fig. 1. Step wise summary of the proposed work is given in Algorithm 1. A detailed description of each step is discussed in the next subsections.

Fig. 1
figure 1

Overview of the proposed framework

Algorithm 1
figure a

Hindi text sentiment analysis of COVID -19.

3.1 Data collection

Hindi tweets about COVID-19 is used in the present work for sentiment analysis. All tweets are collected from Twitter by using Python twint library with the different search keyword namely, #coronavirus, #indiafightscorona, #stayhome, #staysafe, #coronavirusindia, #lockdownindia, #coronaindia, #coronavirusinindia, #COVIDindia, and #COVID-19 as shown in Fig. 2. The Python is used for implementation of scraping script. Total 9,527,810 COVID-19 Hindi tweets are collected from Twitter of the 15 March 2020 to 24 May 2021. The collected tweets contains unstructured, unlabelled, informal, and noisy text. The collected dataset also contains a large number of text features with enormous density. The sample Hindi tweets used for experiment is shown in Fig. 3.

Fig. 2
figure 2

Keywords applied for tweets crawling

Fig. 3
figure 3

Contents of raw dataset

3.2 Data pre-processing

The pre-processing step is used for (i) data cleaning, transformation, and normalization, (ii) feature extraction, and (iii) feature selection. The pre-processing step is summarized diagrammatically in Fig. 4. This step reduces the memory requirements and accelerates the next data processing task. Each pre-processing steps are discussed in the next subsections.

Fig. 4
figure 4

Pre-processing steps

Data cleaning

Unwanted data is removed from the collected tweets data in the cleaning process. Regular expression is used in this step which includes (a) Removal of unnecessary words, (b) Elimination of HTML tags, (c) Elimination of emojis and data numbering patterns, (d) Removal of additional characters in sentences, and (e) Elimination of spaces, emails, URLs, and punctuation.

Removal of stop word

First, a unique dictionary of 576 Hindi stop words is generated as shown in Fig. 5. All stop words are eliminated from the cleaned data.

Fig. 5
figure 5

List of Hindi stop words

Tokenization

The tokenization process partitioned the large chunks of text into single piece of word known as token. The indic NLP library is used for the tokenization process [26].

Data labelling

Data labelling process is followed by data cleaning for labelling of unlabelled data. The HindiSentiWordNet (HSWN) is used for data labelling and sentiment extraction associated with dataset [23]. This step generates vocabulary which contains Hindi sentimental expressions and their corresponding positive, neutral, and negative polarity.

Vectorization of sentences

Vectorization is the final step of pre-processing which maps the pre-processed data into a vector of real numbers. Vectorization is done by indexing technique and mathematically expressed as:

$$ S_{m}=\left[W_{m, 1}, W_{m, 2}, \ldots, W_{m, n}\right] $$
(1)

here, Sm, m represents the mth sentences and order of the mth sentence, respectively. Wi,1, Wi,2, . . . , Wi, n denotes the weighting vector. The vector length is represented by n. The generated weight matrix is represented mathematically as:

$$ weight\_matrix=\left[\begin{array}{llll} W_{(1,1)} & W_{(1,2)} & {\cdots} & W_{(1, n)} \\ W_{(2,1)} & W_{(2,2)} & {\cdots} & W_{(2, n)} \\ {\vdots} & {\vdots} & {\cdots} & {\vdots} \\ W_{(m, 1)} & W_{(m, 2)} & {\cdots} & W_{(m, n)}\\ \end{array}\right]\\ $$
(2)

The weight matrix is given as input for the feature selection.

3.3 Grey wolf optimization

Meta-heuristic-based Grey wolf optimization technique (GWO) is used to select the optimal features from the pre-processed data [29]. This technique is used in this work as (i) It is simple and user-friendly and (ii) The convergence rate is faster than other optimization techniques such as Salp swarm algorithm, Firefly, and Harris Hawks. The Updated_weight_matrix is given as initial population to GWO algorithm which generates optimized_weight_matrix. The steps of the GWO algorithm is summarized in Algorithm 2.

Algorithm 2
figure b

Algorithm of GWO technique.

Grey wolves represent the Canidae family members and contain strong class structures [15]. Grey wolf always lives in a pack which is a group of five to twelve members [18]. Prey hunting is the most common activity among a pack of wolves. The GWO hierarchy initially covers four ranks of wolves denoted as alpha(α), beta(β), delta(δ), and omega(ω) as shown in Fig. 6.

Fig. 6
figure 6

Social hierarchy of wolves

The α is most powerful wolf among all four groups which takes decisions for various activities such as hunting, discipline, sleep, and get-up time. The second powerful wolf group β supports the most powerful group α for decision-making activities. The delta group occupies third place in the wolf social hierarchy. Lastly, ω group ensures security and competency for all wolf packs [1, 39].

3.3.1 Fitness function

The mean absolute difference (MAD) is employed as the fitness function in GWO technique to determine the word relevancy. The fitness function evaluates the word relevancy based on its weight. The weight_matrix is given as input to the fitness function. It calculates the weight of each word based on its fitness value, and returns the updated weight matrix as output. A relevance score is assigned by the fitness function to each text by comparing their mean value by using a mathematical equation expressed as:

$$ MAD(Xh_{m})=\frac{1}{Xh_{m}}{\sum}_{p=1}^{n}\mid r_{m},p-\overline{r_{b}}\mid $$
(3)
$$ \overline{r_{b}}=\frac{1}{X h_{m}} {\sum}_{P=1}^{n} r_{m, P} $$
(4)

Here, Xhm denotes the number of text features taken from the sentences X, rm represents the mean value of vector m, the weighting value of feature P is denoted by \(\overline {r_{b}}\). n denotes the Total number of text features.

3.4 Mathematical model of Grey wolf algorithm

Gray wolf algorithm incorporates four quantitative methods, namely, (i) Hierarchical structure of Grey wolves, (ii) Prey encircling, (iii) Prey hunting, and (iv) Prey searching and attacking. Each method is briefly discussed in the following subsection.

3.4.1 Hierarchy structure of Grey wolves algorithm

The Grey wolf algorithm focuses on the quantitative hierarchy of the wolf pack leadership. Alpha (α) represents the highest level of intellectual hierarchy, whereas (β) denotes the second and third most essential traits, respectively [20].

3.4.2 Prey encircling

All grey wolves hunt their prey by encircling process which demonstrates the wolves movement in the surroundings. This quantum mechanics can be represented mathematically as [8]:

$$ H=\mid G \cdot B_{U}(y)-B(y)\mid $$
(5)
$$ B(y+1)=B_{U}(y)-E \cdot H $$
(6)

here, y represents the current iteration, BU(y) and B(y) denotes the position vector of prey and wolf, respectively. H denotes the distance between prey and wolf. E and G represent random vectors which are defined as:

$$ E=2 d \cdot u_{1}-d $$
(7)
$$ G=2 \cdot u_{2} $$
(8)

here, u1and u2 denotes random vectors between 0 and 1. These vectors determine the closeness of wolves and prey.

3.4.3 Prey hunting

The whole hunting procedure is lead by Alpha wolf group. The grey wolves follows alpha, beta, and delta groups in search of an ideal hunting spot [17, 41]. The whole hunting search process is mathematically represented as:

$$ H_{\alpha}=\mid G_{1} \cdot B_{\alpha}-B\mid $$
(9)
$$ H_{\beta}=\mid G_{2} \cdot B_{\beta}-B\mid $$
(10)
$$ H_{\delta}=\mid G_{3} \cdot B_{\delta}-B \mid $$
(11)
$$ B_{1}=B_{\alpha}-E_{1} \cdot H_{\alpha} $$
(12)
$$ B_{2}=B_{\beta}-E_{2} \cdot H_{\beta}, $$
(13)
$$ B_{3}=B_{\delta}-E_{3} \cdot H_{\delta} $$
(14)

Updated position of the grey wolf is represented by the following equation:

$$ B(y+1)=\left( B_{1}+B_{2}+B_{3}\right) / 3 $$
(15)

3.4.4 Prey searching and attacking

In this step, the grey wolves attack the victim and stop their movement. The random value of vector E differentiates the wolf from the prey. The value of vector E is given in the range of [− d,d] and the value of d is computed as:

$$ d = 2-\left( 2 * t / Q_{v}\right) $$
(16)

As a result, if E < 1, the wolf is obligated to exterminate the victim. If E > 1, the wolf distinguishes itself from the victim and seeks the fittest prey. The prey searching process is affected by the relative position of all three α, β, and δ wolves. The possible hunting location for a grey wolf is shown in Fig. 7.

Fig. 7
figure 7

Wolves hunting movement

The goal of entire process is to determine the calibration of both vectors E and G. The exploitation and exploration are emphasized in almost every dimension. Finally, the GWO algorithm returns an optimal weight matrix denoted as optimized_weight_matrix and given as input to the classification model.

3.5 Classification model

A hybrid deep learning model of CNN and LSTM is applied for sentiment classification as positive, negative, or neutral. The architecture of the hybrid model is depicted in Fig. 8. The brief description of the CNN and LSTM model is given as:

Fig. 8
figure 8

Architecture of hybrid proposed model

Convolution neural network model (CNN)

CNN model contains convolution layer which minimizes the dimension of input data [16, 34]. Total 50 filters of 3x3 window is used for feature extraction from input data. The extracted features are derived from a set of words by applying the mathematical equation expressed as:

$$ M=\textit{q}(\textit{T} \cdot h_{a}+\textit{r}) $$
(17)

Here, M denotes the total number of extracted features, ha and T denotes the set of words and the filter weights, respectively, r represents a biased factor. The non-linear activation function used by the convolutional layer is given by q.

Long short-term memory model(LSTM)

LSTM is a specific recurrent neural network (RNN) model. It uses three gates to maintain and control the long-term dependency along with the state information of each node. The model can also solve the vanishing gradient problem. The gates and cells of LSTM model is mathematically represented as:

$$ \text P u_{z}=\sigma\left( D_{a u} \cdot\left[d p_{r}-1\right], z_{r}+a_{a u}\right) $$
(18)
$$ \text E_{r}=\tanh \left( D_{E}\left[d p_{r}-1\right], z_{r}+a_{i}\right) $$
(19)
$$ \text q_{r}=\sigma\left( D_{q} \cdot\left[d_{p q}-1\right], z_{r}+a_{q}\right) $$
(20)
$$ \text q_{e}=\sigma\left( D_{e}\left[d_{p r}-1\right], z_{r}+a_{e}\right) $$
(21)

here, a denotes the bias vector, D and zr represents input weight and input vector at time r, respectively. The term Pu, qe, Er, and q represents the input gate, output gate, cell memory, and forget, respectively.

Hybrid model

Hybrid model is constructed by the hybridization of both CNN and LSTM models. It combines the functionality of convolutional network with the LSTM. The architecture of the hybrid model is shown in Fig. 8. First, embedded words are fed into the CNN-LSTM model through a convolution layer. Output of CNN model is given as input to the LSTM layer for further processing.

Dropout

Dropout is an essential trick of deep learning model to prevent over-fitting [36]. It skips the non-participating neurons of the back-propagation process. Dropout strategy removes the neurons to avoid co-adaptation during the training of the model.

Embedding layer

This layer embeds the pre-processed dataset with a unique ID and provides a meaningful sequence of words. Index-based weight matrix is employed for word embedding. This layer also assigned a random weight to each word of the training dataset.

Convolution layer

The embedding layer transmits sentences to the convolutional layer which uses a pooling layer for convolution with the given input. The pooling layer controls network over-fitting by reducing the input phrases representation.

Global max-pooling

Finally, global maximum pooling (max-pooling) is implemented to obtain the best global results.

Activation function

The rectified linear unit (RELU) is utilized as an activation function. It returns the output based on the value of input neurons. The mathematical expression of RELU function is given as:

$$ R(W)=\left\{\begin{array}{cc} W & W>0 \\ 0 & W<=0 \end{array}\right\} $$
(22)

here, W denotes the input neuron.

Dense layer

This layer is also referred to as a linked layer in which every neuron is connected with every subsequent neuron. This layer is used to classify input features obtained from the convolution layers.

Soft-Max

The Soft-Max function is utilized as the last layer of the hybrid neural network. It is mathematically expressed as:

$$ \operatorname{soft-max}\left( S_{p}\right)=\frac{\exp \left( S_{p}\right)}{{\sum}_{q} \exp \left( S_{q}\right)} $$
(23)

here, S represents the output produced by the last layer. The exponential function serves as the non-linear function. The output is normalized by the sum of exponential values and converted into probabilities.

4 Experimental results and discussions

The proposed algorithm is implemented by using Python 3.7 with an i8 E-2236 processor, 32 GB RAM, and NVIDIA P2200 display card.

4.1 Preprocessing results

Total 9,527,810 COVID-19 Hindi tweets from 15 March 2020 to 24 May 2021 are collected. Total 21,914 sarcastic, non-Hindi, and non-subjective tweets are removed from the original dataset. The remaining 9,505,896 tweets is given as input to the proposed framework for further processing. Sample of original text data and extracted tokens are given as:

Sample of original text data

figure c

Sample example of extracted tokens

figure d

4.2 Results of data labelling and sentiment analysis

A sample of labelled text is shown in Fig. 9. The distribution of COVID-19 Hindi tweets based on outcome of sentiment analysis is shown in Table 1. Total 4,763,905, 3,811,124, and 930,867 tweets are analyzed as negative, positive, and neutral, respectively.

Fig. 9
figure 9

Labeled data

Table 1 Result of sentiment analysis

4.3 Classification results and discussion

A deep learning-based hybrid model is used for sentiment classification as positive, negative, and neutral. Three different hybrid models namely, (a) CNN-LSTM, (b) LSTM–CNN, and (c) CNN-LSTM + LSTM–CNN are applied for the classification. Performance of the proposed hybrid model is evaluated in terms of accuracy, precision, recall, and F-score. Definition of evaluation parameters are given as:

$$ Accuracy=\frac{\sum (G\_D, G\_Z) }{\sum (G\_D , P\_D ,G\_Z , P\_Z)} $$
(24)

here, G_D, G_Z, P_D, P_Z denote the properly identified, wrongly identified, rejected in the proper manner, and wrongly rejected, respectively. Precision defines the number of data tagged as positive by machine learning models.

$$ Precision=\frac{ (G\_D) }{\sum (G\_D , P\_D )} $$
(25)

Recall is also referred to as sensitivity or positive predictive value. Mathematically it is defined as:

$$ Recall=\frac{ (G\_D) }{\sum (G\_Z , P\_Z)} $$
(26)

F-Scores is defined the harmonic mean of precision and recall and denoted as [42]:

$$ F-score=\frac{ 2* (Precision* Recall) }{\sum (Precision,Recall)} $$
(27)

The various parameter values set as input for all three hybrid models are presented in Table 2.

Table 2 Parameter values set as input for all three hybrid models

All classification models are fine-tuned to obtain the best value of each parameter. Vector dimension of the embedding layer is set as 300. Total 64 neurons are used for the hidden layer of the LSTM-CNN+CNN-LSTM model. The learning rate is fixed as 0.001 with the Adam optimizer. A dropout layer is added with a value of 0.5 between the hidden and output layer of both LSTM and CNN models. A fully linked layer of 32 neurons is used to avoid over-fitting. The outcome of CNN and LSTM models are aggregated with distinct hyper-parameters values (Figs. 10 and 11).

Fig. 10
figure 10

Comparison of the proposed ensemble hybrid model with CNN, LSTM, CNN-LSTM, and LSTM-CNN

Fig. 11
figure 11

Comparison of the proposed hybrid deep learning model with standard machine and deep learning model in terms of accuracy

The performance of all three hybrid models are compared with traditional machine learning and deep learning models namely, Random Forest(RF), Decision Tree, K nearest neighbor (KNN), Naive Bayes (NB), SVM, LSTM, and CNN.

4.4 Comparison of proposed work with existing work

Performance of the proposed model is compared with the existing machine and deep learning models as shown in Tables 3 and 4. The highest accuracy of 89.11%, 85.54%, 87.89%, 88.41%, 87.75%, 91.46%, 88.72%, 82.34%, 91.54%, and 95.54% is achieved with SVM, NB, KNN, DT, RF, CNN, LSTM, CNN–LSTM, LSTM–CNN, and ensemble hybrid model (LSTM-CNN+CNN-LSTM), respectively. It is observed from the result that the performance of the ensemble hybrid model (LSTM-CNN+CNN-LSTM) is better than traditional machine learning models.

Table 3 Comparison of the proposed model with traditional machine and deep learning models
Table 4 Comparison of proposed model with existing work

Convergence plot and boxplot of GWO are compared with three other optimization techniques namely, Salp swarm algorithm, Firefly, and Harris Hawks, as shown in Figs. 12 and 13, respectively. The X and Y axis of the convergence plot represent the number of iterations and fitness values, respectively. It is analyzed from the convergence plot that the suggested technique converges faster than the Salp swarm algorithm, Firefly, and Harris Hawks optimization technique.

Fig. 12
figure 12

Comparison of the convergence rates of the GWO and Salp swarm algorithm, Firefly, and Harris Hawks optimization algorithm

Fig. 13
figure 13

Comparison of GWO with Salp swarm algorithm, Firefly, and Harris Hawks optimization algorithm

The X-axis and Y-axis of the boxplot represent the applied techniques and fitness value, respectively. The boxplot shows that the proposed methodology gives a better result for all parameter values.

4.5 Analysis of computational complexity

The computational complexity of the entire prediction process is analyzed as:

Data-cleaning process requires a total of O(Twitter posts × total word count). Feature extraction step requires T(f) = O(f2) + parsing time. Here, f is the total number of tweets in the sample. Computational complexities of GWO algorithm for feature selection is summarized as:

(i) GWO initialization requires O(W × q) time, here W and q denote population density and problem dimension, respectively. (ii) Calculation of GWO process parameters require O(W × q). (iii) O(W × q) time is required to update the wolf position. (iv) Evaluation of fitness value requires O(W × q) time. The entire time complexity of the GWO is denoted as O(W × q × MI) and here, MI denotes the optimum iterations. The computational complexity of supervised learning algorithms depends on the number of iterations or the number of classification classes. The time complexity of CNN-LSTM model is given as O(b × s(ac + cx + xy)), here, a, c, x, and s denote the number of input layer nodes, second layer node, third layer node, and training examples, respectively. by and y denotes the total number of epochs and output layer nodes, respectively. The prediction step requires a temporal complexity of O(1). The overall time complexity is O(f2). Time complexity of each stage is shown in Table 5.

Table 5 Computational complexity of proposed model

5 Conclusions

The meta-heuristic based Grey wolf optimization technique and hybrid deep learning model is presented in this work. Data preprocessing and labelling using natural language processing plays a important step in the proposed model to construct a feature vector. The optimal features are selected by applying the Grey wolf optimization technique. Finally, the sentiment classification has been done by applying a hybrid model of CNN-LSTM and LSTM-CNN. In addition, results of the proposed model are compared with traditional machine learning models also. Proposed model gives highest 95.54%, 91.44%, 89.63%, and 90.87% of classification accuracy, precision, recall, and F-score, respectively. The experimental results show that the proposed model is more effective than other machine learning models for the sentiment classification of Hindi tweets. This study shows that people are optimistic about COVID-19. Positive tweets suggest that individuals are enthusiastic about COVID-19, while negative tweets indicate that users are scared by its effects. The neutral tweets extracted from the dataset demonstrated that many peoples are confused about the impact of COVID-19. The findings of this study will be beneficial for the government, policymakers, and healthcare management to understand the effects of COVID-19 on society. The findings of this work can be utilized to build more effective strategies for boosting public attitude during the various phases of pandemics. COVID-19 vaccine sentiment analysis will be performed on the Hindi dataset in future work. This will assist the policymakers to address the people concerns prior to mass vaccination. Multi-model framework for sentiment analysis will also be developed by integrating text messages and voice tones about COVID-19 issues.