The Longest Month: Analyzing COVID-19 Vaccination Opinions Dynamics From Tweets in the Month Following the First Vaccine Announcement

The coronavirus outbreak has brought unprecedented measures, which forced the authorities to make decisions related to the instauration of lockdowns in the areas most hit by the pandemic. Social media has been an important support for people while passing through this difficult period. On November 9, 2020, when the first vaccine with more than 90% effective rate has been announced, the social media has reacted and people worldwide have started to express their feelings related to the vaccination, which was no longer a hypothesis but closer, each day, to become a reality. The present paper aims to analyze the dynamics of the opinions regarding COVID-19 vaccination by considering the one-month period following the first vaccine announcement, until the first vaccination took place in UK, in which the civil society has manifested a higher interest regarding the vaccination process. Classical machine learning and deep learning algorithms have been compared to select the best performing classifier. 2 349 659 tweets have been collected, analyzed, and put in connection with the events reported by the media. Based on the analysis, it can be observed that most of the tweets have a neutral stance, while the number of in favor tweets overpasses the number of against tweets. As for the news, it has been observed that the occurrence of tweets follows the trend of the events. Even more, the proposed approach can be used for a longer monitoring campaign that can help the governments to create appropriate means of communication and to evaluate them in order to provide clear and adequate information to the general public, which could increase the public trust in a vaccination campaign.


I. INTRODUCTION
The coronavirus outbreak caused by the novel coronavirus SARS-CoV-2 has brought a series of changes in many aspects of people's economic and social life. Since its occurrence, the coronavirus pandemic has continued to monopolize the different parts of the world, reaching 220 countries and territories by December 9, 2020 [1]. Governments have tried to address the outbreak by considering a series of measures, not all of them in accordance with the general public opinion.
The associate editor coordinating the review of this manuscript and approving it for publication was Derek Abbott .
In all this time, the rapid growth of the number of cases globally has produced panic, fear and anxiety among people [2].
Due to the current situation generated by the lockdown in some parts of the world and social distancing in others, the use of social media globally has intensified [2], as it succeeds in connecting people from geographically different places and allows them to exchange ideas and information related to a series of aspects that have occurred in this period. Even more, people seem to rely on the information posted on social media. As a result, social media platforms have become mediator channels between each individual and the rest of the world and have gained more and more attention, being VOLUME 9, 2021 This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/ one of the fastest growing information systems for social applications [3], [4]. On this channel, individuals show their different views, opinions and emotions during the various events that occur due to the coronavirus pandemic [3]. Among some of the well-known social media platforms, Twitter has gained a particular attention as the users can easily broadcast information about their opinions on a given topic through a public message, called tweet [5]. Besides the information voluntarily offered by the user, a tweet may also retain information related to the location of the user and might contain links, emoticons and hashtags which can help the user in better expressing his/her sentiments, making it a source of valuable information [5], [6]. Even more, Twitter has been used by government officials and political figures for informing the general public either regarding their activity or in the case of major events occurrence [7].
The vaccination topic has been, over time, one of the themes which have raised a series of questions in social media, most of them related to the safety of the entire process. As a result, a series of studies have analyzed the impact of different social media campaigns on vaccination hesitancy [21]- [23] or the general public sentiment in connection with the vaccination process [5], [24]. Additionally, compared to other vaccination situations studied in the scientific literature, the COVID-19 vaccination comes with new inquietudes related to the relatively short period of time needed for the vaccine development. As known, the process of developing a vaccine typically takes a decade [25]. Note, however, that the fastest vaccine development before has been four years [26] in the case of mumps vaccine and that, almost forty years after the discovery of HIV, no effective vaccine has yet been developed. However, the vaccine timelines for COVID-19 are reduced due to the emergency [25]. On December 18, 2020, the web site COVID-19 Vaccine Tracker, 1 held by Milken Institute, shows 236 vaccines are in development, 38 are now in clinical testing and 7 have reached a regulatory decision. Nevertheless, on December 8, 2020 the first vaccine has been administrated in UK.
In this context, the present paper analyzes the public opinion related to the vaccination process in the case of COVID-19, by considering the messages posted on Twitter. The period between November 9, 2020 -when Pfizer and BioNTech announced the development of a vaccine that is more than 90% effective, to December 8, 2020 -when the vaccination process has started in UK, has been considered. A number of 2 349 659 tweets have been collected and a cleaned dataset containing 752 951 tweets has been extracted. The performance for stance detection of several machine learning algorithms (both classical machine learning and deep learning algorithms) has been compared on an annotated dataset. The best performing algorithm has been selected and used for analyzing both the entire and the cleaned datasets.
The contribution of the paper is three-folded: we have collected and annotated a COVID-19 vaccination dataset, we have determined the best performing classifier for COVID-19 vaccination stance detection and we have put in relation the number of tweets and the stance (e.g. in favor, against or neutral) with the events reported by the media in the analyzed period.
The chosen approach can be easily integrated in a system which can allow interested organizations a proper monitoring of the public opinion regarding the vaccination process in the case of the new coronavirus.
The remainder of the paper is organized as follows. Section 2 provides a literature review structured in two main parts: natural language processing -focusing on sentiment analysis and stance detection from social media messages, and recent studies analyzing public opinion based on COVID-19 data extracted from Twitter. Section 3 describes the proposed methodology, while Section 4 focuses on the dataset collection and annotation process. Section 5 describes the steps required for stance detection and analyzes the performance of the classification algorithms. Section 6 presents the dynamics of opinions in the analyzed period. The limitations of the present study are mentioned in Section 7. The paper closes with a conclusion section and references. A series of supplementary materials accompany the paper, in the form of the collected and annotated datasets, along with the extracted unigrams, bigrams and trigrams for each day in the selected period.

II. LITERATURE REVIEW
In the following, a short literature review regarding sentiment analysis and stance detection is conducted in order to underline the current approaches in the research literature. Afterwards, a series of studies that have analyzed the public opinion, in the context of the COVID-19 pandemic, using data extracted from Twitter are discussed.

A. SENTIMENT ANALYSIS AND STANCE DETECTION
Opinion mining is a growing area of the Natural Language Processing field commonly used to determine viewpoints towards targets of interest using computational methods [27]. It is also known as sentiment analysis and includes many sub-tasks, such as polarity detection -in which the goal is to determine whether a text has positive, negative or neutral connotation [28], emotion identification -in which the objective is to uncover specific emotions such as happiness, fear or sadness [29], subjectivity detection -in which the goal is to determine if the text is objective or subjective [30].
Stance detection [31], [32] is an opinion mining task used in debate analysis, for determining the opinions towards a specific target. It can be formalized as the task of identifying the tuple < t, s >, in which t represents the target entity, while s represents the opinion. The target entity (t) can be any discussion topic, including products, services, economic measures, or life choices, such as vaccination. The opinion (s) towards the target is identified as in favor, against or neutral [27].
While similar in some respects to polarity detection, stance detection is a different natural language processing task, given the fact that positive tweets can be against the target entity, while on the contrary, negative tweets can sometimes express a favorable view of the target entity. Moreover, when compared to polarity detection, stance detection always determines the agreement or disagreement in relation to a specific target, even in cases in which the target is not explicitly mentioned in the analyzed text [5].
The types of approaches that can be used for polarity analysis and stance detection include: lexicon-based methods [33], machine learning methods [34] and hybrid methods -in which lexicons and machine learning are combined [35], [36].
Lexicon based methods rely on sentiment lexicons, such as Bing Liu's opinion lexicon [37], MaxDiff [38], Senti-ment140 [39], VaderSentiment [40], SentiWordNet [41] or SenticNet [42], which contain words and sequences of words, together with the polarity score, indicating the strength of the positive, neutral or negative perception. For performing polarity detection, the sentiment lexicons are used together with semantic methods, which typically consider negations and booster words [40]. A simple rule-based model incorporating a sentiment lexicon, as well as grammatical and syntactical conventions, called Vader, is proposed by Hutto and Gilbert [40]. The authors show that the proposed model outperforms individual human raters. When compared to classical machine learning algorithms (such as Support Vector Machines, Naïve Bayes and Maximum Entropy), the authors show that Vader offers a better performance on the datasets collected from Twitter, Amazon reviews and NYT editorials. Given the fact that the creation of lexicons is time consuming, Cotfas et al. [33] have shown that multiple existing lexicons can be combined to create more comprehensive lexicons through the advantages brought by the grey systems theory. Compared to machine learning, lexicon-based approaches have the advantage of not requiring the collection and annotation of training data, making them preferable when the volume or the quality of the training data is not sufficient [43], [44].
Machine Learning approaches use supervised classification algorithms to extract knowledge regarding the sentiment polarity or the stance of a text. As a preliminary step, before applying machine learning, the text needs to be first converted into numerical vectors, using schemes such as Bag-of-Words and word embeddings. The Bag-of-Words approach is a flexible text representation scheme that describes the number of occurrences of words in the encoded document. As a disadvantage, this scheme does not consider the sequence in which the words appear in the document, thus ignoring the context in which they are used [45]. Word embeddings are a text representation approach in which each word is mapped to a vector having the values computed in such a way that allows words which frequently appear in similar contexts to have a similar representation [46]. The main benefit of this representation is that additional clues become available for the classification algorithms. Another advantage resides in the fact that the number of required dimensions is greatly reduced when compared to a sparse vector representation, such as one-hot encoding, in which each term is as a binary vector that contains only zeros, besides a single one-value, corresponding to the term's index in the vocabulary [45]. Among the most popular word embedding techniques, one can mention: embedding layer, Word2Vec [47], GloVe [48] and FastText [49].
Machine learning approaches include classical machine learning and deep learning algorithms. Frequently used classical machine learning algorithms for stance detection are Support Vector Machines (SVM) [5], [31], [50] and Naïve Bayes (NB) [5]. In the context of the ''SemEval-2016 Task 6: Detecting Stance in Tweets'' [51], the SVM classifier with unigram features, used as a baseline for the algorithms developed by the competing teams, has achieved and F-Score of 63.31. By incorporating also word n-grams (unigrams, bigrams and trigrams) and character n-grams (with lengths {2, 3, 4, 5}) the F-Score has increased to 68.98, higher than all the scores recorded by the algorithms proposed during the competition [51]. D'Andrea et al. [5] have compared several classical machine learning (including SVM and NB) and deep learning algorithms for detecting the stance towards vaccination in Italian tweets, achieving the best results when using SVM. The approach proposed by D'Andrea et al. [5] has constituted the basis for the current study.
Deep Learning algorithms have become particularly popular in recent years for both stance detection [31] and sentiment analysis [52]. The Deep Learning based techniques have predominantly used Convolutional Neural Networks (CNN) [5], [53] and Recurrent Neural Networks (RNN) [54], [55], with its variant Long Short-Term Memory (LSTM) [5], [56]- [58]. Zarrella and Marsh [58] have proposed a LSTM approach that has achieved an F-Score of 67.82, one of the highest scores among the competing teams at ''SemEval-2016 Task 6: Detecting Stance in Tweets''. However, the algorithm has performed worse than the baseline SVM n-grams algorithm.
As an alternative to RNN and CNN, Vaswani et al. [59] have proposed transformers, an attention-based architecture, replacing the recurrent layers with multi-headed selfattention, achieving state of the art results for machine translation [59], document generation [60] and syntactic parsing [61]. Transformer-based language models, pre-trained on large and diverse corpuses of unlabeled data, such as Generative Pre-trained Transformer (Open-AI GPT) [62] and Bidirectional Encoder Representations from Transformers (BERT) [63] can be afterwards easily fine-tuned for a wide range of Natural Language Processing (NLP) tasks [62], [63]. While Open-AI GPT uses a unidirectional left-to-right architecture, BERT relies on a bidirectional approach, providing better results on many NLP tasks, including sentiment analysis [63].
Hybrid methods feature a combination of lexicons and machine learning algorithms. Aloufi and Saddik [35] have performed polarity detection from football-specific tweets using several machine learning algorithms and a sentiment lexicon automatically generated starting from a manually labeled dataset. Even though some improvements have been noticed by the authors in comparison to using general lexicons, the best results have been achieved by SVM with unigrams.
Comparisons between various stance analysis approaches used in social media analysis are included in Wang et al. [31] and Mohammad et al. [51].

B. TWITTER SENTIMENT ANALYSIS ON COVID-19 DATA
In the case of epidemics, Merchant and Lurie [64] have observed that besides the role assumed by social media of becoming the fastest channel of communication between people found in situations of social distancing due to lockdown, the social media can also act as a tool which can be used for anticipating the circumstances related to the spread of epidemics around the world. The authors have observed a high correlation between the information posted on Twitter regarding the evolution of an epidemic and the official data released by the Center for Disease Control and Prevention. As a result, the authors have concluded that Twitter can provide real-time estimations and predictions in the case of epidemic-related activities. Based on this research, Kaur et al. [65] have used the data extracted from Twitter to monitor the dynamics of emotions during the first months after the COVID-19 has become known to the public. A total number of 16 138 tweets have been extracted and analyzed using IBM Watson Tome Analyzer. As expected, the number of negative tweets exceeded the number of neutral and positive tweets in all the three months considered in the paper. Comparing the sentiments extracted for June with the ones extracted for February, it has been observed that the proportion of negative sentiments has decreased (from 43.92% to 38.05%), while the positive sentiments proportion has increased (from 21.38% to 27.01%). The proportion of the neutral sentiments has been almost the same (34.07% in February vs. 34.94% in June).
The prevalence of negative sentiments over the positive ones in the case of the COVID-19 pandemic has been also underlined by Singh et al. [66], while Boon-Itt and Skunkan [67] have recorded a high discrepancy between the negative sentiments (covering 77.88% of tweets) and the positive sentiments (covering the rest of 22.12%).
Xue et al. [68] have analyzed the public sentiment related to 11 selected topics determined using Latent Dirichlet Allocation on COVID-19 tweets. The authors have concluded that fear is the most dominant emotion in all the considered topics and that the findings are in line with other studies on COVID-19 which state that human psychological conditions are significantly impacted by the coronavirus outbreak [68].
On the other hand, Bhat et al. [69] found that the most prominent sentiment was positive in the analysis conducted in their paper. The authors state that the occurrence of the positive sentiments in 51.97% of tweets can be a sign that the users who have posted the messages are hopeful and enjoy the socialization experience shared with the family in this period of lockdown and limited social interaction.
At regional level, Kruspe et al. [70] have analyzed geotagged tweets in Europe regarding COVID-19 through the use of a neural network, featuring a multilingual version of BERT, which has been trained on an external dataset, not connected to the COVID-19 outbreak. Based on their results, the authors state that they have observed a general downward trend of the negative sentiments as the time passes.
At national level, several studies have been conducted for different countries around the world. For example, in a study conducted on tweets extracted for Nepal, Pokharel [71] observed that the public opinion faced positive sentiments (58% of the tweets), while the negative sentiments have only been expressed in 15% of the tweets. The study used a Naïve Bayes model applied on a limited number of tweets (615 tweets). Barkur et al. [72] determined that in the case of the tweets from India, the positive sentiment was dominant when analyzing the national lockdown situation announced by the government. Similar conclusions have been reached by Khan et al. [73] in a research that has used Naïve Bayes classifier. The difference between the reactions towards the pandemic in different cultures has been studied by Imran et al. [74] through sentiment and emotion analysis, implemented with deep learning classifiers. Besides the correlation between tweets' polarity from different countries, the authors also state that NLP can be used to link the emotions expressed on social platforms to the actual events during the coronavirus pandemic. Samuel et al. [75] have shown insights related to the evolution of the fear-sentiment over time in the United States.
At regional level, Zhou et al. [76] analyzed the sentiments in local government areas located in Australia and found that the general sentiment during the COVID-19 pandemic was a positive one, but there have been observed decreases in the positive polarity as the pandemic advanced, with significant changes from positive to negative sentiments depending on the government policies or social events. Wang et al. [77] made a comparative analysis between the tweets posted in California and New York and concluded that California had more negative sentiments than New York and that the fluctuation in sentiment scores can be correlated with the severity of COVID-19 pandemic and policy changes. Pastor [78] analyzed the sentiment of the Filipinos located in Luzon area and concluded that most Filipinos had negative sentiments, most of them due to the extreme community quarantine.
Some other analyses on Twitter in the context of COVID-19 have focused, but have not been limited to: topical sentiment analysis regarding the use of masks [79], monitoring depression trends [80], sentiment dynamics related to cruise tourism [81], identifying discussion topics and emotions [82], thematic analysis [83], detecting misleading information [84].
As shown above, the prominent sentiments related to COVID-19 have been found to be either positive or negative. The expressed sentiments have been shown to depend on the geographic area, government decisions and number of recorded cases. A more in-depth analysis related to the studies on sentiment analysis featuring COVID-19 and other infectious diseases can be found in Alamoodi et al. [3].
In this context, the present paper aims to analyze the stance of the Twitter users in connection to the new upcoming vaccines for COVID-19 in the first month after Pfizer and BioNTech announced their results on the new vaccine. The methodology and data collection process are presented in the following sections.

III. METHODOLOGY
The steps taken in order to analyze the public's opinion regarding COVID-19 vaccination from social media messages are shown in Figure 1.
The initial step is to collect a COVID-19 vaccination stance dataset containing English language tweets. A randomly sampled subset from this dataset has been afterwards manually annotated as neutral, in favor, or against vaccination, in order to be used in the training phase of the stance classification algorithms.
Given their unstructured nature and informal writing style, in the following step, the tweets from the collected dataset have been pre-processed, with the purpose of improving the performance of the stance classification algorithms.
For text representation and classification, four approaches have been investigated: 1) Bag-of-Words representation followed by classical machine learning, 2) Word embeddings followed by classical machine learning, 3) Word embeddings followed by deep learning and 4) Bidirectional Encoder Representations from Transformers.
In order to determine the best performing classification algorithm, the text has been represented using both Bag-of-Words and word embeddings schemes. In the present paper, the performance of multiple classical machine learning and deep learning algorithms has been evaluated based on the following widely used metrics: Accuracy, Precision, Recall and F-score. Accuracy, which indicates the ratio of correctly predicted observations to the total observations is defined as shown in (1), in which TP, TN, FP and FN refer to true positive, true negative, false positive and false negative. Thus, TP represents the number of real positive tweets classified as positive, FP is the number of real negative tweets classified incorrectly classified as positives, TN represents the number of negative tweets correctly classified as negative and FN is the number of real positive tweets incorrectly classified as negative.
Precision, which represents the ratio of correctly predicted positive observations to the total predicted positive observations, is computed as shown in (2).
Recall, representing the ratio of correctly predicted positive observations to all the observations in the actual class, is computed as shown in (3).
Starting from Precision and Recall, the F-Score can be computed as a weighted average, as shown in (4).
Finally, the best performing algorithm has been used to analyze the evolution of the public stance towards vaccination in the considered period. The evolution has been correlated with the major events and news that have followed the announcement of the Pfizer and BioNTech vaccine results.
Steps of the proposed stance-detection approach. VOLUME 9, 2021

IV. COVID-19 VACCINE STANCE DATASET
A machine learning approach has been chosen for detecting the stance of the tweets, which requires a labeled dataset for training the classification models. Since we have not identified an already labeled dataset for stance towards COVID-19 vaccination in the scientific literature, a domain-specific dataset, having Twitter as a data source, has been collected and manually annotated. It should be also mentioned, that according to [31], there is a general lack of annotated corpuses for stance detection.

A. DATASET COLLECTION
Several public datasets including large-scale collections of tweets related to the coronavirus pandemic have been proposed in the scientific literature, including the ones presented in [85]- [88]. Some of the datasets, such as [86], [88], are multi-lingual, while others, such as [85], [87] are language specific, including only tweets written in English.
In order to collect a dataset centered around COVID-19 vaccination, a hybrid approach has been chosen, in which the tweets that we have fetched through the Twitter API for the keywords in Table 1, have been supplemented with the ones in the dataset described in [86], selected using the same keywords.
Gathering the tweets from the Twitter API has been performed through the Twitter Filtered Stream API, with the help of the TweetInvi 2 library.
While the approach proposed in this paper can be extended to other languages, in the present study only tweets written in English have been considered. Thus, between 2 https://github.com/linvi/tweetinvi

B. DATASET ANNOTATION
To ensure the quality of the annotated dataset, that will be used for training the machine learning algorithms, duplicated tweets have been discarded, as well as retweets. The retweets have been easily identified due to the presence of the ''RT'' symbol. This choice is in accordance with the approach from other studies, including, but not limited to [5], [35]. The remaining number of tweets in the cleaned dataset is 752 951, representing 32.04% of the initial dataset. Table 2 includes for each day in the considered period both the total number of tweets, as well as the remaining number of tweets after the duplicates and retweets have been eliminated.
From the cleaned dataset we have randomly selected and manually annotated 7530 tweets, representing approximately 1.00% of all the tweets in the dataset. The number is higher than the one used in other stance detection approaches, such as D'Andrea et al. [5] and Mohammad et al. [89]. D'Andrea et al. [5] have trained the algorithms on a manually labeled dataset containing 693 tweets. The dataset proposed by Mohammad et al. [89] is organized on several topics, with the largest topic numbering 984 tweets.
In the present approach, the stance of the tweets towards vaccination has been evaluated by three independent human raters into three classes: in favor, against and neutral. Disagreements between the annotated tweets have only been recorded between the in favor and neutral or between the neutral and against stances. No disagreement has been recorded between in favor and against annotations. In the case of disagreement, the class chosen by most annotators has been associated with the tweet.
The distribution of the tweets in the annotated dataset in the three considered categories is illustrated in Table 3.
Tweets that have been assigned to the class in favor express a positive opinion regarding the vaccination. Tweets belonging to the against vaccination class express a negative opinion towards COVID-19 vaccination. The neutral class  mainly includes news related to the development of vaccines, tweets that do not express a clear opinion, such as questions regarding the vaccine, informative tweets concerning vaccination, as well as off-topic tweets, many of them related to the 2020 presidential election in the United States, which was held nominally, just a few days before the analyzed period, on November 3, 2020. Several examples of manually labeled tweets belonging to the three categories are included in Table 4.

V. COVID-19 VACCINATION STANCE DETECTION
The main components of the stance detection process are the pre-processing, the feature extraction and the machine learning classification. In the pre-processing step the text is cleaned, while in the feature extraction the raw textual data is converted to feature vectors. The classical machine learning and deep learning classifiers, that have been compared in this paper, are described within this section.

A. PRE-PROCESSING
Given the fact that social media messages are frequently written using a casual language, a pre-processing step has been used in order to prepare the tweets in the annotated dataset for training the machine learning classifiers. This step is considered crucial by D'Andrea et al. [5] for the success of the entire system, while Bao et al. [90] provide a comprehensive discussion regarding the importance of preprocessing in social media analysis. The impact of the different pre-processing steps, such as the removal of links, on the performance of classical machine learning classifiers has been discussed by Jianqiang and Xiaolin [91].
During this pre-processing step, all the user mentions, easily identified through the presence of the @ symbol at the beginning of the message have been normalized, since they do not provide any useful information for the classification process. All the links and email addresses have been normalized as well. The emoticons have been replaced with the corresponding words. Minor spelling mistakes have been automatically corrected to improve performance. Contractions and hashtags have been unpacked, while elongated words have been corrected and annotated. Finally, all the letters have been converted to a lowercase representation. The pre-processing has been implemented with the help of the ekphrasis library [92]. Additional processing has been performed through Natural Language Toolkit (NLTK) library [93] and the ''re'' python module.

B. FEATURES
In order to use machine learning algorithms for text classification, the text content has to be first converted into numerical feature vectors. The Bag-of-Words (BoW) scheme converts the text to a numerical representation, having as a starting point the frequency of the words. Given a vocabulary V = {w 1 , . . . , w N , containing N tokens, denoted using w i , a tweet, or any other textual document d, belonging to a corpus D, can be represented using a feature vector X = {x 1 , .., x N , in which x i can either represent a binary variable that indicates whether the word w i appears in the text or a numeric variable indicating the number of times the word w i appears in the text.
Given the fact that very frequent words can sometimes carry little ''informational content'', the performance of classification algorithms that rely on word frequencies can be improved using a more complex feature representation, called Term Frequency -Inverse Document Frequency (TF-IDF), that reduces the weight associated to words that frequently appear in all the documents in the corpus. TF-IDF is computed as shown in (5): where TF (w i ) represents the number of appearances of the word w i , |D| stands for the number of documents and DF (w i ) is the number of documents containing the term w i . The TF-IDF statistical measure is used throughout the present study for features representation. By only focusing on the number of times a word occurs in a given text, the Bag-of-Words approach does not provide any information regarding the succession of the words. This issue can be addressed if the n-gram language model is used, in which the text is represented through successions of N consecutive words. Common types of n-grams include grams of size one, called unigrams (1-grams), grams of size two, called bigrams (2-grams), and grams of size three, called trigrams (3-grams) [94].
In the present study, various combinations of unigrams, bigrams and trigrams have been considered as features for the machine learning algorithms, as shown in Table 5. Besides the Bag-of-Words representation, word embeddings have been used. In word embeddings the words are mapped to vectors, having similar representations for the words which frequently appear in the same context. Compared to one-hot encodings, word embeddings provide a denser representation that requires a smaller number of dimensions for representing the words. The similar representation of words with close meanings provides additional clues for the classification algorithms. The following word embeddings have been considered in the present study: Datastories, 3 GloVe 4 and Fast-Text. 5

C. LEARNING ALGORITHMS
A machine learning approach has been used in order to accurately determine the stance towards vaccination in the collected tweets. Starting from the annotated dataset, the performance of several popular classification algorithms has been investigated: Multinomial Naive Bayes (MNB), Random Forest (RF), Support Vector Machine (SVM), Bidirectional Long Short-Term Memory (Bi-LSTM) and Convolutional Neural Network (CNN).

1) MULTINOMIAL NAIVE BAYES
Naive Bayes classifiers are a family of probabilistic classification algorithms that apply the Bayes theorem. They are called naïve because they perform the classification under a strong assumption that every feature is independent from the other features. Despite their simplicity, this family of algorithms has been demonstrated to be fast, reliable and accurate in many NLP classification tasks [95]. The Multinomial Naive Bayes [96] classifier implements a variant of the Naïve Bayes algorithm which can be used with multinomially distributed data, such as the frequencies of n-grams in text classification problems.

2) RANDOM FOREST
Random Forest (RF) [97] is an ensemble classifier that consists of multiple decision tree classifiers, trained in parallel with bootstrapping followed by bagging. According to Misra and Li [98] the RF classifier offers better results when compared to other classification methods in terms of accuracy and does not require feature scaling. Furthermore, the RF classifier has been determined to be more robust in the selection of training samples. Even though the RF might be hard to interpret, its hyperparameters can more easily be turned than in the case in which a decision tree classifier is used [98].

3) SUPPORT VECTOR MACHINE
Support Vector Machines (SVM) [99] are a family of supervised learning algorithms used for classification, regression and other tasks such as outlier detection. While other classification algorithms suffer from overfitting, one of the advantages of SVM is that they are less prone to this situation [100]. Another advantage resides in the fact that besides binary classification, multiclass classification can be performed by combining several binary classification functions. For this, each class is considered individually at a time, and for each class a classifier is searched that separates it from the other classes [101].

5) CONVOLUTIONAL NEURAL NETWORK
Convolutional Neural Networks (CNN) are a type of neural networks that are specialized for processing data that features a grid-like topology [46]. CNNs have already been successfully used in different NLP tasks, including stance classification [5], [104], [105].
In the current paper, we have followed the approach used by Cliché [104] and Baziotis et al. [106] regarding the filter lengths of [3]- [5]. Additionally, the architecture of the network is similar to the one presented in [106]. In the approach used in the current paper, the word embedding layer (none, 50, 300) is followed by a Gaussian noise layer (none, 50, 300) and by a dropout layer (50,300). After this layer, three 1-D convolutional layers (using ReLU activation) have been added, each followed by a max pooling layer and a flattening layer. The outputs of these layers are merged in a concatenation layer (none, 7000). The dropout layer (none, 7000) and a dense layer (none, 3) conclude the network.

6) BIDIRECTIONAL ENCODER REPRESENTATIONS FROM TRANSFORMERS
Bidirectional Encoder Representations from Transformers (BERT) [63] is a pre-trained transformer-based language model. Compared to word embeddings such as GloVe, BERT has the advantage of taking into account the context for each occurrence of a given word. The model has been pre-trained on a diverse corpus of unlabeled text extracted from the English Wikipedia and the BookCorpus [107].
Pre-trained BERT models with a wide range of sizes exist, varying the number of layers L from 2 to 24 and the hidden size H from 128 to 1024 [63], [108]. In the present paper, the BERT BASE 6 model has been chosen, having L = 12, H = 768 and the number of self-attention heads A = 12. The neural network architecture has a total of 110M parameters. In comparison, BERT LARGE (L = 24, H = 1024, A = 16), having 340M parameters, has been shown to provide 6 https://huggingface.co/bert-base-uncased improvements in accuracy of no more than 5% [63], while being far more compute intensive.

D. EXPERIMENTS AND RESULTS
The classical machine learning algorithms have been implemented using the scikit-learn [109] library, while the deep learning algorithms have been implemented using the Keras 7 library, having TensorFlow 8 as a backend.
Cross-validation using either 5-folds [51] or 10-folds [5] is a widely-used approach for comparing and selecting classifiers. In this paper, following the approach described by Mohammad et al. [51], the classifiers have been evaluated through a 5-fold cross-validation procedure, during which the classification model is trained using k-1 of the folds as training data, while the resulting model is validated on the remaining part of the data. The performance of the classifier is then computed as an average of the values computed during the k consecutive runs. Since the balanced dataset includes 3249 tweets (1083 in each class), at each iteration, the classification models are trained using 2600 tweets and evaluated using the remaining 649 tweets. The results of the considered methods are shown in Table 6 and further discussed in the sub-sections below.

1) BOW AND CLASSICAL MACHINE LEARNING
The best parameters for the developed natural language processing pipeline have been determined through the grid search approach. Thus, different numbers of features have been tested, including using all the features and reducing the number of features, F, to a maximum of 1500, 2000 and 3000 values.
Different n-gram combinations, ranging from (1,1) to (1,3), as listed in Table 5, have been investigated for the string vectorizer, as well. Additionally, the algorithms have been evaluated considering both the case in which the general stop words are kept and the one in which they are excluded. The stop word list that has been considered is the one included in the NLTK library. In the case of the corpus-specific stop words, the document frequency thresholds, maxDF, that have been considered are 0.5, 0.75 and 1.0. Besides, the evaluation has also analyzed whether applying Term Frequency (TF) or Term Frequency -Inverse Document Frequency (TF-IDF) can improve the stance classification results.
We have experimented with different settings for the classifiers, including varying, in the case of the SGDClassifier, the alpha parameter, which multiplies the regularization term. Different regularization terms have been tested, including ''l1'', ''l2'' and ''elasticnet''. The loss function of the SGD-Classifier has been configured as ''hinge'', corresponding to a linear SVM.
For each considered classical machine learning classifier (C1-C6), Table 6 includes both the results achieved using the parameters determined through grid search (C1, C3, C5) and the results achieved through the n-gram model (1, 2), corresponding to unigrams and bigrams, and limiting the maximum number of features to F = 2000. For the C1-C6 classifiers, the best results have been achieved when using TF-IDF, without excluding the general stop words.
In the case of the Multinomial Naïve Bayes classifier (C1 and C2) the best results have been achieved for the C1 classifier, for which the maximum number of features has been reduced to F = 3000, while including both unigrams and bigrams as features, keeping maxDF = 1.0.
It has been observed that the Random Forest classifier (C3 and C4) performed best when using only unigrams, without limiting the number of features, while applying a frequency threshold, maxDF, for corpus specific stop words of 0.5, namely in the case of C3.
In the case of the Support Vector Machines classifier (C5 and C6), the best results have been achieved in the case of C5, configured with the n-gram model (1,2), without limiting the maximum number of features, an alpha parameter value of 0.0001, a maxDF threshold equal to 1.0 and choosing ''elasticnet'' as a regularization term.
As expected, C1, C3 and C5, for which the parameters have been determined through grid search have performed better than the corresponding classifiers of the same type, C2, C4 and C6.
The overall best performing classifier has been C5, a SVM classifier which had 76.23% accuracy, followed by C6, with 74.20% accuracy. In terms of precision and F-score, C5 overperformed all the other classifiers for each of the three considered classes, in favor, against and neutral. A small difference is recorded in the case of recall, where the value for the neutral class is slightly lower for C5 than for C6 (76.18% versus 77.84%).
The worst performing classifier has been C4, a RF classifier, with an accuracy of 70.79%. In terms of precision and recall, the classifier C4 performed worse than C5 and C6 on all three classes, in favor, against and neutral.

2) WORD EMBEDINGS AND CLASSICAL MACHINE LEARNING
Starting from the algorithm that has provided the best results in the context of the Bag-of-Words approach, C5, in the following we have analyzed if the performance can be further improved by considering pre-trained word embeddings. Similar approaches, using word embeddings with classical machine learning algorithms, have been investigated in [5] and [110].
To this end, a word embedding, called glove.6B, that includes six billion tokens, created through the Glove approach from a corpus extracted from Wikipedia and from the news archive Gigaword [111], has been used. This implementation is marked in Table 6, as C7.
As shown in Table 6, the values of all the four considered metrics (precision, recall, F-score and accuracy) of the C7 classifier are worse than those achieved in the case of C5.

3) WORD EMBEDINGS AND DEEP LEARNING
In the case of the deep learning classifiers, in the present paper, the Adam approach has been applied for tuning the learning rate [112]. The resulting classifiers are listed in Table 6 under the C8 -C13 classifiers.
As shown in Table 6, among the Bi-LSTM classifiers (C8-C10), the best results have been achieved by the C9 classifier, with an accuracy of 74.70%, higher than in the case of C8, (73.41%) and C10 (68.36%). The C9 classifier has used the word embeddings created through the Glove approach from a corpus composed of 2 billion tweets.
In the case of CNN classfieirs (C11-C13) the best results have been achieved by the C13 classifier, using the word embeding created through the FastText approach on a corpus extracted from Wikipedia and news stories (accuracy 69.01%).
Classifiers C8 and C11, ranked second in the Bi-LSTM category (73.41%) and third in the CNN category (65.71%) based on accuracy, have used the Datastories word embeddings, created from a corpus of 330 million tweets, by applying the GloVe approach.
The best performing deep learning classifier has been C9, implementing Bi-LSTM, which outperforms C8, C10-C13 classifiers, both in terms of accuracy and F-score.

4) BIDIRECTIONAL ENCODER REPRESENTATIONS FROM TRANSFORMERS
In order to establish the best values for the hyperparameters of the BERT language model (C14), the approach recommended by Devlin et al. [63] has been followed during the fine-tuning procedure in regarding the batch sizes (16,32), learning rate (5e-5, 3e-5, 2e-5) and number or epochs (2,3,4). The best results have been achieved when using a batch size of 16, a learning rate of 3e-5 and a number of epochs equal to 3. Having an accuracy of 78.94%, the C14 classifier outperforms all the other classifiers. Moreover, it clearly outperforms the second-best performing classifier, C5, in terms of precision, recall and F-score, for all the considered classes.

E. DISCUSSION
The results achieved by the deep learning classifier C9 are worse than the ones obtained in the case of the classical machine learning classifier C5 in terms of accuracy and F-score. This result is consistent with the ones in other studies, such as D'Andrea et al. [5], in which classical machine learning algorithms have outperformed deep learning approaches, such as CNN and LSTM, in the case of vaccine stance classification.
As noted in the review paper of Wang et al. [31], stance detection approaches typically do not perform extremely well. The reasons mentioned by the authors include the sparsity, the colloquial language and the absence of large, labeled datasets that could be used for training. Moreover, Mohammad et al. [51] summarize the results of the ''SemEval-2016 Task 6: Detecting Stance in Tweets'' mentioning that the SVM baseline with n-grams has performed relatively well compared to other machine learning approaches.
In the following we have used the best performing classifier, C14, to analyze the tweets collected over the considered period of time. The model has been trained on all the tweets in the annotated dataset.

VI. ANALYZING SOCIAL MEDIA
The evolution of the daily number of tweets is discussed in this section in connection with the major events which have occurred around the world related to COVID-19 vaccination, with an accent on the English-speaking countries.

A. MAJOR EVENTS
The major events have been extracted from the news published online in each day of the analyzed period using google.com search engine by selecting the ''News'' section and ''COVID'' keyword and by pointing one-by-one the days in the mentioned period. Each time, the first 10 pages of News titles have been considered and the most relevant news have been extracted in connection to the COVID-19 vaccination theme, relevance being given by the connection to the COVID-19 vaccination and the amount of news on a specific topic. VOLUME 9, 2021 As a result, it has been observed that in all the analyzed days there have been news regarding the COVID-19 vaccination theme, starting from the announcement of the vaccine effectiveness by different producers, the amount of money funded by various organizations for COVD-19 vaccine, adverse events encountered in the pre-test phase, ethical issues related to whom should have first access to the vaccine, the predicted quantity of vaccines to be distributed in different countries and areas and ending with the vaccination in Russia and UK.
From the events presented above, we have selected two events, one that has generated a large number of tweets, namely E10, and another one that has generated a comparatively smaller number of tweets, namely E6.

B. STANCE ANALYSIS
In the following, the best performing classifier -determined in Section V, BERT (C14) -is used to perform stance analysis on the gathered dataset. As it will be observed, not all the 19 https://www.economist.com/science-andtechnology/2020/12/01/britain-becomes-the-first-country-to-license-afully-tested-covid-19-vaccine (accessed December 9, 2020) 20 https://www.bbc.com/news/uk-55181665 (accessed December 9, 2020) 21 https://www.bbc.com/news/uk-55227325 (accessed December 9, 2020) news published in the analyzed period have generated the same amount of interest from the general public, a series of local peaks being identified in some of the analyzed days.

1) CLEANED TWEETS STANCE ANALYSIS
The evolution of the stance expressed and the distribution of the stances on the three considered categories: in favor, against and neutral is depicted in Figure 2, which considers only the cleaned tweets. By simply considering the stances' evolution, one can easily observe that there have been variations in the number of tweets published, especially in the days following a major announcement or news. From Figure 2 it can be observed that the general dominating stance is neutral, as it was expected from the initial stage in which most of the annotated tweets belonged to this category. Based on the tweets in the neutral category, it has been noticed that most of them deal with presenting news related to the occurrence of the COVID-19 vaccine.
As for the number of against and in favor tweets, it can be observed that they have oscillated during the analyzed period. Based on Figure 2 it can be observed that between November 9 and December 1 the number of against tweets has kept a constant trend, with a few ''spikes'' on November 9, November 16 and November 23. In all of these days, the news released in the media were speaking about the efficiency of COVID-19 vaccines in clinical trials: for Pfizer and BioNTech on November 9 (event E1), for Moderna on November 16 (E4) and for Oxford AstraZeneca on November 23 (E7). Two major turning points for increasement in the number of against tweets have been represented by December 2 and December 8, these being the days in which UK has authorized the Pfizer BioNTech COVID-19 vaccine (E10, 6242 against tweets) and the day in which the COVID-19 vaccination started in the UK (E12, 8429 against tweets).
Considering the evolution of the graphic containing the cleaned tweets from all the three categories, it can be observed that the number of against tweets follows on a smaller scale the evolution of the number of neutral tweets (the curve of the against tweets being more flattened than in the case of the neutral tweets), while the in favor tweets follow more precisely the trend imposed by the number of neutral tweets (Figure 2; Figure 7 and Figure 8). As a result, for the in favor tweets one can observe an increasement in most of the major events cases presented above. Even more, in the events announcing the effectiveness of the COVID-19 vaccines from different companies (E1, E4, E5 and E7) the in favor tweets have overpassed the number of against tweets. Even after the vaccine authorization in UK, the situation has not changed and the number of in favor tweets overpassed daily the number of against tweets (with approximately, on average, 2050 tweets per day).

2) ALL TWEETS STANCE ANALYSIS
As the difference, in the number of tweets, between the in favor and against stances in the cleaned dataset seem to have close values during the analyzed period, a stance analysis considering all tweets dataset has been performed.
In the study conducted on vaccination in Italian tweets, D'Andrea et al. [5] considered that one should analyze the whole tweets dataset (including retweets) as, some of the users who retweet a certain opinion or piece of information, generally believe in it and, instead of writing their own words to a particular situation, they might decide instead to share the information. As a result, we have run the stance analysis over the entire tweet-dataset and the results are presented in Figure 3.
As it can be observed from Figure 3, on November 9, when Pfizer and BioNTech announced their vaccine effectiveness, the number of retweets has been only slightly different from the ones of tweets (a difference of 497 tweets). On November 10, the number of cleaned tweets has been half compared to November 9, while the total number of tweets doubled, showing a high increasement in the number of retweets. As the events marking November 10 were mostly referring to Even in this case, one can notice that the evolution of the number of in favor tweets resembles more closely to the evolution of the neutral tweets. The increasement in the number of tweets marked as in favor is better visible in the case of the occurrence of the events mentioned above E1-E12, more precisely in the cases of E2, E4, E5, E7, E10 and E12 (Figure 9), than in the other days of the analyzed period. For the against tweets, except for the increasement observed in the period following E10 (Figure 10), the evolution is almost constant, recording an average number of against tweets of approximately 6826 tweets/day. After E10, the number of daily against tweets recorded an increasement of 95.28%.
Comparing the number of tweets recorded in the three categories for the period between December 2 -December 8 (E10-E12) with the period between November 9 -December 1 (E1-E9), it can be observed that the higher relatively increasement has been recorded in the against tweets category (95.28%), followed by neutral (87.86%) and in favor (54.86%). The smaller difference noticed between the number of tweets in the against category compared to the in favor category can be attributed to the fact that starting from December 2 (E10) the potential of having a vaccine was no longer a ''dream'' but it became ''reality'', which might have ''activated'' the anti-vaccination community.

3) CONTEXT RELATED EVENT COMPARISON
For better understanding the opinion polarity in connection with a given event, we have considered in the following, three major events recorded in this month: first announcement of a possible effective vaccine (E1), first authorization of a vaccine (E10) and first vaccination (E12), all of them creating spikes in the number of tweets.
As a result, it has been decided to aggregate the opinions expressed in the days corresponding to the three events for both the cleaned and for the entire data set. The results have been summarized in Figure 4.   A first observation is related to the fact that the stance is overall neutral for the considered events. Particularly, in terms of relative distribution, one can note that that percentage of the neutral tweets in total number of tweets has recorded a slight increase of 17% (from 56% for E1 to 73% for E12) when all the tweets are considered and an increase of 7% on cleaned tweets (from 56% for E1 to 63% for E12) -please see Figure 4.
A higher percentage variation, in relative terms, can be observed for the in favor tweets: on the entire dataset the percentage dropped from 31% to 20%, while in absolute value, the number of in favor tweets almost doubled from 17 526 tweets for E1 to 44 447 tweets for E12 (with 27 487 tweets for E10).
A decrease in the percentage of tweets from the entire sample can be also observed in the case of against tweets, even though, in this case, the decrease is of only 6% (from 13% to 7%). Considering the cleaned tweets, the percentage of the against tweets has faced an increase of 1% (between E1 and E10-E12). Even in this case, the absolute number of against tweets on all dataset increased from 7604 tweets on E1 to 12 541 on E10 and to 14 359 on E12.
Another observation can be made in connection to E1 and E10: it can be observed that in the case of the against tweets, the increasement of the number of these tweets happened mostly in the following days after the events occurred, not being so visible in the day of the event. This observation is in line with the study conducted by D'Andrea et al. [5], in which the authors have shown that sometimes the effect of an event is immediately visible, while in some other cases, it might need some hours / a day in order to be visible in the tweets. A possible explanation for this lag in response might be related to the fact that the users posting against tweets might have needed some time in order to look for information related to the event prior to tweeting.

4) CUMULATIVE STANCE ANALYSIS
As expected, the predominant stance of the cumulative set is neutral both in the entire set ( Figure 5) and in the cleaned set.
The percentage of the against and in favor reaches the same value in the case of cleaned tweets, even though in absolute value, there can be noticed a difference in favor of the against tweets ( Figure 6). On the entire dataset, the in favor tweets have a 10% difference compared to the against tweets ( Figure 6). As the difference between the in favor and against tweets is not as visible as expected ( Figure 6) and a global vaccination campaign is expected, the involved agencies and governments should try to provide more information regarding the vaccination process, its advantages and its presumed disadvantages, offering to the general public all the needed instruments and information for increasing their trust in the decisions taken at a macro-scale, with impact on everybody's life.

VII. LIMITATIONS OF THE STUDY
A potential limitation of the current study is represented by the classification algorithms selected in the paper. As the Natural Language Processing field is in continuous development, better algorithms could be developed over time providing an improved classification performance. Another limitation is related to the selected dataset, which only includes tweets extracted between November 9, 2020 and December 8, 2020, written in English. This limitation opens the path towards VOLUME 9, 2021    other possible extensions, which might include, but not limited to analyzing the tweets written in other languages or from specific geographical areas, which could be identified using the associated GPS coordinates. Additionally, a different period of time could be considered due to the vaccination process dynamics worldwide.

VIII. CONCLUSION
In the current paper, the one-month period passed between the first announcement of a coronavirus vaccine and the first actual vaccination process started outside the limited clinical trials has been analyzed using machine learningbased stance detection. Multiple classical machine learning and deep learning algorithms have been compared and the best performing classifier has been chosen based on four performance metrics.
The proposed approach has classified the tweets into three main classes, namely in favor, against and neutral regarding COVID-19 vaccination, employing BERT with an accuracy of 78.94%.
The aim of the paper has been to monitor the evolution of the stance towards COVID-19 vaccination from tweets, by matching the number of Twitter messages with the main events reported by the media in the analyzed period.
The main stance was neutral at both a daily level and for the entire period, on either the cleaned dataset or on the complete dataset (dataset including all the tweets). In terms of neutral tweets' percentage recorded at the beginning of the period, November 9, versus the end of the analyzed period, December 8, it has been observed that the percentage has increased with 17% on the complete dataset. The evolution of the in favor tweets has been characterized by a series of spikes, closely following the evolution of the neutral tweets, but at a reduced scale. For both the neutral and in favor tweets it has been observed that some of the events in media have entrained a series of spikes, which are not encountered in the case of against tweets, where the major spike has been represented by the authorization in the UK of the Pfizer BioNTech COVID-19 vaccine.
Even more, it has been observed that in the case of the against tweets there is sometimes a lag between the day in which a certain event has occurred and the change in the number of against tweets. The existence of such a lag in tweets posting has also been mentioned by D'Andrea et al. [5] in a paper regarding vaccination in Italy.
The correspondence between the information in the tweets and the events in the media has been confirmed by analyzing the n-grams, underlining even more the fact that the tweets reflect the hot topics in the society at large.
The early detection of an opinion shift might be highly useful in the context in which many countries from all around the world are planning to start the COVID-19 vaccination process, as it would allow governmental decision makers to promote actions aimed at limiting the distribution of fake news and increasing the general public's confidence towards vaccination. The analysis can be performed on a daily basis, in order to have a real-time overview of the stance evolution.
Possible future research directions include the development of better performing stance classification algorithms, as well as extending the analyzed period, especially given the fact that the vaccination process is expected to take a relatively long period of time.

ACKNOWLEDGMENT
Liviu-Adrian Cotfas and Camelia Delcea would like to thank the members of the ELLIADD Laboratory, University of Bourgogne Franche-Comté and Pays de Montbéliard Agglomération for the whole support offered during the time when this work has been carried out. VOLUME 9, 2021 LIVIU-ADRIAN COTFAS received the Ph.D. degree in economic informatics from the Bucharest University of Economic Studies, Bucharest, Romania.
In 2018, he has been a Visiting Professor with Université de Technologie Belfort-Montbéliard, France. He is currently with the Economic Cybernetics and Informatics Department, Bucharest University of Economic Studies. His research interests include machine learning and deep learning techniques, semantic Web, agent-based modeling, social media analysis, sentiment analysis, recommender systems, grey systems theory, and geographic information systems.
Dr. Cotfas has received several research awards, including the Georgescu Roegen Award for excellent scientific research. He is an active member of the Grey Uncertainty Analysis Association and of the INFOREC Association.
CAMELIA DELCEA received the Ph.D. degree in economic cybernetics and statistics from the Bucharest University of Economic Studies, Bucharest, Romania.
She is currently with the Economic Cybernetics and Informatics Department, Bucharest University of Economic Studies. She is an active member in the area of grey systems theory, agent-based modeling, and sentiment analysis. Since 2009, she has obtained 25 international and national awards (Best Paper Award, Georgescu Roegen for excellent scientific research, Excellent Paper Award, and Top reviewers) and has been invited to deliver a keynote speech on grey systems themes in 2013, 2016, and 2017, at the IEEE GSIS conference and in 2018 at the GSUA conference. She is an active member of the Grey Uncertainty Analysis Association.
IOAN ROXIN received the Ph.D. degree in statistics from the Bucharest University of Economic Studies.
He is currently with the University of Franche-Comté and the Head of the ELLIADD Laboratory. His research interests include the area of knowledge organization, artificial intelligence, the IoT, and semantic Web, with applications in social networks and learning. Over the years, he has managed or has acted as a Board Member in several research projects on topics related to learning and communication. He has been a Visiting Professor or an Invited Speaker at several universities from Canada, Finland, Japan, Liban, Mexic, Romania, and Vietnam.
Dr. Roxin is an active member of the ACM and IEEE associations.
CORINA IOANĂŞ received the Ph.D. degree in economic studies from the Bucharest University of Economic Studies, Bucharest, Romania. Over the last years, she has been a member in eight national and international projects, having ten books written in the economic field. She is currently with the Accounting and Audit Department, Bucharest University of Economic Studies. Her main research interests include accounting, audit, consumer behavior, economic modeling, economic policies, and risk analysis.
Dr. Ioanăş is an active member of the Centre for the Accounting and Management Information Systems, ARACIS (Romanian Agency for Ensuring the Higher Education Quality), and CECCAR (The Body of Expert and Licensed Accountants of Romania).
DANA SIMONA GHERAI received the Ph.D. degree from the Faculty of Economics and Business Administration, Timisoara West University.
Over the last years, she has been a member in five European research projects, having four books written in the economic field. She is currently with the Finance-Accounting Department, University of Oradea. Her main research interests include accounting, audit, consumer behavior, economic policies, economic entities, economic modeling, ethics, and risk analysis.
Dr. Gherai is an active member of the CECCAR (The Body of Expert and Licensed Accountants of Romania).
FEDERICO TAJARIOL received the M.S. degree in cognitive science from the National Polytechnic Institute, Grenoble, and the Ph.D. degree in cognitive science from Grenoble University. After an experience as a Researcher with IT international company, he joined the University of Franche-Comté, in 2009, where he currently works as Associate Professor in information science. He is also with the University of Franche-Comté and a member of the ELLIADD Laboratory. His research interests include the user experience design and evaluation of digital services in learning and crisis settings. He is a member of the Design Society, ICA, and ISCRAM scientific associations. VOLUME 9, 2021