Abstract In today’s world, everyone is expressive in one way or other. Many social websites and android applications whether being Facebook, WhatsApp or Twitter, in this highly advance and the modernized world is flooded with views and data. One of the most global and popular platforms is Twitter. This is seen as the main source of sentiments where almost every enthusiastic or social person tends to express his or her views in form of comments. These comments not only express the people but also give the understanding of their mood. Text present on these medias are unstructured in nature, so to process them firstly we need to pre-process, six pre-processing techniques are used and then features are extracted from the pre-processed data. There are so many feature extraction techniques such as Bag of Words, TF-IDF, word embedding, NLP(Natural Language Processing) based features like word count, noun count etc. In this paper we analysed the impact of two features TF-IDF word level and, N-Gram on SS-Tweet dataset of sentiment analysis. We found that by using TF-IDF word level (Term Frequency-Inverse Document Frequency) performance of sentiment analysis is 3-4% higher than using N-gram features, analysis is done using six classification algorithms(Decision Tree, Support vector Machine, K-Nearest Neighbour, Random Forest, Logistic Regression, Naive Bayes) and considering F-Score, Accuracy, Precision, and Recall performance parameters.
[1]
Joel D. Martin,et al.
Sentiment, emotion, purpose, and style in electoral tweets
,
2015,
Inf. Process. Manag..
[2]
U. Rajendra Acharya,et al.
Author's Personal Copy Biomedical Signal Processing and Control Automated Diagnosis of Epileptic Eeg Using Entropies
,
2022
.
[3]
Hiroshi Nagahashi,et al.
Parameterized AdaBoost: Introducing a Parameter to Speed Up the Training of Real AdaBoost
,
2014,
IEEE Signal Processing Letters.
[4]
Michael Paraskevas,et al.
Examining the Impact of Feature Selection on Sentiment Analysis for the Greek Language
,
2017,
SPECOM.
[5]
Gaël Varoquaux,et al.
Scikit-learn: Machine Learning in Python
,
2011,
J. Mach. Learn. Res..
[6]
Leo Breiman,et al.
Random Forests
,
2001,
Machine Learning.