In past years, quite a few studies have been carried out on the subject of "Sentiment Analysis on Twitter" by a number of scholars.[1]
Go and L.Huang (2009) proposed an answer for sentiment evaluation for twitter statistics via the use of remote supervision, wherein their education information consisted of tweets with emoticons, which served as noisy labels. Applying Naive Bayes, MaxEnt, and Support Vector Machines (SVM), they build models. It had positive, negative, and neutral characteristics. They came to the conclusion that SVM outperformed several trends and that unigram had been a more potent characteristic.
Pak and Paroubek(2010) a version that would classify the tweets as favourable, neutral, or negative. By compiling tweets using the Twitter API and automatically annotating the ones that included emotions, they produced a Twitter corpus. They developed a sentiment classifier based mostly on the multinomial Naive Bayes method using that corpus.
Barbosa.et.al.(2010) created a segment-automated sentiment evaluation method for putting tweets into different categories. They classified tweets as objective or subjective before classifying the subjective tweets as positive or unfavourable in the subsequent step. Retweets, hash tags, links, punctuation, exclamation points, as well as features like previous polarity of phrases and POS, were all included in the function area.
Bifet and Frank(2010) utilised Twitter streaming data provided by the Firehouse API, which provided all real-time messages from each user that were visible to the public. They came to the conclusion that the SGD-based model turned out to be superior to the others employed when utilized with the right learning rate.
Davido.et.al.,(2010) offered a method to leverage punctuation, single words, n-grams, and styles as specific function types that are then combined into a single function vector for sentiment classification in order to employ Twitter user-described hash tags in tweets as a kind of sentiment. By building a function vector for each instance, they were able to assign sentiment labels using the K-Nearest Neighbour approach.
Agarwal.et.al.(2011) developed a three-way approach for categorizing sentiment into good, terrible, and neutral categories. They experimented with many models, including the unigram version, a version based on features, and a version based on a tree kernel. They displayed tweets as a tree for the version that was mostly based on the tree kernel. One hundred capabilities are used in the entire version based on functions, and over 10,000 in the unigram version. They came to the conclusion that functions that combine the previous polarity of phrases with their parts-of-speech (pos) tags are essential and play a top-notch role in the typing work. The contrary models were implemented by the tree kernel, which was based on the entire version.
Turne.et.al used bag-of-phrases technique for sentiment evaluation, wherein the relationships among phrases were never taken into consideration and a report was represented as only a series of words. To decide the sentiment for the entire document, the sentiments of each phrase were decided and people's values were united with a few aggregation functions.
Kamps.et.al. worked on WordNet's lexical database to identify a phrase's emotional meaning combined with its extraordinary dimensions. On WordNet, they developed a distance metric and determined the semantic polarity of adjectives.
Luoet. al. highlighted the demanding situations and green strategies to mine reviews from Twitter tweets. Spam and wildly varied language make opinion retrieval inside Twitter a tough task.