3.1. Theoretical implication for online review mining
Customers are interested in observing the opinions of other customers via online product reviews, blogs, and social networking sites, among others. Therefore, customer feedback is crucial in determining purchasing behavior and attitude toward online shopping (Sohail et al., 2016). Customer reviews on the internet are a significant source of information for consumers evaluating items before making a purchase (Zhao et al., 2018). Customers who are considering making a purchase frequently read online product reviews given by other buyers. Such reviews give consumers a glimpse of other users' purchasing and usage experiences. Shoppers see these reviews as more reliable than seller-sponsored marketing (Thakur, 2018). Due to the pervasiveness of social media on the internet, opinion mining has become a necessary method for analyzing data, which, if successful, aids in making decisions (Sohail et al., 2016; Sun et al., 2017). Customers can use online reviews to make more informed purchasing decisions because they have read the reviews of other customers prior to making a purchase. After consumption and experience, a comparison is made between the actual outcomes and the customer's expectations. Then, feedback is provided based on performance perceptions employing sentiments as a post-purchase behavior to demonstrate purchase satisfaction.
On the other hand, compared to merchant-provided recommendations or reviews on shopping platforms, consumers prefer information provided on social networking platforms (Li and Ku, 2018; Qazi et al., 2017). Opinions are divided into two categories: comparative and non-comparative (El-Halees, 2012; Eldefrawi et al., 2019; Xu et al., 2011). Researchers determined four types of comparable opinions: non-equal gradable, equative, superlative, and non-gradable (Kansaon et al., 2020).
Alternatively, according to (Nasiri and Shokouhyar, 2021), non-comparative opinions were classified as regular opinions. They initially selected and classified eight customer satisfaction dimensions (CSDs) into three categories. "features of the product" was the most important of these categories. These features will be accessible to customers who acquire a refurbished smartphone. The literature demonstrates the significance of review mining. Generally, two types of opinions exist: comparative and non-comparative. In opinion mining, we focus on equative-type comparative opinions and regular non-comparative opinions.
Sentiment analysis, which is a component of opinion mining, is used to identify views and arguments in a given text. If there are any positive, negative, or neutral statements in comments or reviews, the algorithm attempts to uncover assertions of agreement or disagreement (Wang and Zhang, 2020). After selecting the target and utilizing the sentiment identification algorithm (Giatsoglou et al., 2017), the textual documents are classified as positive, negative, or neutral based on their polarity. For the sentiment identification stage, various strategies may be used, some of which are unsupervised and others supervised. Lexicon-based approaches are frequently used in unsupervised scenarios; lexical resources are utilized to assign polarity ratings to individual words to determine a text's overall emotion. When it comes to unsupervised scenarios, machine learning is frequently employed, with algorithms such as Nave Bayes, Neural Networks, Decision trees, Linear regression, and Support Vector Machines (SVM) used for sentiment detection tasks serving as classification problems.
3.2. Practical methods
This research study classifies online reviews of refurbished products into two groups: (i) direct evaluations that focus on specific features of the refurbished product and analyze the sentiment towards them using machine learning, and (ii) indirect evaluations that rely on the claims made by the manufacturers of refurbished products that they are comparable to new ones. In this part, we will explain the methods we used to conduct our research and gather data on customer satisfaction with refurbished smartphones in both developed and developing markets. Using the three distinct procedures depicted in Fig. 1, it will be disassembled into its component parts and described in detail. To comprehend how satisfied customers feel after making a purchase, it is crucial to collect online reviews, preprocess them, and analyze both comparative and non-comparative opinions.
To execute the steps in Fig. 1, Twitter API was used to collect data from developed and developing countries. The reviews were then preprocessed using tokenization, stop word removal, stemming, and part-of-speech (POS) tagging. Then, we separated reviews into opinions that were comparative and non-comparative. We identified comparative sentences in comparisons using entities and related keywords. Subsequently, we removed sentences that were not connected. Afterward, we extracted customer satisfaction dimensions (CSDs) using Latent Dirichlet Allocation (LDA) and word frequency statistics for non-comparative opinions. A support vector machine (SVM) determined the sentimental inclination of reviews and CSDs.
Fig .1. Post-purchase satisfaction framework for refurbished smartphones in developed and developing countries.
Step 1. Data collection
Utilizing the Twitter API to obtain data from the target countries was the first step in the data collection process. This study employed tweets from developed countries, including the United States, Canada, the United Kingdom, Australia, and Germany, and tweets from developing countries, including India, South Africa, Malaysia, Saudi Arabia, and Brazil. Twitter has become one of the most important data sources for data sharing, with most users sharing their opinions on topics such as purchasing a new device, device performance, and others. Due to the widespread use of Twitter, users' tweets will reveal concerns, challenges, or issues encountered by users of refurbished mobile phones.
Step 2. Data preprocessing
Cleaning or preprocessing data is essential before the machine learning system can prepare and highlight attributes that require processing (Drikvandi and Lawal, 2020). Many open-source NLP tools, such as the Natural Language Toolkit (NLTK), OpenNLP, and Stanford CoreNLP, support major NLP tasks (Sun et al., 2017). Python is excellent at handling text data, which has led to the development of several popular libraries for information retrieval, natural language processing, and text analytics, such as NLTK, Gensim, and spaCy (Sarkar, 2019). An extensive platform with a variety of corpora and lexical resources is the Natural Language Toolkit (NLTK). In addition, it includes the tools, interfaces, and methods required to manage and evaluate text data (Lauriola et al., 2021). Some techniques for cleaning or preprocessing text data include tokenization, stop-word removal, stemming, and part-of-speech (POS) tagging. These procedures must be implemented in our text collection (Drikvandi and Lawal, 2020). In the following sections, we will examine the strategies mentioned above are examined in detail.
"Tokenization" refers to the process of dividing a block of text into smaller units, such as words, phrases, or symbols, that are meaningful for a specific task. Although this task may seem straightforward when working with machine-readable formats, it still requires addressing certain challenges, such as removing punctuation marks and hyphens at the end of lines. Nonetheless, tokenizers can correctly process other characters, including brackets and hyphens. Additionally, tokenization can help establish the credibility of the documents being analyzed (Vijayarani and Jahani, 2016).
"Stop words removal" refers to the practice of removing words from a piece of writing that have little or no significance in order to preserve the words with the most meaning and context. Examining the frequency of solitary tokens in a large corpus of text is the most common method for identifying stop words. Stop words include words such as "a," "the," "and," and others. Every discipline and language has its own collection of stop words; therefore, there is no exhaustive list of all stop words (Sarkar, 2019). Moreover, in certain scenarios like sentiment analysis, specific words like "no" or "not" that have a negative connotation cannot be eliminated from the text since it affects the intensity of the sentiment being conveyed (Kaczmarek et al., 2021).
To fully comprehend the stemming process, we must first identify the word stems pertaining to stemming. Stems and affixes combine to form morphemes; adding affixes to word stems modifies their meaning or creates a new term; attaching affixes to word stems generates new words. This process is called inflection. Stemming is obtaining the base form of a word from its inflected form. Stemming enables various applications, including text classification, text clustering, and information retrieval, by standardizing words to their base stem regardless of their inflections. Words such as "JUMPS," "JUMPED," and "JUMPING," for instance, are created by adding affixes to the word "JUMP." This is the word stem for the base word "JUMP" (Sarkar, 2019).
"Part of speech (POS) tagging" is a process in natural language processing where words are labeled with their respective part of speech, such as nouns, verbs, adjectives, etc. It is a crucial aspect of natural language processing and serves as the basis for named entity resolution, query resolution, and word sense disambiguation (Kulkarni and Shivananda, 2021). POS extracts text features by utilizing the network's capacity for self-encoding without taking linguistic expertise into account (Xiao et al., 2022).
Step 3. Opinion mining
3.2.1 Comparative opinion mining
Comparative opinion mining is a specific area within the broader field of opinion mining. Its main purpose is to extract comparative information and relationships from text data. Comparative opinion mining focuses on identifying and analyzing comparative forms and expressions in text, which can help in understanding the relationships and opinions expressed in the text (Guo et al., 2022). Comparative opinion mining is extremely significant when attempting to assess something, as it provides a basis for comparison (Varathan et al., 2017). (Xu et al., 2011) developed a model based on mobile phone data sets. This model was used by (Nasiri and Shokouhyar, 2021) to compare the functionality of refurbished and new smartphones on Amazon and Backmarket with the following comparative relationship:
R (P1, P2, A, S)
In this research, we utilize comparative opinion mining to compare refurbished and brand-new smartphones in developed and developing countries. The products' names are represented by P1 and P2, while A denotes the characteristic name and S represents the sentimental phrase taken from customers' reviews. However, in this study, we exclude the sentimental phrases and only focus on comparing the two types of smartphones. Consumers of refurbished smartphones in both developed and developing countries believe that their functionality is identical to that of new phones. Thus, refurbished smartphones appear brand new.
Entity and keyword attributes are both required for comparative relationship extraction. In this instance, we compare two distinct products: new and refurbished. We used the keyword set in both extraction tasks as the attribute set for machine learning (ML). In this particular instance, the SVM learner produces the best results. Then, irrelevant sentences were removed to improve classification results.
3.2.2 Non-comparative opinion mining
Aspects of customer satisfaction were extracted from online reviews to model customer satisfaction. Then, sentiment orientation analysis of reviews was conducted on each customer satisfaction dimension based on the SVM learner. The process of extracting customer satisfaction dimensions (CSDs) from online reviews includes two steps, (1) extraction of CSD based on LDA and (2) sentiment orientation. Detailed descriptions of each step's particulars are provided below.
• Extraction of CSD based on LDA
This section provides a brief introduction to Latent Dirichlet Allocation (LDA). LDA is an unsupervised learning technique that does not require labeled data, making it adaptable to different languages and domains. The model automatically generates semantically related word clusters to represent each topic. Unlike many comparative opinion mining tools, the LDA model exclusively uses text data and provides more in-depth information. LDA is a reliable and interpretable approach (Guo et al., 2022). Consequently, we used LDA in our non-comparative analysis due to its reliability.
The LDA model treats each online review as a probability distribution consisting of a blend of subjects and phrases. To form a review, one must repeatedly choose a topic based on its distribution and then choose words for that topic based on the probability distribution of words associated with it (Blei et al., 2003). After conducting online reviews, we calculated the probability of each phrase using the probability formula proposed by (Nasiri and Shokouhyar, 2021):
$$p(\text{word }\mid \text{ review })=\sum _{\text{topic }} p(\text{ word }\mid \text{ topic })\times p(\text{ topic }\mid \text{ review })$$
This probability formula can be expressed as a matrix, "Review-Word" matrix is the frequency of a word in a review, the "Review-Topic" matrix is the probability of a topic in a review, and the "Topic-Word" matrix is the probability of a word in a topic. The "Review- Word" matrix is denoted as\({x}_{m\times n}\), where \(M\) is the number of online reviews and \(N\) is the number of words in all the documents. The trained LDA model produces the "Review-Topic," "Topic-Word," and topic lists. Using the LDA generation process, the LDA model can be trained using the \({x}_{m\times n}\) acquired from the preceding formula matrix. Since we have used Twitter API to extract opinions from developed and developing countries. To evaluate the features of the refurbished smartphones themselves, we have removed all comments relating to fraud, shipping, Service and behavior of sellers. Then, in order to obtain more reliable results, we manually filtered out noisy words.
After analyzing the initial topics generated by LDA, the authors introduce a new method called SKP-LDA which incorporates sentiment word co-occurrence and knowledge pair feature extraction algorithms. They also create a word bag that takes into account the co-occurrence of emotive words, which helps to capture the emotional polarity of short texts such as microblogs. Furthermore, the SKP-LDA model utilizes topic-specific words and topic-related words as knowledge pairs that are added to the LDA model to improve its accuracy and effectiveness (Wu et al., 2021). Thus, semantic information can be derived with greater precision. Finally, we labeled each topic that can be considered for CSD for online reviews in developed and developing countries.
Our objective was to utilize SVM to identify the sentiment orientations of reviews related to each CSD. We employed the approach recommended by (Bi et al., 2019). In the set of online reviews \(R\), \({R}_{i}=\left\{{r}_{i1}^{l},\dots ,{r}_{ic}^{k},\dots ,{r}_{i{C}_{i}}^{g}\right\}\)denotes the set of reviews for the \(i\)th CSD, where \({r}_{ic}^{k}\) represents the \(c\)th review in \({R}_{i}\), and \(R\),\({C}_{i}\)denotes the total number of reviews for the \(i\)th CSD, \(l,k,g\in \{1, 2, . . . ,\text{M}\}\).To derive \({R}_{i}\), we classified the online reviews in set \(R\) into sentences based on punctuations. By extracting sentences containing \({\text{w}\text{o}\text{r}\text{d}}_{ij}\)in\(R\), \(i=\text{1,2},\dots ,I\),\(j=\text{1,2},\dots ,{J}_{i}\), \({R}_{i}\) can be obtained according to \({f}_{i}=\left\{{\text{w}\text{o}\text{r}\text{d}}_{i1},{\text{w}\text{o}\text{r}\text{d}}_{i2},\dots ,{\text{w}\text{o}\text{r}\text{d}}_{iJ}\right\}.\)If multiple sentences in a review relating to the same CSD, they can be consolidated.
Each \({R}_{i}\) review's sentiment is determined using SVM, a cutting-edge machine-learning technique for sentiment classification. Two phases are required for SVM sentiment classification: (I) feature construction using a bag-of-words model and (ii) training using tagged reviews. Each \({R}_{i}\) review's sentiment can be identified using the SVM sentiment classifier. The sentiment orientation of \({f}_{i}\) (the \(i\)th CSD) in each online review\({r}_{m}\) can be calculated based on the sentiment orientation of each review in \({R}_{i}\), where\(i=\text{1,2},\dots ,I\), \(m=\text{1,2},\dots ,M\). As demonstrated in Table 1, we represent the \({f}_{i}\)'s sentiment orientation in the online review \({r}_{m}\) where ∗ \(\in \left\{\text{P}\text{o}\text{s},\text{N}\text{e}\text{g}\right\}\),\(i=\text{1,2},\dots ,I\), \(m=\text{1,2},\dots ,M\). The nominally coded data can be transformed into structured data using the equation below, as illustrated in Table 1. From Table 1, it is evident that if the sentiment orientation of \({f}_{i}\) in the online review \({r}_{m}\) is positive,\({S}_{im}^{Pos}=1\) and \({S}_{im}^{Neg}=0\); if the sentiment orientation is negative, \({S}_{im}^{Pos}=0\) and \({S}_{im}^{Neg}=1\); and if the sentiment orientation is missing, \({S}_{im}^{Pos}=0\)and \({S}_{im}^{Neg}=0\),
$$i=\text{1,2},\dots ,I, m=\text{1,2},\dots ,M.$$
$${S}_{im}^{*}=\left\{\begin{array}{c}1,\hspace{0.25em}\hspace{0.25em}\hspace{0.25em}\hspace{0.25em}\text{ }\text{i}\text{f}\text{ }\text{t}\text{h}\text{e}\text{ }\text{s}\text{e}\text{n}\text{t}\text{i}\text{m}\text{e}\text{n}\text{t}\text{ }\text{o}\text{r}\text{i}\text{e}\text{n}\text{t}\text{a}\text{t}\text{i}\text{o}\text{n}\text{ }\text{i}\text{s}\text{*}\text{ }\\ 0,\hspace{0.25em}\hspace{0.25em}\hspace{0.25em}\hspace{0.25em}\text{ }\text{o}\text{t}\text{h}\text{e}\text{r}\text{w}\text{i}\text{s}\text{e}\text{ }\end{array}i=\text{1,2},\dots ,I,m=\text{1,2},\dots ,M\right.$$
Table 1
Data on the structure of online reviews.
| CSDs |
\({f}_{1}\) | \({f}_{2}\) | \(\cdots\) | \({f}_{7}\) |
Online review | \({S}_{1}^{\text{P}\text{o}\text{s}}\) | \({S}_{1}^{\text{N}\text{e}\text{g}}\) | \({S}_{2}^{\text{P}\text{o}\text{s}}\) | \({S}_{2}^{\text{N}\text{e}\text{g}}\) | \(\cdots\) | \({S}_{7}^{\text{P}\text{o}\text{s}}\) | \({S}_{7}^{\text{N}\text{e}\text{g}}\) |
\({\varvec{R}}_{1}\) | 0 | 1 | 0 | 0 | \(\cdots\) | 1 | 0 |
\({\varvec{R}}_{2}\) | 1 | 0 | 0 | 0 | \(\cdots\) | 0 | 0 |
\(\cdots\) | \(\cdots\) | \(\cdots\) | \(\cdots\) | \(\cdots\) |
\({\varvec{R}}_{\varvec{M}}\) | 0 | 0 | 1 | 0 | \(\cdots\) | 0 | 1 |