0% found this document useful (0 votes)
21 views8 pages

ISSS609 Project Proposal Group 7

The project proposal outlines the development of a sentiment analysis tool aimed at understanding consumer perceptions of Sephora's products through user-generated content. It involves data collection from Sephora's website, preprocessing of reviews, and the application of various machine learning and deep learning models for sentiment classification. The project aims to provide actionable insights for marketing, customer service, and product development teams to enhance their strategies based on consumer feedback.

Uploaded by

ayushi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views8 pages

ISSS609 Project Proposal Group 7

The project proposal outlines the development of a sentiment analysis tool aimed at understanding consumer perceptions of Sephora's products through user-generated content. It involves data collection from Sephora's website, preprocessing of reviews, and the application of various machine learning and deep learning models for sentiment classification. The project aims to provide actionable insights for marketing, customer service, and product development teams to enhance their strategies based on consumer feedback.

Uploaded by

ayushi
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

SMU Classification: Restricted

ISSS609
Project Proposal

“Analyzing Beauty: Insights into Sephora


Feedback via Sentiment Analysis”
Group 7
Alvin LIM Li Xian
Ayushi SHAKYA
Debanjan DATTA
Junzhe Huang
Neha GOYAL
NGUYEN Thuy Hanh Duyen

29 Sep 2024
SMU Classification: Restricted

1. Introduction
With social media platforms and e-commerce websites being highly prevalent among all
consumer segments, businesses now have access to a wealth of user-generated content that
offers valuable insights into their sentiment. This project aims to develop a sentiment analysis tool
that categorizes and processes the sentiments behind user reviews, helping a popular e-
commerce platform Sephora understand how their brands and products are perceived. The tool
will enable businesses to respond more effectively to customer feedback by classifying
sentiments in captions and comments as positive, negative, or neutral.
Project Components:

• Data Sources: The dataset (downloaded from Kaggle) was collected via Python scraper
from Sephora US website (March 2023) and contains 2 data tables, “Products” and
“review”. (See Appendix 1)
• Challenges: User reviews content often includes informal language, slang, and emojis,
making natural language processing (NLP) essential for cleaning and normalizing the data.
The language this model will be trained on is English, it will not be able to analyze mixed
language posts and comments.
• Machine Learning Models: A mix of traditional machine learning techniques and deep
learning architectures will be applied for sentiment classification.
Benefits for different business functions include:

• Marketing teams can leverage consumer sentiment insights to fine-tune campaigns and
better target audiences
• Customer service teams can respond quickly to negative sentiments, addressing concerns
• Product development teams can gather feedback on product features and identify areas
for improvement and new value propositions
By providing insights into consumer opinions, this tool allows businesses to track and adjust their
strategies, addressing customer concerns and improving overall product satisfaction. Staying
attuned to consumer sentiment can help businesses remain responsive, customer-centric, and
competitive.

2. Proposed Methodology
Analyzing user reviews poses various challenges for sentiment analysis, especially because of
the subtle language nuances often found in the data. This makes it difficult to detect ambiguity,
sarcasm, or irony, given that this heavily depends on context which models may not capture
explicitly. Moreover, traditional models often struggle with contextual understanding, leading to
misclassification of sentiments. Data quality such as comments containing noise, slang,
abbreviations, and emojis can affect model performance. Additionally, comments may be written
in languages other than English, which might not be supported by the models. Training Large
Language Models (LLMs) like BERT [2] and GPT [3] requires significant computational resources.
SMU Classification: Restricted

This project aims to develop and compare sentiment analysis models using traditional machine
learning techniques, deep learning architectures, and Large Language Models (LLMs). The
methodology is structured into three core steps: data collection and preprocessing, model
development, and comparison and evaluation. Figure 1 shows the flowchart for the proposed
methodology [1]:

Figure 1: Flowchart for sentiment analysis

Data Collection and Preprocessing

We will utilize datasets of user reviews and product information. Data preprocessing will remove
noise such as punctuation, stop words, emojis, slang, and abbreviations by cleaning, normalizing
words (stemming or lemmatization), and tokenizing the text. After preprocessing, methods like
word embeddings will be used to capture semantic relationships. The cleaned dataset will be split
into training and testing sets using an 80-20 split.

Model Development
To desmonstrate different approaches to sentiment classification, both traditional machine
learning models and LLM-based models will be implemented. Feature extraction techniques will
be employed to transform text data into numerical representations for traditional machine learning
models algorithms. This will facilitate an assessment of how engineered textual features impact
model performance. Deep learning models, including CNN and transformers like BERT and GPT,
will be used for their ability to capture complex contextual relationships. Comparing these
methods will reveal the trade-offs between computational efficiency and capturing nuanced
sentiment.

Comparison & Evaluation:


The models will be evaluated using a standard set of performance metrics including accuracy,
precision, recall, F1-score, and confusion matrix analysis. The goal is to identify the strengths and
weaknesses of each approach based on their performance on the same test data.
SMU Classification: Restricted

3. Solution Details

Data Pre-processing
Data cleaning and pre-processing such as stop word removal, stemming, lemmatization and
tokenization will be done before being used for model training.

Classic Machine Learning Models


Classic machine learning models are simple lightweight model which can be used as a baseline
model to assess the performance of other more sophisticated models.

Feature Extraction:

Classic machine learning models perform better with numbers of inputs compared to text inputs.
We will explore multiple methods for converting text data into numerical representations:

a. TF-IDF (Term Frequency-Inverse Document Frequency): A statistical measure to


evaluate how important a word is within a document relative to the corpus.
b. Word Embeddings: Using pre-trained word embeddings such as Word2Vec or
GloVe to convert words into vectors that capture semantic meaning.

Model Selection:

Logistic Regression (LR): The LR model is a simple classification model which we will use as
a baseline model. The LR model will be trained on the preprocessed text data to classify the
sentiments of each text. We will explore using 1. a binary classifier to classify posts into positive
or negative, and 2. a multiclass classifier to classify posts into positive, neutral or negative.

Support Vector Machine (SVM): SVM is a more powerful classifier that works well with high-
dimensional data (i.e. its better able to accept a larger corpus than LR). Similar to LR, the SVM
model will be trained using the preprocessed text data and classify the sentiment of the text into
positive and negative sentiment. In addition, we will explore the use of the One vs All approach
to do a 3-class classification with SVM.

Deep Learning Model

Deep Learning models replicates artificial neural networks to learn the underlying patterns from
large amount of data. They are capable of handling more complex data compared to classic
machine learning models and are therefore less reliant on data pre-processing.

Model Selection:

Convolutional Neural Networks (CNN): CNNs are traditionally used for image recognition
tasks but we can use a 1-Dimension CNN models for the purpose of text sentiment analysis.
SMU Classification: Restricted

The CNN model will be trained on the tokenized data to do a 3-class classification of the
sentiment.

Transformers: Transformers are neural networks that uses self-attention mechanisms to


process text, allowing it to better capture the context of each sentence. They are the state-of-
the-art model for sentiment analysis.

We will explore using various transformers model such as BERT (Bidirectional Encoder
Representations from Transformers) [2] and GPT (Generative Pretrained Transformer) [3] to
perform sentiment analysis and compare their performance.

Model Advantage Limitation


LR • Lightweight and fast • Not able to fully capture context
• Predictions are easy to interpret • Performance is highly dependent
and explain on data preprocessing and
feature extraction
SVM • Requires more computation • Not able to fully capture context
resource than LR but less than • Performance is highly dependent
deep learning models on data preprocessing and
• Performs better than LR with feature extraction
complex data such as social
media post
CNN • CNN is better at capturing context • Requires more computation
than classic machine learning resource for training
models
• Less data pre-processing is
required
Transformers • Expected to have the best • Most computationally intensive to
performance train and fine-tune
• Able to process complex data,
highly capable of understanding
textual context

4. Proposed Experiments

For the sentiment analysis experiment, we plan to create a systematic process for data
preparation, model training, and evaluation.
Our first step will involve text preprocessing, where tokenization is performed first to break down
the text into individual words or tokens. Following this, we will remove noise such as
punctuation, emojis, and special characters. Stop-word removal will come next, eliminating
common words that do not add significant meaning to the text. Finally, we will apply stemming
or lemmatization to normalize words to their root forms, ensuring consistency across the
SMU Classification: Restricted

dataset. This process is critical for reducing noise and improving the overall performance of the
sentiment analysis models, as detailed in our methodology.
Next, we will partition the data into three sets: 80% for training, 10% for validation, and 10% for
testing. This will allow us to fine-tune the model using the validation set and evaluate the
model's generalization using the test set.
For the experiment, we will use Python, leveraging key libraries such as Pandas for data handling,
NLTK and spaCy for text preprocessing, and Scikit-learn for traditional machine learning models.
Advanced deep learning models will be implemented using frameworks like TensorFlow and
PyTorch, ensuring efficient model training on GPU resources available through Google Colab.
Our approach involves running two phases of experiments:
1. Phase 1: We will start with traditional machine learning models, such as Logistic
Regression (LR) and Support Vector Machine (SVM). These will be used as baselines to
classify sentiments using TF-IDF and Word2Vec for feature extraction.
2. Phase 2: We will then advance to more sophisticated models like Convolutional Neural
Networks (CNN) and transformers (such as BERT and GPT-3) to capture complex
relationships between words and improve contextual understanding. Fine-tuning will be
conducted for the transformers to optimize performance.
Evaluation metrics will include accuracy, precision, recall, and F1-score to compare the
performance of traditional models against deep learning models.
This step-by-step approach allows us to assess the efficiency and accuracy of various models,
ultimately determining the most effective method for social media sentiment analysis.

5. Project Schedule and Work Division


• Week 2-3: Literature review and dataset selection
Responsible: <Neha Goyal> ,<Ayushi Shakya>
• Week 4: Data cleaning, feature extraction, and data partitioning
Responsible: <Debanjan Datta>
• Week 5-6: Model training for traditional machine learning models
Responsible: <Ayushi Shakya>
• Week 7-8: Fine-tuning and evaluation of LLM models
Responsible: <Alvin Lim>, < Nguyen Thuy Hanh Duyen>
• Week 9: Model comparison, analysis, and final report preparation
Responsible: <Neha Goyal>, <Junzhe Huang>
SMU Classification: Restricted

References
[1] Arun, K. & Srinagesh, Ayyagari. (2020). Multi-lingual Twitter sentiment analysis using machine
learning. International Journal of Electrical and Computer Engineering (IJECE). 10. 5992.
10.11591/ijece.v10i6.pp5992-6000.
[2] Devlin, J., Chang, M., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of Deep
Bidirectional Transformers for Language Understanding. North American Chapter of the
Association for Computational Linguistics.
[3] Radford, A. (2018). Improving language understanding by generative pre-training.
SMU Classification: Restricted

Appendix I – Data Table Schema


The dataset consists of two data tables:

- “Products” containing information about all beauty products (over 8,000) from the
Sephora online store, including product and brand names, prices, ingredients, ratings,
and all features.

- “review” includes user reviews (about 1 million on over 2,000 products) of all products
from the Skincare category, including user appearances, and review ratings by other
users

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy