Definition
Sentiment analysis assigns a polarity label — typically positive, negative, or neutral — to a piece of text. Many systems also output an intensity score, a confidence value, and sometimes an emotion label (anger, frustration, joy). It is one of the oldest natural-language-processing tasks and the most commonly requested capability in any feedback tool.
The methods form a progression. Keyword and lexicon systems count positive and negative terms. Classical machine-learning models (logistic regression, SVMs) train on labeled datasets and learn term weights. Deep-learning models (LSTMs, CNNs) learn richer patterns. Modern systems use pretrained transformer models — BERT derivatives or general LLMs — either fine-tuned on domain data or prompted zero-shot. Most production feedback tools run a hybrid: a fast classifier on every record plus LLM-based classification on ambiguous or business-critical text. The hybrid pattern keeps per-record cost low while preserving accuracy where it matters — on the subset of reviews a product manager actually reads.
Why it matters
Sentiment is the first-pass filter that makes a large feedback corpus navigable. A million reviews is unreadable; the same corpus bucketed by sentiment, SKU, and theme is a working queue. For product teams, sentiment lets them triage: read the negative first on a launch SKU, track positive on a messaging test, set alerts on sentiment drops. For CX, it routes the urgent; for QA, it flags the emerging.
The pitfalls for product reviews are real. Sarcasm ("works great — if you like replacing batteries weekly") flips naive classifiers. Aspect conflicts are the norm: one review can be positive on battery life, negative on the screen, and neutral on packaging. Domain drift means a model trained on general reviews underperforms on appliances or cosmetics. Serious feedback work pairs sentiment with aspect-based sentiment analysis (ABSA) and thematic analysis, not sentiment alone.
Example
A beauty brand ingests 8,400 reviews on a moisturizer SKU across Amazon, Target, and Ulta's Bazaarvoice-powered page. A lexicon-only sentiment system rates the SKU at 72% positive. Indellia's sentiment model, paired with ABSA, rates overall positive at 68% but breaks it down: positive 91% on scent, 78% on packaging, 61% on texture, 42% on "breakout" mentions. The Consumer Insights team flags the texture and breakout aspects for the formulation team; the headline sentiment number alone would have buried both. Over the following six weeks, negative texture mentions track back to a thickener change with a single supplier — a finding the lexicon-only rollup would never have produced.