Pazarlama Araştırmaları

Analyzing Brand Perception in LLMs

Research Questions

  1. To what extent do LLMs exhibit bias toward specific brands?
  2. Which metrics can be used to measure this bias?
  3. How can an appropriate brand comparison test be designed?
  4. What results emerge when these metrics are applied across different LLMs and brands?

Results

  • All LLMs showed consistent (transitive) brand preferences, producing the ordering Apple > Samsung > Huawei.
  • ChatGPT-4 was the most decisive model, whereas Gemma produced the most uncertain responses.
  • ChatGPT models displayed strong positive bias toward Apple and Samsung, while Gemma showed weaker positive bias and higher uncertainty.
  • The methodology can be applied not only to brands but also to a wide range of human-centered concepts.

Findings

  • Preference Dimension:

    • All models demonstrated logical consistency; the aggregate preference ordering was: Apple > Samsung > Huawei
  • Sentiment Dimension:

    • Gemma was more consistent but exhibited lower bias.
    • ChatGPT-3.5 and ChatGPT-4 showed strong positive bias toward Apple and Samsung (e.g., 93.6% positive response rate for Apple).
    • Responses toward Huawei varied significantly across models.
  • Consistency Metrics:

    • ChatGPT models: low entropy, high decisiveness.
    • Gemma: entropy levels 4–5× higher, indicating greater uncertainty.
    • Overall alignment across question formats; Gemma showed the largest deviations.
  • LLM Models: 5

  • Synthetic Data: 1

  • Method: 5

  • Speed: 2

  • Ethics: 1

  • Accuracy: 4

  • Demographics: 0

If you would like to access more detailed information about this article, click here to view the supplementary material.

2 min read

Related Articles