Leading academics reveal that Generative AI causes unintended changes in sentiment

TL;DR:

  • Leading academics from the Gillmore Centre for Financial Technology reveal that Generative AI and Large Language Models (LLMs) unintentionally alter the sentiment of original text.
  • Their research highlights that the widespread adoption of LLMs fundamentally transforms the linguistic features of content, primarily shifting sentiment towards increased neutrality.
  • The study, involving the examination of 50,000 tweets, demonstrates that LLMs tend to move both positive and negative sentiments closer to neutrality, impacting sentiment analysis.
  • Potential biases arise from LLMs’ application in paraphrasing, rewriting, and content creation, calling for mitigation strategies to enhance reliability in user-generated content (UGC).
  • Dr. Yi Ding emphasizes the growing use of Generative AI, with millions of users worldwide and businesses leveraging its potential as a valuable tool.

Main AI News:

In the realm of Generative AI and the deployment of Large Language Models (LLMs), a profound revelation has emerged. Leading academics from the Gillmore Centre for Financial Technology at Warwick Business School have uncovered a disconcerting phenomenon—the unintended transformation of sentiment in the original text. This revelation comes to us courtesy of a paper titled “Who’s Speaking, Machine or Man? How Generative AI Distorts Human Sentiment,” authored by eminent scholars from the Gillmore Centre for Financial Technology. Their research delves deep into the impact of the burgeoning presence of LLMs on public sentiment, leading to a stark conclusion: modifications introduced by LLMs render existing outcomes unreliable.

This ground-breaking research, based on the replication and adaptation of well-established experiments, assumes a pivotal role in expanding the discourse on Generative AI and user-generated content (UGC). It underscores that the widespread integration of LLMs fundamentally alters the linguistic characteristics of any given text.

This startling phenomenon came to light through a comprehensive analysis encompassing the examination of a staggering 50,000 tweets. The researchers harnessed the formidable power of the GPT-4 model to rephrase the text, and the “Valence Aware Dictionary for Sentiment Reasoning” (VADER) methodology was employed to compare the original tweets with their GPT-4 rephrased counterparts. The outcome was as startling as it was consistent—LLMs predominantly steer sentiment towards heightened neutrality, effectively steering text away from both positive and negative orientations.

Ashkan Eshghi, the Houlden Fellow at the Gillmore Centre for Financial Technology, weighed in on these findings, stating, “Our research reveals a discernible shift towards neutral sentiment in LLM-rephrased content compared to the original human-generated text. This transformation affects both positive and negative sentiments, ultimately diminishing the variability in content sentiment. While LLMs do tend to nudge positive sentiments closer to neutrality, the shift in negative sentiments towards a neutral position is more pronounced. This overarching shift towards positivity could have significant implications for the application of LLMs in sentiment analysis.”

Ram Gopal, the Director of the Gillmore Centre for Financial Technology, voiced a pertinent concern: “A vast body of literature already exists on the multifaceted use of UGC, spanning from predicting stock prices to evaluating service quality. However, we’ve discovered that the extensive adoption of LLMs introduces a significant concern—potential bias. This bias emanates from the utilization of LLMs in tasks such as paraphrasing, rewriting, and even content creation, leading to sentiments that may diverge from those expressed by individuals without the involvement of LLMs.”

In response,” Gopal continues, “our research proposes a mitigation strategy aimed at reducing bias and bolstering the reliability of UGC. This entails predicting or estimating the sentiment of original tweets by analyzing the sentiments of their rephrased counterparts.”

Yet, the question remains: Would other linguistic attributes of UGC undergo a transformation if AI were employed? Emotion, sentence structure, and the proportion of specific words in a sentence are among the myriad aspects that warrant further investigation.

Dr. Yi Ding, Assistant Professor of Information Systems at the Gillmore Centre for Financial Technology, sheds light on the broader context: “We observe an ever-expanding user base, with approximately 180 million OpenAI monthly active users globally, and an increasing number of businesses embracing AI’s potential as a business tool. Conducting this study, which juxtaposes Generative AI with human sentiment, stands as a pivotal milestone in the ongoing evolution of LLMs. It promises to enhance output, rectify biases, and elevate efficiency for all who harness its capabilities.

The academic luminaries intend to employ additional predictive models to infer authentic human sentiments and chart further mitigation strategies in forthcoming research endeavors.

Conclusion:

This research underscores the profound implications of Generative AI’s influence on sentiment. As LLMs increasingly shape the linguistic characteristics of content towards neutrality, businesses must be cautious when employing them for tasks like sentiment analysis. Mitigation strategies are necessary to rectify potential biases and enhance the reliability of user-generated content (UGC). The expanding use of Generative AI in the market requires a thoughtful approach to ensure accurate and unbiased insights.

Source