Online sentiment toward ChatGPT shows strong wave of skepticism
The predominance of negative sentiment aligns with early public reactions to transformative technologies, where concerns about privacy, job loss, misinformation and ethical boundaries often amplify skepticism. As ChatGPT gained popularity, debates intensified over its ability to produce human-like text, automate tasks and influence behavior across education, media and business sectors.
A large-scale machine learning investigation has found that public sentiment toward ChatGPT on the social media platform X (formerly Twitter) was dominated by negative opinions during the study period, revealing global skepticism toward the rapidly evolving artificial intelligence system.
The peer-reviewed article, “Analyzing Global Attitudes Towards ChatGPT via Ensemble Learning on X (Twitter),” published in Algorithms, examines more than 219,000 tweets and uses an enhanced ensemble machine-learning model to classify sentiment into positive, negative and neutral categories. The paper presents a three-class categorization approach designed to better capture the nuances of public conversation surrounding the AI tool.
The findings reveal that the majority of tweets expressed negative sentiment, a result that suggests unresolved public concerns around AI use, safety, transparency and long-term impact. The research goes further by evaluating how different machine learning models interpret these sentiments, ultimately demonstrating that ensemble methods provide superior accuracy when analyzing large volumes of diverse social media opinions.
What does global public opinion toward ChatGPT look like?
The study focuses on determining the global mood surrounding ChatGPT by examining posts from users across many regions. To answer this, the researchers built a large dataset of English-language tweets over a one-month period, capturing a wide variety of reactions ranging from enthusiasm about ChatGPT’s capabilities to concerns about misinformation, job displacement and ethical boundaries.
The data collection process produced a dataset containing more than 219,000 tweets, offering substantial coverage of public discourse from casual users, technical professionals, educators and AI observers. After comprehensive data cleaning, which removed emojis, links, hashtags, special characters and other nonessential symbols, the researchers applied tokenization and vectorization techniques to prepare the text for machine-learning processing.
Once the dataset was refined, the sentiment distribution became clear: negative tweets dominated. Nearly half of all collected messages expressed unfavorable views of ChatGPT, while the remaining portion was split between positive sentiment and more reserved, neutral attitudes. The heavy concentration of negative posts is a critical finding, especially given ChatGPT’s widespread adoption and prominence.
The study highlights that neutral sentiment is an essential category that many previous analyses ignored. By including a third classification beyond simple positive–negative polarity, the research captures the blended reactions common in discussions about emerging technologies. Neutral posts often encapsulated observations, troubleshooting comments or ambivalent reactions that lacked the emotional intensity seen in the other categories.
Through this lens, the sentiment analysis reveals a polarized but mostly cautious global audience. The pattern reflects ongoing public debate over the promises and risks of advanced AI models, as well as the rapid speed at which ChatGPT integrated into both everyday life and professional environments.
Which machine-learning approaches most accurately capture public sentiment?
Next up, the study examines which machine-learning algorithms best capture sentiment patterns in large-scale social media data. To test performance, the authors compared four approaches: Naïve Bayes, Support Vector Machine (SVM), Random Forest and an Ensemble Learning method combining all three.
Each model was trained using a 75–25 split between training and testing data. Before training, the researchers applied random oversampling to correct the imbalance between negative, positive and neutral tweet categories. This balancing step was essential to avoid skewed model predictions, especially since negative tweets significantly outnumbered the others.
The evaluation produced clear results: the Ensemble Learning method achieved the highest overall accuracy at 86 percent. The combined model outperformed SVM at 84 percent, Random Forest at 79 percent and Naïve Bayes at 66 percent. The ensemble method also showed superior recall and F1-score, particularly in identifying neutral tweets, which tend to be the most challenging to classify due to their lack of strong emotional indicators.
The study explains why an ensemble approach works better than individual models. Naïve Bayes is effective at identifying general word patterns but struggles with more nuanced expressions. Support Vector Machine performs strongly in high-dimensional classification tasks but requires additional calibration to handle probability outputs. Random Forest captures nonlinear relationships but may misinterpret subtle linguistic cues. By combining the strengths of all three, the ensemble model achieves a more balanced and precise interpretation of sentiment.
Additionally, performance curves showed the ensemble model had better discriminatory power, confirming its ability to differentiate between nuanced tweet types. This result underscores the importance of hybrid-learning systems in sentiment analysis—particularly when analyzing fast-moving, chaotic environments like social media.
The research provides a compelling demonstration that ensemble methods represent a more reliable approach for analyzing large, real-world datasets, especially when sentiment categories are unevenly distributed. This has implications for policy makers, companies and institutions seeking to understand public attitudes toward new technologies.
Why does public opinion toward ChatGPT lean negative?
Although the researchers do not conduct qualitative analysis of individual tweets, the large-scale trend suggests a collective unease around the expanding role of AI in society.
The predominance of negative sentiment aligns with early public reactions to transformative technologies, where concerns about privacy, job loss, misinformation and ethical boundaries often amplify skepticism. As ChatGPT gained popularity, debates intensified over its ability to produce human-like text, automate tasks and influence behavior across education, media and business sectors.
The study indicates that negative sentiment may signal broader discomfort with the pace of AI development. Users expressed specific worries around the reliability of AI outputs, the potential for harmful usage and uncertainties about long-term societal impact. This pattern demonstrates that public trust in generative AI remains fragile, despite widespread usage.
The research also points out that neutral sentiment plays a significant role in shaping overall discourse. Many tweets offered observations or inquiries without strong emotional reaction. This suggests a segment of users is still evaluating the technology, forming opinions based on evolving experiences rather than predetermined stances.
The implications for developers and policy makers are substantial. The authors argue that understanding global sentiment is essential for shaping ethical guidelines, communication strategies and transparency efforts. If negative attitudes remain dominant, AI developers may face resistance in deploying new features or expanding integration into sensitive sectors such as education, healthcare or finance.
The study also calls attention to the limitations of TF-IDF vectorization and traditional machine-learning techniques in capturing deeper semantic signals, including sarcasm or cultural nuance, which are common on Twitter. As a result, the authors recommend future research using contextual embeddings like BERT or RoBERTa to achieve richer sentiment comprehension.
- FIRST PUBLISHED IN:
- Devdiscourse

