Tackling messaging scams with generative AI

As scams become more sophisticated, technologies like ScamGPT-J will play a crucial role in levelling the playing field. This study not only introduces a powerful anti-scam tool but also sparks a broader conversation about how AI can be used to protect individuals in an increasingly deceptive digital landscape.


CO-EDP, VisionRICO-EDP, VisionRI | Updated: 03-01-2025 09:28 IST | Created: 03-01-2025 09:28 IST
Tackling messaging scams with generative AI
Representative Image. Credit: ChatGPT

Messaging scams are a growing menace, costing billions globally each year and preying on unsuspecting individuals through increasingly sophisticated tactics. These scams exploit psychological vulnerabilities through deceptive conversations, luring victims into financial losses or identity theft. To address this challenge, researchers from the National University of Singapore have developed an innovative solution called ScamGPT-J. Their study, ScamGPT-J: Inside the Scammer’s Mind, A Generative AI-Based Approach Toward Combating Messaging Scams, presented at the International Conference on Information Systems (ICIS) 2024, outlines a breakthrough in using generative AI to prevent messaging scams by helping users identify deceptive interactions.

Tackling messaging scams with generative AI

Messaging scams are particularly effective because they exploit trust, emotions, and cognitive biases. Scammers often create a sense of urgency, appeal to authority, or play on emotions like fear or love to manipulate their victims. The conversational nature of these scams makes them more believable and harder to detect than static phishing emails. For example, a love scam might begin with weeks of trust-building before transitioning into financial extortion. Existing anti-scam technologies, which rely on detecting keywords or patterns, often miss such nuanced tactics.

ScamGPT-J addresses these challenges by focusing on the entire conversational flow rather than isolated messages. The tool helps users see through the manipulative strategies used by scammers, providing a deeper understanding of how scams unfold and empowering them to disengage before falling victim.

At the heart of ScamGPT-J is a fine-tuned generative AI model designed to replicate scammer behaviours. Unlike detection systems that simply flag potential threats, ScamGPT-J predicts how a scammer might respond in a given conversation. Users can input suspicious messages, and the tool generates scam-like replies based on the context, allowing users to compare these responses with the actual conversation they are having. This real-time simulation makes the scammer’s tactics more transparent, helping users identify red flags they might otherwise overlook.

To develop this tool, the researchers created a dataset of 902 synthetic scam conversations across four scam categories: authority scams, job scams, love scams, and investment scams. The dataset was designed to capture the subtleties of scammer interactions, using AI to model both generic and specific tactics. This innovative use of synthetic data ensures that ScamGPT-J can address a wide range of scam scenarios, from tax fraud to fake job offers.

During rigorous testing, ScamGPT-J demonstrated remarkable accuracy in replicating scammer-like behavior. In simulated evaluations, the tool outperformed general-purpose language models like GPT-J by a significant margin. For example, ScamGPT-J provided contextually appropriate scam responses in 14 out of 16 scenarios, compared to GPT-J’s 3 out of 19. Users also rated ScamGPT-J highly for its ability to identify scams, with an average usefulness score of 4.4 out of 5.

The tool’s focus on conversational context rather than individual messages sets it apart from existing scam detection methods. By analyzing the flow of dialogue, ScamGPT-J provides insights into the psychological tactics employed by scammers, such as trust-building and urgency creation. This makes it not just a detection tool but also an educational resource for users, teaching them how to recognize and avoid scams.

One of ScamGPT-J’s most impactful contributions is its ability to educate users about scam tactics. By simulating scammer interactions, the tool encourages critical thinking and helps individuals identify red flags in real time. For instance, users can learn how scammers exploit cognitive biases like the anchoring effect or authority bias to manipulate their targets. This proactive approach empowers users to make informed decisions, turning them from passive recipients into active participants in their digital safety.

Challenges and future directions

While ScamGPT-J is a significant advancement in scam prevention, it is not without limitations. Its reliance on synthetic data, while innovative, may not fully capture the complexity of real-world scams. Incorporating real scam conversations into the training dataset could further enhance its robustness. Additionally, the tool must address potential risks such as overreliance by users or false positives, which could lead to legitimate messages being misclassified as scams.

Looking ahead, the researchers propose expanding ScamGPT-J’s capabilities to include multi-lingual support and advanced scam detection techniques. They also highlight the importance of integrating such tools into larger cybersecurity ecosystems, allowing for seamless collaboration between users, technology providers, and regulatory bodies.

As scams become more sophisticated, technologies like ScamGPT-J will play a crucial role in levelling the playing field. This study not only introduces a powerful anti-scam tool but also sparks a broader conversation about how AI can be used to protect individuals in an increasingly deceptive digital landscape. With continued advancements and collaboration, ScamGPT-J could become a cornerstone of global efforts to combat the ever-evolving threat of messaging scams.

  • FIRST PUBLISHED IN:
  • Devdiscourse
Give Feedback