Blog
AI

From Deepfakes to Fake Identities: How Threat Actors Are Weaponising AI

CYJAX has identified a growing trend of AI-based tools being developed and shared across threat actor forums for fraudulent and cybercriminal purposes, highlighting the need for proactive monitoring and defence against evolving AI-driven threats.

October 20, 2025
13
min read
Cymon
Table of contents
Share

CYJAX has observed a rise in the number of alleged AI-based tools being developed across threat actor forums, whether for fraudulent or cyber purposes.

As seen below, an example of such a tool was developed and shared in a Russian-language forum to attract user feedback. In this case, CYJAX’s Russian-language specialists noted that the tool’s output failed to bypass Gemini. The tool, reportedly developed to create a “fake identity”, generates information including social media, job, and financial data. Whilst still in active development, this is a good example of what is currently being organically produced within threat actor circles.

Whilst CYJAX has observed examples of threat actors developing tools, we have also identified “key sellers”. Such key sellers offer keys to most, if not all, commercially available AI tools, often at a reduced cost compared to the RRP.

In terms of fraud methods related to AI, this is not typically offered, as most threat actors with access to commercially available AI tools prefer to keep their tooling useful for both onboarding and wider fraud. Historically, once a method becomes public (through purchase or public sharing), it becomes increasingly easy to detect. For this reason, CYJAX is seeing fewer threat actors offering methods around the use of AI, to protect their own operations.

That said, CYJAX analysts have found references in monitored chats. Below is a list of commercially available tools and their current applications in fraud.

  • Deepfake creation (face-swapping)
    Fraudsters use this tool to create fake videos or images by swapping faces, often to impersonate celebrities, public figures, or victims. This is typically used in impersonation scams such as fake endorsements or investment schemes.
  • AI-powered face-swapping app
    Used to create deepfakes of people’s faces on video clips, resulting in onboarding scams, impersonation scams, or fake promotional videos. Threat actors also use it to generate fake social media posts or video testimonials.
  • AI for generating synthetic images (faces, landscapes, etc.)
    Used to create realistic but fake profile pictures for fraudulent social media accounts, often employed in romance scams or to establish false credibility for investment schemes.
  • AI-driven natural language processing for generating text-based content
    Threat actors use ChatGPT to automate phishing emails, create fake reviews, write deceptive product descriptions, or manipulate users through social engineering in fake conversations.
  • AI-based image generation and art manipulation
    Used to generate fake advertisements, deceptive product images, or counterfeit promotional materials for fraudulent products such as fake art sales or scam offerings.
  • AI for creating and editing images, video, and audio
    Used to generate fake promotional videos or audio clips, such as forged celebrity endorsements or fake interviews, to add credibility to scams.
  • AI-based writing, voice synthesis, design, and deepfake creation
    Threat actors often hire freelancers using AI tools for services like creating fake reviews, fake voiceovers for scam phone calls, or manipulating media content (audio or video) for fraudulent campaigns.
  • AI-based video generation (synthetic avatars that speak in videos)
    Used to create fake spokesperson videos promoting fraudulent products, services, or investment opportunities. These synthetic videos appear realistic, making scams seem more credible.
  • AI for creating synthetic voices and voiceovers
    Used to clone voices of real people (such as friends, celebrities, or public figures) to create fake audio recordings for extortion or fraud (e.g., fake phone calls, blackmail).
  • AI to generate realistic images and videos
    Used to create fake product images (such as electronics or fashion items) that can be advertised as counterfeit goods on online marketplaces or fraudulent e-commerce sites.
  • Natural language processing and text generation
    Used to create fake social media posts, scam ads, or automated responses to manipulate users, promote fraudulent services, or gather personal information.
  • AI music generation
    Used to create custom music tracks or sound effects for scam videos, making them more convincing when promoting fake schemes or fraudulent services.
  • AI for image recognition and manipulation
    Used to create fake product images that are indistinguishable from legitimate ones, typically in scams involving counterfeit goods sold as authentic items.
  • Facial recognition AI
    Allows threat actors to track down individuals’ public photos across the web and use their likeness for identity theft, fake profiles, or impersonation scams on social media.
  • AI-driven copywriting tool
    Used to generate convincing scam emails, sales pages, or fake product descriptions, making fraudulent offers appear legitimate and well-crafted.
  • AI-based visual recognition for scanning QR codes
    Used to create fake QR codes that lead to phishing websites or malicious apps. When users scan these codes, they are redirected to fraudulent sites that steal personal information or money.
  • AI that helps with customer relationship management (CRM) and sales automation
    Used to create automated fake sales outreach, offering fake services or discounts on social media to gather sensitive personal information.
  • AI tool for transcribing speech to text
    Used to transcribe fake voice messages or phone calls from supposed authorities, encouraging users to act on fraudulent offers.
  • AI-powered content scheduling and automation for social media
    Used to automate fake social media posts and promotional campaigns for fake products, services, or schemes (e.g., fake cryptocurrency or online businesses).

Stay Ahead with CYJAX
As threat actors continue to experiment with AI for malicious purposes, understanding their methods and tools is crucial for maintaining resilience. CYJAX provides trusted, validated, and actionable threat intelligence to help organisations stay ahead of emerging risks.

➡️ Learn more about how CYJAX monitors, analyses, and mitigates threats: www.cyjax.com

Subscribe for weekly updates

Receive our latest cyber intelligence insights delivered directly to your inbox

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.
Share
Get started

Get Started with CYJAX CTI

Empower Your Team. Strengthen Your Defences.CYJAX gives you the intelligence advantage: clear, validated insights that let your team act fast without being buried in noise.

Link Copied