AI Fakers, Deepfake Candidates, and North Korean IT Workers: The Hiring Crisis No One Saw Coming

Fake candidates aren’t new, employers just haven’t been great at spotting them. From AI-generated resumes to North Korean IT workers infiltrating tech firms, the hiring crisis is here. The question is: Can your hiring process handle it?

AI Fakers, Deepfake Candidates, and North Korean IT Workers: The Hiring Crisis No One Saw Coming
Photo by Aideal Hwa / Unsplash

The Job Applicant That Didn’t Exist

Imagine this: A job candidate aces your coding test, completing it faster than any other applicant. Their resume is flawless, their credentials are top-tier, and their interview answers are polished.

But something feels… off.

That’s exactly what happened to Vidoc Security, a cybersecurity startup, when they nearly hired a completely fake AI-generated applicant.

During the final video interview, the candidate refused to place their hand in front of their face, a simple request that would have broken their deepfake disguise.

Vidoc caught the fraud just in time. Two months later, it happened again.

They exposed another deepfake candidate who had manipulated their video feed in real-time, fooling multiple rounds of screening.

Take a read of the full postmortem by Gergely Orosz...

AI fakers exposed in tech dev recruitment: postmortem
A full-remote security startup nearly hired a backend engineer who doesn’t exist, after a candidate used an AI filter as an on-screen disguise in video interviews. Learnings for tech companies

And Vidoc isn’t alone.

Hiring fraud is exploding. AI tools allow anyone to fake expertise, while state-sponsored operations like North Korean IT worker scams infiltrate global companies under false identities.

This isn’t just a hiring scam, it’s an existential threat to recruitment.


AI Enters the Hiring Chat (Literally)

Hiring managers have long relied on AI-driven tools to streamline applications, automate screenings, and even analyze candidate speech patterns.

But job seekers have flipped the script.

🔹 Resumes crafted by ChatGPT
🔹 Portfolios designed by Midjourney
🔹 Real-time AI whispering interview answers through earpieces

The result? A credibility crisis.

Who’s truly qualified, and who’s just exceptionally good at using AI?

Even worse, some “candidates” aren’t candidates at all.

Hiring fraud is no longer just about bad actors gaming the system—it’s an international security threat.

Blame the Game, Not the Players

AI-generated candidates may be new, but gaming the hiring system is not.

People have been keyword-stuffing resumes, hiring ghostwriters for cover letters, and inflating job titles for years.

AI just makes cheating faster, easier, and nearly undetectable. So instead of panicking about AI fakers, employers should be asking:

Why does our hiring process reward fakery in the first place?

A buzzword-heavy resume passes through ATS filters.
A candidate who memorized “Top 10 Interview Answers” breezes through an interview.
A technically weak but well-rehearsed applicant outshines a skilled but introverted one.

And now, AI is amplifying these hiring loopholes at scale.

If it’s easy for an AI-assisted job seeker to look like a perfect hire, what’s stopping cybercriminals from doing the same?

North Korean IT Workers: The Global Hiring Fraud You Should Be Worried About

For most companies, hiring fraud is an inconvenience. For some, it’s a national security risk.

While AI-generated resumes and deepfake candidates are concerning, there’s an even more sophisticated operation at play, one that goes far beyond individual job seekers looking to game the system. State-sponsored cybercriminals are using hiring fraud as a tool for financial and intelligence warfare.

According to Google Cloud’s Threat Intelligence report, North Korean IT workers have infiltrated companies worldwide, primarily in tech, finance, and cybersecurity roles. The goal? To funnel millions of dollars back to the DPRK regime while embedding themselves inside Western companies.

Staying a Step Ahead: Mitigating the DPRK IT Worker Threat | Google Cloud Blog
North Korea’s IT workforce presents a persistent and escalating cyber threat.

Why Is North Korea Targeting Remote Tech Jobs?

🔹 Generating Revenue for the DPRK Regime

  • Fraudulent salaries fund weapons programs, cyberwarfare, and sanctions evasion.

🔹 Access to Critical Infrastructure

  • DPRK operatives have used stolen credentials to introduce malware and exfiltrate sensitive data.

🔹 Exploiting the Remote Work Boom

  • Remote hiring makes it easier to fake identities, manipulate resumes, and evade background checks.

The $6.8M Hiring Scam

A North Korean IT worker facilitator helped fake candidates secure over 300 jobs, using:

Stolen Identities – Real names, fake workers
Laptop Farms – U.S. company laptops rerouted to offshore locations
Multiple Simultaneous Jobs – One person, 5-10 salaries

The total damage? $6.8 million in fraudulent salaries.

How to Spot AI Fakers & Hiring Fraud: Key Red Flags

🚩 1. AI-Optimized Resumes That Cheat the System

  • AI-generated resumes mirror job descriptions perfectly.
  • Solution: Use AI-detection tools like GPTZero or Originality.ai to flag suspicious applications.

🚩 2. Inconsistent Digital Footprints

  • Fake candidates have limited online presence or questionable LinkedIn connections.
  • Solution: Cross-check GitHub, LinkedIn, and industry contributions for real engagement.

🚩 3. Suspicious Video Interview Behavior

  • AI fakers avoid direct eye contact, use filters, or refuse to perform basic actions (like turning their head).
  • Solution: Record video interviews and require candidates to turn off all filters.

🚩 4. Work Experience That Doesn't Add Up

  • Fraudulent candidates list non-existent companies or inconsistent job titles.
  • Solution: Contact past employers directly instead of relying on LinkedIn.

How Employers Can Defend Against AI Hiring Fraud

Vidoc Security and Google’s cybersecurity team recommend the following strategies to detect and prevent AI-generated candidates:

Implement AI Detection for Resumes

  • Use GPTZero, Copyleaks, or Originality.ai to scan for AI-written cover letters and CVs.

Require Live Video Verification

  • Conduct real-time video calls, and ask candidates to perform physical actions (like moving their head or touching their face).

Cross-Check Digital Footprints

  • Verify social media engagement, GitHub repositories, and past employer references.

Use Identity Verification Tools

  • Require notarized identity proof and multi-step background checks.

Monitor Remote Work Access

  • Block suspicious VPNs, remote desktop tools, and unusual login patterns.

Educate HR and Hiring Teams

  • Train recruiters to spot AI fakers, deepfake interviews, and stolen identities.

📄 Download Vidoc’s Guide on Detecting Fake Candidates

If AI Can Fake It, Your Hiring Process Needs to Change

If Vidoc Security, a company specializing in fraud detection, nearly hired two deepfake candidates, what are the chances your hiring team has already been fooled?

The future of hiring isn’t about eliminating AI, it’s about building a hiring process that AI (and fraudsters) can’t game.

🚨 The next hiring scandal is just a bad hire away. Are you ready to adapt, or will your company be the next headline? 🚀

Takeaways

How are AI fakers bypassing hiring systems?

AI-generated candidates exploit keyword-heavy ATS filters, use AI whisper tools to ace interviews, and deploy deepfake technology to manipulate video calls. Traditional hiring processes are struggling to keep up with AI-driven deception.

Why are North Korean IT workers targeting Western companies?

These state-sponsored operatives secure remote jobs under fake identities, using their salaries to fund North Korea’s weapons program. Some even gain privileged access to sensitive corporate networks, making this not just a hiring issue, but a national security threat.

How can recruiters spot AI-generated candidates?

🔹 Look for “too perfect” resumes AI-written CVs often mimic job descriptions exactly.
🔹 Analyze digital footprints legitimate professionals usually have consistent online activity (LinkedIn, GitHub, industry forums).
🔹 Require live video verification ask candidates to turn off all filters, move their head, and perform basic actions during interviews.

Should companies ban AI-assisted applications?

Not necessarily. AI can enhance productivity and creativity, but it should be used transparently. Instead of banning AI outright, assess how candidates use it and whether it improves their actual job performance.

What changes should hiring managers make today?

AI-detection tools for resumes use GPTZero, Originality.ai, or Copyleaks to scan applications.
Mandatory live video interviews deepfake fraud collapses under real-time scrutiny.
Stronger identity verification require notarized ID proof and multi-step background checks.
Monitoring of remote work tools flag suspicious VPN usage and remote desktop access patterns.
HR training in AI fraud detection recruiters should know how to spot deepfake behaviors and AI-generated deception.

What’s the employer branding risk of hiring fraud?

A single hiring scandal can destroy trust, attract negative press, and expose companies to lawsuits. Organizations known for lax hiring security will struggle to attract top talent especially in cybersecurity, tech, and finance.


Did you enjoy reading this? Elevate your employer branding game. Subscribe for more fresh content and expert insights - it's free!


EBN Jobs - Global Employer branding Roles

Let’s be real: a great employer brand is about more than quirky seating and trendy perks. Your next career move should be built on substance—where your strategy defines the culture, not the furniture.

Ready to make a real impact? Join Employer Branding Jobs and discover roles where strategy matters more than style.

EBN Jobs