Identity Fraud Detection Playbook
This playbook is a collaboration between the Federal Chief Information Security Officer Council Identity, Credential, and Access Management Subcommittee and the Digital Identity Community of Practice, and the DOD DARPA SemaFor Program team. For additional context and to request a conversation or demo of the tools mentioned within this document please contact icam@gsa.gov.
Version Number | Author | Date | Change/Updates |
---|---|---|---|
1.0 | OGP Identity Assurance and Trust Access Division | February 2025 | Developed template and completed the document. |
Acknowledgments
This playbook was developed by GSA’s Office of Technology Policy’s Identity and Trusted Access Division. Technical contributions pertaining to media forensics were made by the members of the Defense Advanced Research Projects Agency Semantic Forensics (SemaFor) program team. OGP IATAD appreciates the collaboration and ongoing support from DARPA on this effort and looks forward to jointly collaborating on other high impact engagements as opportunities arise.
Executive Summary
In today’s increasingly digital world, federal agencies face a growing threat of identity fraud. Sophisticated actors continually develop new tactics to exploit vulnerabilities and compromise sensitive information. This playbook serves as a resource and provides a foundational understanding of identity fraud techniques, detection methods, and mitigation strategies. By establishing a common terminology and framework, this guide empowers agencies to:
- Recognize the diverse landscape of identity fraud techniques.
- Detect suspicious activities and potential threats.
- Implement effective mitigation strategies to protect sensitive data and systems.
- Foster cross-agency collaboration and information sharing to combat identity fraud collectively.
This playbook will delve into the intricacies of identity fraud and equip federal agencies with the knowledge and tools necessary to safeguard their operations.
Figure 1: Identity Fraud Detection Overview
Purpose
This playbook provides Federal Identity, Credential, and Access Management program managers with the knowledge and tools needed to combat identity fraud. It provides clear definitions of key fraud terminology to establish a common understanding to be used as a baseline for effective communication and cross-agency collaboration. The playbook offers insights by listing identity fraud techniques, use cases, and scenarios which explain how criminals exploit vulnerabilities and compromise information. It provides effective fraud detection and mitigation strategies to help identify and prevent fraudulent activities. Additionally, this playbook provides guidance on assessing an agency’s unique risks and tailoring its approach accordingly. Finally, it supports setting objectives and aligning fraud prevention programs with federal policy.
Scope
The scope of this playbook covers critical fraud detection techniques such as deepfake and forgery detection techniques for government agency use. While this playbook provides a comprehensive overview of identity fraud, it does not delve into specific technical implementations or vendor-specific solutions. The focus remains on equipping agencies with the foundational knowledge and strategic guidance necessary to develop effective fraud prevention programs.
Key Terms
These are key terms used throughout this document:
- Behavioral Biometrics - Behavioral biometrics establish identity by monitoring the distinctive characteristics of movements, gestures, and motor-skills of individuals as they perform a task or series of tasks.
- Biometrics - Unique biological markers (e.g., fingerprints, facial patterns) used for identity verification.
- Blockchain - A distributed ledger technology used for secure and immutable credential verification.
- Credential Stuffing - Credential stuffing is the automated injection of stolen username and password pairs (“credentials”) into website login forms, in order to fraudulently gain access to user accounts.
- Deepfake - A deepfake is a video, photo, or audio recording that seems real but has been manipulated with AI.
- Digital Content Forgery - Digital content forgery technologies enable adversaries to create or manipulate digital audio, visual, or textual content, to distort information, undermine security and authority, and ultimately erode trust in each other and in our government.
- Facial Recognition - Facial recognition technology is a contemporary security solution that automatically identifies and verifies the identity of an individual from a digital image or video frame. This technology can be compared to other biometric technologies, and used for a number of activities.
- FISMA - A U.S. law requiring federal agencies to protect information and information systems.
- Generative Adversarial Networks - A generative adversarial network is an artificial neural network with a distinctive training architecture, designed to create examples that faithfully reproduce a target distribution.
- Hardening - A process intended to eliminate a means of attack by patching vulnerabilities and turning off nonessential services.
- Spoofing - The deliberate inducement of a user or resource to take incorrect action. Note: Impersonating, masquerading, piggybacking, and mimicking are forms of spoofing
- Iris Recognition - Iris recognition is the process of recognizing a person by analyzing the random pattern of the iris.
- NIST Standards - Guidelines set by the National Institute of Standards and Technology for cybersecurity and identity verification.
- Phishing - The act of tricking individuals into disclosing sensitive personal information through deceptive computer-based means.
- Social Engineering - An attempt to trick someone into revealing information (e.g., a password) that can be used to attack systems or networks.
- Synthetic Media - Synthetic Media is an all encompassing term to describe any type of content whether video, image, text or voice that has been partially or fully generated using artificial intelligence or machine learning.
Audience
The primary audience for this playbook are agency digital identity and security program managers. Table 1 lists stakeholders and stakeholder types digital identity program managers engage with during identity fraud prevention, detection and mitigation. An internal stakeholder is within the agency and external stakeholders are outside the agency.
Stakeholder | Stakeholder Type | ||||||||||||||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Digital Identity and Security Program Managers</th> | Internal | </tr>||||||||||||||||||||||||||||||||
Security Operations</th> | Internal | </tr>||||||||||||||||||||||||||||||||
IT / System Administrators</th> | Internal | </tr>||||||||||||||||||||||||||||||||
Human Resources</th> | Internal | </tr>||||||||||||||||||||||||||||||||
Users</th> | Internal/External | </tr>||||||||||||||||||||||||||||||||
Third-Party Assessors</th> | External | </tr>||||||||||||||||||||||||||||||||
Law Enforcement Agencies</th> | External | </tr> </tbody> </table> ## Disclaimer This playbook was developed by the General Services Administration Office of Government-wide Policy with input from federal IT practitioners. This document shouldn’t be interpreted as official policy or mandated action, and doesn’t provide authoritative definitions for IT terms. Instead, this playbook supplements existing federal IT policies and builds upon the [Office of Management and Budget Memorandum 19-17 (OMB M-19-17), Enabling Mission Delivery through Improved Identity, Credential, and Access Management](https://www.whitehouse.gov/wp-content/uploads/2019/05/M-19-17.pdf){:rel="noopener noreferrer"}{:class="usa-link usa-link--external"}, as well as existing federal identity guidance and playbooks. Privileged user access (e.g., superusers, domain administrators) is out of scope for this playbook. ## Identity Fraud 101 ### What is it? *“The impact of identity theft on the government sector reaches far and wide, costing Americans tens of billions of dollars every year”* *__(LexisNexis, 2016, pg.9)__[^1]*. Identity fraud in the federal government happens when someone uses a stolen identity or a fake identity to gain access to a government system. This type of fraud has evolved with time and has grown increasingly sophisticated. Awareness of different identity fraud techniques will help agency fraud prevention, encourage compliance with federal policies, and promote unified interagency approach. ### Key Types of Identity Fraud Techniques This section of the playbook highlights a few key fraud techniques, methods, and their impact when exploited for perspective and does not intend to be all inclusive. **Identity Fraud** - **Synthetic Identity Fraud** - The creation of false identities using a mix of real and fabricated information. It often involves combining real data, such as a real social security number, with fake information, like a fictitious name or date of birth. - **Identity Theft** - The use of stolen personal information to impersonate another individual. This can lead to unauthorized transactions or activities being conducted in the victim's name. **Types of Identity Fraud Techniques** - **Social Engineering** - Psychological manipulation of people into performing actions or divulging confidential information. Often involves tricking people into breaking normal security procedures. - **Phishing** - Fraudulent attempts to obtain sensitive information such as usernames, passwords, and credit card details by disguising oneself as a trustworthy entity in electronic communications. - **Email Phishing** - The most common form of phishing, where attackers send emails that appear to be from reputable sources. These emails often contain malicious links or attachments designed to steal personal information or install malware. - **Spear Phishing** - A more targeted form of phishing where attackers customize their attack to a specific individual or organization. The emails used in spear phishing attacks are highly personalized and often contain information relevant to the target to make the deceit more convincing. - **Whaling** - A form of spear phishing aimed at high-profile targets within a company, such as executives or decision-makers. Whaling attacks often involve attempts to steal sensitive data or execute fraudulent transactions. The emails are usually crafted to look like important or official communications. - **SMS Smishing** - Phishing conducted through SMS messages. Victims receive text messages that appear to be from reputable sources, often asking them to follow a link or call a phone number to resolve a fake issue like a compromised bank account or unpaid bill. - **Vishing (Voice Phishing)** - Involves phone calls instead of electronic communication. Attackers use spoofed caller IDs and pretend to be from trusted organizations like banks or tech support to trick victims into providing personal or financial information. - **Deepfake Technology** - AI-based manipulation of audio, video, and images to create realistic but false representations of reality. Contemporary technologies enable realistic fabrications that can be used for various fraudulent activities. - **Audio Deepfake** - Audio deepfakes use advanced AI algorithms to create a synthetic replica of a person's voice. This technology can convincingly mimic a person's speech patterns, intonation, and cadence after being trained on recordings of that person's voice. - **Video Deepfake** - Video deepfakes are alterations of a person’s appearance in a video to make it seem as though they are saying or doing something they never did. This process often uses sophisticated AI models, such as GANs, to achieve realistic results. - **Visual Deepfake** - Visual deepfakes use AI to create or alter images, often by swapping faces, morphing features, or generating entirely new images of non-existent people. GANs are a common method for generating and manipulating images with a high level of realism. - **Enabling Tech** - AI and machine learning models for generating deepfakes. - **Credential Stuffing** - Using stolen or leaked credentials from one system (typically usernames and passwords) to access others. This threat is based on the common reuse of passwords across multiple accounts. - **Account Takeover** - Illicit access and control of another person's account, which can result in unauthorized transactions or activities. - **Forgery Techniques** - Advanced document and credential falsification, including altering or counterfeiting government documents such as passports, driver's licenses, or diplomas. - **Data Aggregation and Data Breaches** - Unlawful access to sensitive data through breaches which can be used for various fraudulent activities including identity theft, synthetic identity fraud, and credential stuffing. **Combinations of Identity Fraud Techniques** In today's interconnected digital world, identity fraud has become increasingly intricate, exploiting advanced technologies and social engineering tactics. Fraudsters often blend multiple techniques to create sophisticated and highly effective fraud schemes, significantly enhancing their ability to deceive victims and evade detection. By understanding the synergy between these various methods, we can better equip ourselves against such multifaceted identity fraud attacks. The following are some examples of damaging combinations of identity fraud techniques. - **Email Phishing with Credential Stuffing** - **Mechanism** - Attackers initiate a phishing campaign by sending deceptive emails that prompt victims to reveal their login credentials. The obtained credentials are then used in credential stuffing attacks to access additional systems where the same usernames and passwords have been reused. - **Impact** - This combination can compromise multiple accounts, leading to unauthorized transactions, data breaches, and further exploitation of the user's digital footprint across various platforms. - **Spear Phishing with Forgery Techniques** - **Mechanism** - Fraudsters conduct in-depth research on a specific individual or organization to send highly personalized phishing emails. They bolster their credibility by using forged documents, such as fake invoices or official-looking correspondence, to convince the target to provide sensitive information or authorize fraudulent activities. - **Impact** - The success rate of spear phishing increases significantly when paired with forgery techniques, resulting in considerable financial and reputational damage to the targeted entities. - **Whaling with Deepfake Technology** - **Mechanism** - High-profile individuals such as executives are targeted using whaling techniques combined with deepfake technology. Fraudsters create realistic deepfake videos or audio messages impersonating trusted colleagues or partners, persuading these high-level targets to disclose sensitive information or approve significant financial transactions. - **Impact** - The convincing nature of the deepfakes, coupled with the high stakes involved in whaling attacks, can lead to substantial financial losses and breaches of highly sensitive corporate data. - **SMS Phishing with Social Engineering** - **Mechanism** - Attackers use smishing to send deceptive SMS messages that appear to be from reputable sources, often claiming an urgent issue like a compromised bank account. They then employ social engineering tactics, such as creating a sense of urgency or fear, to manipulate victims into clicking on malicious links or calling phone numbers to provide personal information. - **Impact** - The combination of smishing and social engineering can effectively pressure victims into acting quickly and irrationally, leading to the rapid disclosure of sensitive information and unauthorized financial transactions. - **Vishing (Voice Phishing) with Audio Deepfake Technology** - **Mechanism** - Fraudsters use audio deepfake technology to replicate the voice of a trusted individual, such as a company executive or a bank representative. They then make vishing calls using spoofed caller IDs, convincing the target to share confidential information or authorize fraudulent activities. - **Impact** - The use of deepfake audio adds realism to vishing attacks and increases their success rate dramatically, making it more likely for victims to fall for the scam and divulge critical information. ## Effects of Identity Fraud - **Fraud Technique** - Forgery - Document and Credential Falsification - **Scalability** - Digital Tools - The use of sophisticated software tools enables forgers to create highly accurate counterfeit documents. This includes advanced image editing software for precision alteration of document features like holograms, watermarks, and microprints. - **Printing Technology** - High-resolution printers and specialized printing techniques allow the production of near-identical replicas of official documents. - **Distributed Networks** - The proliferation of underground online marketplaces and distribution networks facilitates the wide dissemination and sale of forged documents globally. - **Machine Learning Models** - Emerging utilization of machine learning allows algorithms to replicate security features and anti-counterfeiting measures more effectively. - **Easier Access and Collaboration** - Increased online tutorials, forums, and dark web collaborations provide forgers with the knowledge and resources to enhance their techniques quickly. - **Potential Impact** - Unauthorized access to secured locations and systems by exploiting counterfeit credentials, posing significant security risks to institutions and critical infrastructure. - Creation of false identities allowing criminals to engage in illegal activities, such as financial fraud, without being easily traced. - Erosion of trust in official documents and institutions, as widespread falsification questions the authenticity of legitimate credentials. - Illegal border crossings and bypassing of immigration controls using forged passports and visas, undermining national security. - Individuals obtaining employment, educational opportunities, or professional licenses using fake diplomas, certifications, or professional credentials, potentially leading to unqualified personnel in crucial roles. - Opening bank accounts, taking loans, or conducting financial transactions under false pretenses, resulting in substantial monetary losses for institutions and individuals. - Misuse of falsified documents to obtain sensitive or dangerous materials, such as pharmaceuticals, firearms, or hazardous chemicals. - **Fraud Technique: Deepfake (synthetic media)** - **Scalability** - Image: Image generators are improving, becoming more customizable with better control and are easier to use. - Audio: Improved fidelity of voice generation with less source material required; increasingly refined controls of auditory/vocal qualities, such as environment, accent, gender, emotion, etc. - Video: Longer-lasting and more photorealistic text-guided video generation & editing; more natural audio-driven deep fake video; full body video generation & reenactment - **Potential Impact** - Gaining access to critical resources by cloning senior executives or decision makers’ voices or video. - Information manipulation at scale, exploiting a public figure’s reputation for mass effect. - Election Interference. - Individual targeting by an online account appearing to be that of a popular or public figure. - Potentially irreversible reputational and financial damage. - Generation of non-consensual explicit content. ## Fraud Detection Capabilities ### Introduction The use cases below exhibit sample media forensic workflows drawn from DARPA’s Semantic Forensics (SemaFor) program capabilities, including analytics for image, voice and video identity fraud detection. SemaFor technologies are utilized here to illustrate elements of an exemplar workflow, for which an increasing marketplace of academic and industry products exist. This section provides a discussion of potential requirements of such a technical workflow in government fraud detection goals. ### The Technology Platform Recent industry and academic efforts have created multiple families of technologies designed to help mitigate online threats perpetuated via synthetic and manipulated media. These technologies include algorithms that can detect, attribute, and characterize manipulated or synthesized media multi-modal media assets (e.g., text, audio, image, video). Models like ChatGPT, DALL-E, and Midjourney are all recent tools that enable the automated generation of content, and the widespread adoption of such technologies has resulted in an explosion of AI-generated multimedia that makes it challenging to understand the provenance and authenticity of online material. Additionally, media generation and manipulation technologies are advancing rapidly, and traditional detection techniques that rely on a file’s statistical fingerprints are insufficient for detecting modern manipulated media. The workflow below illustrates a suite of analytics leveraged to detect, attribute, and localize manipulations in media through the identification of artifacts and inconsistencies, with the goal of providing an informative, probabilistic assessment of file manipulation. **Scenario 1 - Image** A digital photograph file, of individual, individual’s location, or documents such as a passport, are submitted electronically for purposes of identification or verification of identity or location or association with another individual (this image may be potentially morphed, generated or modified). **Analyst Task - Determine if presented image is generated or manipulated; localize the manipulation if present** Algorithms, such as those displayed in the DARPA SemaFor program prototype user interface depicted below, analyze the image and provide a finding of “Likely Manipulated,” along with detailed information to assist the analyst in making a judgment on a particular piece of media, in this case a generated image. **Figure 2 - Image**
Acronym | Definition |
---|---|
CISA</th> | Cybersecurity and Infrastructure Security Agency | </tr>
DARPA</th> | Defense Advanced Research Projects Agency | </tr>
FBI</th> | Federal Bureau of Investigation | </tr>
FISMA</th> | Federal Information Security Management Act | </tr>
GAN</th> | Generative Adversarial Network | </tr>
GAO</th> | Government Accountability Office | </tr>
GSA</th> | General Services Agency | </tr>
ICAM</th> | Identity, Credential, and Access Management | </tr>
IRP</th> | Incident Response Planning | </tr>
NIST</th> | National Institute of Standards and Technology | </tr>
OGP</th> | Office of Government-wide Policy | </tr>
OIG</th> | Office of Inspector General | </tr>
OMB</th> | Office of Management and Budget | </tr>
SemaFor</th> | Semantic Forensics | </tr>
SMS</th> | Short Message Service | </tr> </tbody> </table> [^1]: LexisNexis, 2016 The Identity Fraud Prevention Playbook, [https://risk.lexisnexis.com/cross-industry-fraud-files/risk/downloads/assets/id-fraud-prevention-playbook.pdf](https://risk.lexisnexis.com/cross-industry-fraud-files/risk/downloads/assets/id-fraud-prevention-playbook.pdf){:rel="noopener noreferrer"}{:class="usa-link usa-link--external"}