TheAlgoBrief

IAB AI Transparency & Disclosure Framework: Complete Guide 2026

22 min readTheAlgoBrief Editorial Team
IAB FrameworkAI AdvertisingAI DisclosureAd Tech ComplianceC2PASynthetic InfluencersNY Synthetic Performer LawEU AI Act
What Is the IAB AI Disclosure Framework?

The IAB AI Transparency & Disclosure Framework (January 2026) is a voluntary industry standard that tells advertisers, agencies, and ad tech platforms when AI-generated or AI-modified ad content requires a consumer-facing disclosure label — and exactly what that label must say. It covers five content types (images, video, audio, synthetic influencers, and text) and uses a three-part materiality test to determine if disclosure is required: Deception Potential, Material Impact, and Expectation Alignment. It also defines custom C2PA metadata assertions (com.iab.threshold, com.iab.disclosure) for machine-readable compliance signalling across DSPs, SSPs, and publishers.


Table of Contents


1. What Is the IAB AI Transparency & Disclosure Framework?

The IAB AI Transparency and Disclosure Framework is an industry standard published by the Interactive Advertising Bureau (IAB) in January 2026. It is the advertising industry's most comprehensive attempt to answer one question that every brand, agency, and creative team is now asking: When AI creates or modifies ad content, does the consumer need to know?

The framework does not say "always disclose everything AI." Nor does it say "AI is just another production tool, no disclosure needed." It does something more nuanced — and more useful. It defines a materiality threshold: a test that determines whether AI involvement in content creation is significant enough that a reasonable consumer would care if they knew about it. If the answer is yes, disclosure is required and the framework tells you exactly what to say and where to put it.

The IAB is not a regulator. It cannot fine you. But the framework matters enormously for three reasons:

First, it is rapidly becoming the de facto compliance standard that major platforms, DSPs, and publishers are adopting as their enforcement baseline. When Google, Meta, or The Trade Desk decides how to handle AI disclosure requirements at the platform level, they will build toward this framework.

Second, it maps directly onto the disclosure requirements embedded in binding law — particularly the New York Synthetic Performer Law (effective June 2026), California SB 942 (effective August 2026), and the EU AI Act — giving compliance teams a single framework they can use across jurisdictions.

Third, it defines the C2PA custom assertions that ad tech infrastructure uses to signal AI disclosure compliance in machine-readable metadata. If you are building or buying ad tech in 2026, the IAB framework is the specification you are building to.

The IAB AI Transparency Framework at a Glance — timeline showing Jan 2026 framework release, key regulatory deadlines (NY June 2026, CA Aug 2026, EU 2027), and the five content types covered
The IAB AI Transparency & Disclosure Framework: key dates, content types, and regulatory intersections. January 2026.

2. Why the IAB Created the AI Transparency Framework (January 2026)

At a Glance
  • The Problem: Fragmented regulations (NY, CA, EU, S. Korea) and eroding consumer trust.
  • The Solution: A unified industry standard that harmonizes legal compliance with technical ad-serving infrastructure.
  • Key Impact: Moves AI disclosure from a manual "policy check" to a machine-readable technical requirement (C2PA).
Market Context

AI-generated advertising content went from novelty to mainstream in roughly 18 months. By late 2024, a significant share of digital display creative was being generated or substantially modified by AI tools. By 2025, AI-synthesized voices were appearing in radio and podcast ads at scale.

The advertising industry's existing disclosure frameworks were not built for this. Prior standards assumed that the humans creating advertising content were humans, using human labor. The question of "did AI make this?" simply was not on the regulatory or ethical checklist.

Three things converged to make the IAB move:

Regulatory Pressure

Legislators in NY and CA passed laws requiring AI disclosure, with fines up to $5,000 per violation. EU AI Act and S. Korea laws are already in force.

Consumer Trust

Eroding trust from undisclosed AI imagery and voices directly impacts campaign performance and brand authenticity in the long term.

Platform Enforcement

Major DSPs and exchanges now require technical signals (C2PA) to flag AI creative, forcing a move from simple policy to technical standard.

Regulatory Pressure & Real Financial Strings

The legal landscape is no longer hypothetical. Legislators in New York and California have passed laws with sharp teeth and fines upto $5,000.

The New York Synthetic Performer Law and California SB 942 require clear disclosure of AI-generated performers and content. Suddenly, the legal landscape was fragmenting, and the industry needed a unified standard to navigate it without managing 50 different sets of rules.

The IAB's response was the January 2026 framework: the first industry standard that is simultaneously a decision-making guide, a vocabulary of specific label texts, and a technical specification for C2PA metadata.


3. The Three Disclosure Checks

Before requiring a disclosure label, the IAB framework asks three simple yes/no questions. These determine if the AI involvement is significant enough that a consumer needs to know about it. If all three answers are "Yes", a label is required.

The materiality test is intentionally practical — it is designed to let routine AI-assisted production continue without disclosure burdens, while drawing a clear line at the point where AI involvement could actually mislead a consumer about what they are buying or who is endorsing it. Think of the three checks as a filter: most AI post-production work fails Check 1 and never needs a label. Only the cases that clear all three reach the point of required disclosure.

The 3 IAB AI Disclosure Checks Cheat Sheet
The 3 IAB AI Disclosure Checks: A definitive logic map for determining if a disclosure label is required.
01

Is it realistic enough to deceive a consumer?

Could a reasonable person be misled about the nature, origin, or authenticity of this content? If it's a clearly fantastical rocket ship or a cartoon, people know it's not real photography.

IF YES

Continue to Check 2

IF NO

No disclosure required

02

Does it materially change what the consumer thinks of the product?

Does the AI change product features, performance claims, or social proof? Retouching a background is fine; making a product look slimmer or using a fake person to endorse it is a material change.

IF YES

Continue to Check 3

IF NO

No disclosure required

03

Is it beyond what people expect from standard production?

Retouching and color grading are routine. AI-generated people who don't exist or deepfakes of real individuals fall outside standard expectations and require disclosure if they appear realistic.

IF YES

ALL CHECKS PASSED. Disclosure Required.

IF NO

No disclosure required

It is worth noting what the test is not asking. It does not ask whether AI was involved at all — the framework explicitly accepts that AI tools are now part of the standard creative production stack. It does not ask whether the brand intended to deceive. It asks only whether the output, as a consumer would perceive it, clears these three thresholds. An AI tool used in good faith with no deceptive intent can still produce content that requires a label if the output meets the materiality criteria.


4. Content-Type Rules: Images, Video, Audio, Synthetic Influencers, Text

The IAB framework applies the materiality test differently across five content types. Each has specific decision branches, specific exemptions, and specific label texts. Understanding which category your creative falls into is the most important step — the downstream disclosure requirements and legal exposure vary significantly across them.

At a Glance: Does Your AI Creative Need a Disclosure? — visual reference card showing all five content types with key exemptions and requirements
IAB AI disclosure requirements by content type. Download the reference card for your compliance review workflow.

4.1 Images

The most common application of the framework, because AI image generation is now the default production tool for many creative teams. The key variable is photorealism — not the tool used, not whether the image started from a real photograph.

ScenarioStatusDisclosure Rule
Fully Generated🔴 RequiredIf the image looks like a real photograph.
Substantive Edit🔴 RequiredIf the product itself or a real person is modified.
Stylized / Abstract🟢 ExemptIf the artificial nature is obvious (cartoons, 3D art).
Routine Edit🟢 ExemptBackground cleanup, color correction, blemish removal.

A photorealistic AI render of a coffee cup on a kitchen counter requires disclosure. A clearly illustrated or 3D-animated version of the same coffee cup does not. The distinction is not aesthetic — it is whether the image could pass for a real photograph in a casual viewing context.

The trickiest edge case is substantive editing of real photography. The framework draws a precise line here: AI changes that alter the appearance of the product itself or of a real person shown in the ad require disclosure. Background-only replacements are generally exempt, provided the new background does not misrepresent where or how the product performs. Removing a cluttered shelf and replacing it with a clean studio backdrop is fine. Adding a beach backdrop to a moisturiser to imply it's water-resistant when it isn't is not.

4.2 Video

Video rules follow image logic but add a timing requirement: when a label is required, it must remain visible for a minimum duration so that consumers actually register it.

ScenarioStatusDisclosure Rule
Fully Generated🔴 RequiredRealistic video with no clear artificial markers.
Altered Subject🔴 RequiredIf a real person or product is meaningfully changed.
Animated / Stylized🟢 ExemptClearly 2D or 3D animation.
Standard Post-Prod🟢 ExemptNoise reduction, color grading, AI upscaling.

If disclosure is required, the label must appear for at least 2 seconds at the first appearance of the AI-generated or AI-altered elements — not buried at the end of the spot. A label that flashes for a half-second in a corner position does not satisfy this requirement.

AI upscaling, frame interpolation, and noise reduction are all exempt. These are established post-production techniques that consumers neither expect to see disclosed nor would find material to their understanding of the product. The disclosure obligation activates when AI is generating or altering something a consumer would consider real.

4.3 Audio

Audio rules are built around identity and authenticity, and they carry the sharpest legal exposure of any content type for U.S. advertisers. The framework's audio provisions map directly onto the New York Synthetic Performer Law — the binding legislation with the hardest deadline and the clearest penalty structure.

ScenarioStatusDisclosure Rule
Deceased Person🔴 RequiredALWAYS required, even with estate authorization.
Fabricated Claim🔴 RequiredUsing a real person's voice clone to say unverified things.
Generic AI Voice🟢 ExemptText-to-speech with no intended human likeness.
Contracted Clone🟢 ExemptStandard authorized use for scripted commercial copy.

The deceased performer rule is a hard line with no exceptions. It does not matter whether the estate has consented, whether the estate was paid, or whether the use is respectful of the performer's legacy. Any AI-synthesized voice of a deceased person in a commercial context requires disclosure. This is one of the clearest and most consistently enforced provisions across all jurisdictions — IAB, New York, California, and the EU are aligned on this.

For living persons, the framework introduces a context dependency: authorized voice clones used for scripted commercial copy are generally exempt. The line is crossed when the voice clone is used to generate statements about real-world facts, fabricate endorsements, or create impressions that the person holds views they have not expressed.

Since audio ads often have no visual component, the framework specifies that disclosure must appear in the ad metadata and in any visible companion banner or player interface associated with the ad. For purely audio-only contexts (smart speakers, radio), disclosure in the spoken content or a clear spoken cue is required.

Audio: June 2026 deadline for U.S. agencies

The New York Synthetic Performer Law takes effect in June 2026 and carries civil penalties of $1,000–$5,000 per occurrence for undisclosed use of AI-synthesized voices of real individuals. This applies to both living and deceased performers. Any agency running audio campaigns in New York markets Needs a disclosure review in their production workflow before that date.

Urgent: June 2026 deadlineNY Synthetic Performer Law: What Ad Agencies Need to Know Before June 2026If the penalty structure above concerns you, this is the next read. Full breakdown of what the NY law requires specifically, how fines are calculated per occurrence, and what a compliant audio workflow looks like before June.

4.4 Synthetic Influencers

This is the highest-risk category for regulatory fines across both NY and California, and the one where brands and agencies are most likely to have compliance gaps they are not aware of. The category is broad — it covers digital twins of real people, fully synthetic AI personas, and interactive AI chatbots appearing in ads.

ScenarioStatusDisclosure Rule
Deceased Twin🔴 RequiredALWAYS required.
Realistic Persona🔴 RequiredIf they appear as a primary brand endorser.
Interactive Bot🔴 RequiredMust state: "AI-powered assistant."
Stylized Avatar🟢 ExemptClearly non-human avatars or low-fidelity characters.

The digital twin rules for living persons introduce an authorization question that the other content types do not. Using an AI-generated likeness of a real, living person without their explicit authorization is not just a disclosure failure — it is likely a violation of that person's personality rights and potentially a direct breach of the New York Synthetic Performer Law. Authorization is necessary but not sufficient: even with authorization, placing a digital twin of a living person in fabricated events or locations they never visited requires a disclosure label.

Fully synthetic personas — AI-generated people who have never existed — require disclosure if they are realistic in appearance and appear in a primary endorsement role. An AI-generated person in the background of a lifestyle shot is incidental and does not require disclosure. That same AI-generated person speaking directly to camera about a product does. The framework's "primary vs. incidental" distinction is the key variable here.

Not sure if yours qualifies?Does My AI Ad Need a Disclosure Label?If you're now uncertain whether a synthetic persona in your current campaigns needs a label, the Materiality Checker gives you a verdict on your specific creative in under 90 seconds.

4.5 Text and Copy

The most lenient category in the current framework, and the one most likely to evolve as AI-generated content becomes more sophisticated.

ScenarioStatusDisclosure Rule
Standard Copy🟢 ExemptAI-written headlines, emails, or body text.
Testimonials🔴 RequiredNote: Fabricated AI reviews are deceptive per the FTC.

The framework's reasoning for exempting AI-written copy is grounded in consumer expectations: advertising copy has always been written by professionals using whatever tools were available, and the use of AI in that process does not change what the consumer learns about the product. A headline is a headline — the production method does not affect the consumer's understanding of the claim.

The testimonial exception is important and sits slightly outside the IAB framework proper. Fabricated AI-generated testimonials — fake reviews, manufactured customer quotes — fall under the FTC's general deception standards regardless of whether the IAB framework would otherwise require a label. The IAB framework does not need to address this directly because broader consumer protection law already does.

The framework notes that text disclosure requirements may be revisited as AI-generated testimonials and synthetic social proof become more prevalent. Treat the current text exemption as stable for 2026 but watch for updates.


5. What Are the Official IAB AI Disclosure Labels for 2026?

The framework is prescriptive about both what the label must say and where and for how long it must appear. Standardised label text is not stylistic preference — it is what allows consumers to recognise the labels across different ad contexts, and what allows platforms to detect and enforce compliance programmatically.

The Complete IAB Label Reference Table

The table below maps every content type that requires disclosure to its IAB-specified label text, required placement method, and duration. Click any label text to copy it exactly.

Source: IAB AI Transparency and Disclosure Framework, January 2026, pp. 18–19. Table recreated with additional practical notes.

Content TypeRequired LabelPlacement & DurationPractical Note
Synthetic ImageNear the image, clearly visibleNo minimum duration — static placement. Sparkle icon (✨) or C2PA CR icon are accepted alternatives.
AI-Altered ImageNear the image, clearly visibleApplies when the product itself or a real person in the photo has been substantively modified. Background-only changes are exempt.
Synthetic VideoFirst frame → persists for the entire videoThe label cannot be shown briefly then removed. It must remain on screen for the full duration of consumer exposure to AI-generated content.
Synthetic Voice — Deceased PersonSpoken verbal disclosure, before or after AI segmentRequired even with estate authorisation. In audio-only contexts: spoken disclosure must be repeated at least once if the ad exceeds 60 seconds.
Synthetic Voice — Living Person, Fabricated StatementsAudio: verbal pre- or post-content. Video: text overlayApplies when a voice clone makes unverified statements about real-world facts. Standard scripted commercial use with authorisation is exempt. In audio-only contexts: same spoken verbal disclosure rules apply — pre/post placement, and repeated at least once if the ad exceeds 60 seconds.
Synthetic AvatarFirst frame → persists throughout synthetic avatar useA fully AI-generated person who could be mistaken for a real human. If in the ad for its full duration, the label is present for its full duration.
Digital Twin — Deceased PersonFirst frame → persists throughout useAlways required regardless of estate authorisation or nature of content depicted.
Digital Twin — Living Person, Fabricated EventsFirst frame → persists throughout useApplies to authorised AI replicas shown in scenarios that never occurred. Standard authorised endorsement is exempt.
AI Chatbot / Conversational AgentAt initiation → accessible throughout entire ad experienceMust be disclosed before the consumer begins the interaction. Cannot be buried in small print or a terms link.

Three Placement Rules That Most Teams Get Wrong

Critical: Video Label Persistence

The framework does not specify a minimum number of seconds. It specifies that the label must remain visible for the entire duration of the video. Many creative teams assume a brief opening disclosure satisfies the requirement — it does not. If the AI-generated content is on screen, the label must be on screen.

Critical: Audio Verbal Disclosure & the 60-Second Rule

For audio-only contexts where no visual interface exists — radio, podcast feeds, smart speaker ads — a visible label in metadata is not sufficient. The framework requires a spoken verbal disclosure in the same language as the advertisement, delivered at a normal conversational pace. If the ad runs longer than 60 seconds, that verbal disclosure must be made at least twice. Position it before or immediately after the AI-generated segment — not as an end-of-spot afterthought.

Critical: Chatbot Initiation Timing

The disclosure must appear at the moment the consumer begins interacting with the AI — not after they have already engaged. An AI-powered interactive ad unit that reveals its AI nature only in a post-interaction screen does not satisfy this requirement.

Label Format Alternatives

The framework accepts three disclosure formats for visual content types. Text labels are the primary method and the recommended default. Two alternatives are permitted when text labels would materially compromise advertising effectiveness.

Label Format Alternatives — Visual Preview

Discover Our Product
Sparkle / Star Icon

Emerging cross-platform convention. Must include accessible alt text for screen readers (e.g., alt="AI-generated content"). Accepted by several major platforms in lieu of text.

Discover Our ProductCR
CRC2PA Content Credentials Icon

The most future-proof option. Simultaneously provides consumer-facing disclosure and machine-readable compliance via embedded C2PA assertions (com.iab.threshold, com.iab.disclosure).

Discover Our ProductAI-generated image
📝Text Label (Primary / Default)

The recommended default for all content types. Use the exact IAB-specified wording — do not paraphrase. The label must be in the same language as the advertisement.

Platform-applied indicators — such as Meta's "AI info" label — satisfy the IAB requirement when applied automatically based on C2PA metadata, provided they meet IAB conspicuousness standards. The advertiser remains accountable for verifying the platform has actually applied the label. Relying on platform enforcement without that verification is not a compliance strategy.

Key Takeaways for Section 5
  • Synthetic video requires a label that persists for the entire video — not just the first two seconds.
  • AI audio of any real person (living or deceased) requires a spoken verbal disclosure in the same language as the ad.
  • The 60-second rule: For audio-only ads over 60 seconds, the verbal disclosure must be made at least twice.
  • Chatbot / conversational agents must disclose at the moment of initiation — not in fine print or after engagement.
  • All visible labels must meet IAB conspicuousness standards — small print does not qualify.

6. How Does C2PA Automate IAB AI Disclosure Compliance?

The Verification Loop

C2PA moves compliance from a "trust-based" model to a "verification-based" model. By providing cryptographic provenance, the framework enables ad-tech interoperability and programmatic enforcement at scale. Creative assets no longer just say they are compliant; they prove it with a cryptographically signed manifest.

One of the most technically significant aspects of the IAB framework is its integration with the Coalition for Content Provenance and Authenticity (C2PA) standard. C2PA is a technical specification for cryptographically signing digital content to establish its provenance — where it came from, what tools were used to create it, and whether it has been altered since publication.

The Custom IAB Assertions

The IAB framework defines two custom C2PA assertions specific to advertising:

These assertions, embedded in a C2PA Content Credentials manifest attached to the ad creative file, allow DSPs, SSPs, and publishers to programmatically verify compliance at the point of impression.

Which Tools Already Write C2PA?

This is a practical question with an uneven answer. As of early 2026:

Tools that write C2PA metadata natively: Adobe Firefly (and the broader Adobe ecosystem via Content Credentials), Leica cameras, certain Microsoft Designer outputs, and a growing number of stock photo platforms.

Tools that do not write C2PA natively (yet): Midjourney, Runway, Sora, ElevenLabs, and most other independent AI creative tools. Content created with these tools requires a separate C2PA signing step to achieve machine-readable compliance.

The practical implication: if your creative workflow relies on tools that do not natively write C2PA, you need a process to add C2PA metadata before the asset enters your ad serving workflow. The IAB framework's adoption of C2PA creates a market signal that will accelerate tool adoption — but in 2026, workflow gaps are common.

C2PA Content Credentials in the Ad Serving Stack — flow showing creative asset through C2PA signing to ad server to publisher to consumer display
C2PA content provenance in the ad tech stack. IAB custom assertions (com.iab.threshold, com.iab.disclosure) travel with the creative asset and are verified at each handoff.

7. Which Laws Mandate AI Disclosure in Advertising? (2026 Guide)

The Path to Legal Compliance

While the IAB framework is voluntary, the laws it maps to are binding. Complying with the IAB standard is the most effective way to satisfy multiple, overlapping regional laws (NY, CA, SK, EU) through a single technical workflow.

Global Comparison Table: AI Disclosure Requirements

JurisdictionLaw / RegulationEnforcement StatusPrimary Focus
New YorkSynthetic Performer LawJune 2026AI replicas of real people (Voice/Likeness)
CaliforniaSB 942August 2026Broad AI-generated content in commerce
South KoreaTelecom Business ActActive NowAll AI-generated advertising content
European UnionEU AI ActJan 2027High-risk AI & synthetic media

New York: Synthetic Performer Law (Effective June 2026)

New York's law is the most specific and most immediately consequential for U.S. advertisers. It requires disclosure when an advertiser uses a "synthetic performer" — an AI replica of a real person's voice, likeness, or performance — in a commercial context. The definition of "synthetic performer" maps closely to the IAB framework's synthetic influencer and audio categories.

Violations carry civil penalties of $1,000 to $5,000 per occurrence. "Per occurrence" is still being interpreted, but the legal consensus suggests it means per ad unit, per flight — making non-compliance at scale a significant financial risk.

The IAB framework's categories and label texts are designed to be compliant with NY requirements. An advertiser following the IAB framework's synthetic influencer and audio rules should be meeting New York's requirements.

Urgent: June 2026 deadlineNY Synthetic Performer Law: What Ad Agencies Need to Know Before June 2026Full breakdown of what the NY law requires specifically, how fines are calculated per occurrence, and what a compliant audio workflow looks like before June.

California: SB 942 (Effective August 2026)

California's SB 942 requires disclosure of AI-generated content in certain commercial contexts. It is broader in scope than New York's law — applying to a wider range of AI-generated content, not only AI personas — but with somewhat less specificity about what disclosure must look like. Following the IAB framework's label texts satisfies SB 942's requirements for content that falls within both frameworks' scopes.

South Korea: Telecommunications Business Act Amendment

South Korea moved fastest of all the major markets. The amended Telecommunications Business Act, which came into force in early 2026, requires disclosure of AI-generated content in digital advertising. For advertisers running campaigns in South Korea, this is already a live compliance obligation.

EU AI Act

The EU AI Act's provisions are being phased in, with the most relevant commercial obligations coming into full force in 2027. The Act takes a risk-tiered approach and requires disclosure of AI-generated or manipulated media in contexts where it could mislead. The IAB framework is largely compatible with EU requirements, though the EU's obligations also extend into algorithmic transparency in ways the IAB framework does not directly address.

Global coverageAI Ad Disclosure Laws by Country: EU, US, South Korea 2026Full jurisdiction comparison: requirements, effective dates, penalty ranges, and how the IAB framework maps onto each law. Includes a downloadable comparison table.

8. The IAB Materiality Checker — Try It Free

Understanding the framework is one thing. Applying it to a specific piece of creative — with its specific content type, specific method of creation, and specific deployment context — requires working through the decision logic.

We built the IAB Materiality Checker to do that in under 90 seconds.


Try the IAB Materiality Checker

The tool does the following:

  1. Asks you to select your content type (image, video, audio, synthetic influencer, or text)
  2. Asks the relevant follow-up questions for that content type (creation method, editing scope, voice type, etc.)
  3. Applies the IAB decision logic to your specific answers
  4. Returns a verdict: Disclosure Required, Not Required, or Recommended (context-dependent)
  5. If disclosure is required, gives you the exact IAB label text to use and where to place it
  6. Shows a jurisdiction breakdown covering IAB, EU AI Act, NY Synthetic Performer Law, California SB 942, and South Korea

9. Who is Responsible for AI Disclosure Compliance?

The Liability Flow

Compliance is a shared responsibility across the advertising ecosystem. While agencies often handle the technical execution (signing/labeling), the brand holds the ultimate financial and reputational liability for non-disclosure.

Advertisers and Brand Teams

You are ultimately responsible for the content you run. If your creative uses AI-generated imagery, AI-synthesized voices, or synthetic personas, you need to know whether disclosure is required. The practical implication: add an AI disclosure check to your pre-flight creative review process. The Materiality Checker takes under 90 seconds per creative asset.

The larger strategic question is governance: who in your organization knows what AI tools are being used in your creative workflow? Many brands are discovering that their agencies or production vendors have integrated AI tools without formal disclosure, leaving the brand holding liability they didn't know they had.

Agencies and Creative Teams

You are producing AI-assisted content at scale, often faster than your compliance review process can keep up. The framework creates a practical need for AI disclosure to become part of the creative production checklist — not a legal afterthought. Teams using Firefly benefit from C2PA being written natively. Teams using Midjourney or Runway need an additional signing step.

The June 2026 New York deadline makes this urgent for any agency with U.S.-based clients running campaigns in New York. "My client didn't ask me to check" is not a defense.

Ad Tech Platforms — DSPs, SSPs, Ad Servers

The IAB framework is a technical specification, not just a policy guide. DSPs and ad servers that accept AI-generated creative need to build toward C2PA manifest verification and should be building toward automated disclosure label enforcement. The com.iab.threshold and com.iab.disclosure assertions are the signals to read.

Publishers that accept third-party ad creative have a platform-safety interest in ensuring that AI-generated content running on their pages is properly labeled. Advertiser non-compliance creates reputational risk for publishers.

The framework's alignment with binding law makes it the starting point for multi-jurisdiction AI compliance. Using the IAB framework as your compliance baseline reduces the risk of fragmented, jurisdiction-by-jurisdiction interpretations that are more likely to create gaps.


10. How to Implement the IAB AI Disclosure Framework: A 6-Step Guide

A practical six-step checklist for getting your organization IAB-compliant.

Step 1

Audit AI Tools

Make a list of every AI tool used anywhere in your creative production workflow. Include tools used by vendors, agencies, and freelancers, not just tools your team uses directly

Step 2

Map to IAB Types

For each tool, identify which of the five content types it generates or modifies. Some tools span multiple types (a video tool that also generates audio, for example).

Materiality Check

Run each active creative through the Materiality Checker.** For any live campaign or upcoming campaign, run each AI-involved creative through the Materiality Checker. Document the results. If disclosure is required and you're not currently disclosing, that's a priority fix.

Step 4

Apply IAB Labels

Use the exact label text specified in the framework (see Section 5). Follow the placement requirements for the relevant content type. Do not paraphrase or improvise the label text — standardized language is how consumers learn to recognize what the labels mean.

C2PA Metadata Signing

If any of your AI tools do not write C2PA metadata natively, add a C2PA signing step to your production workflow before assets are uploaded to your ad server.

Step 6

Update Creative SOPs

AI disclosure should become a standard item on your pre-flight creative checklist, alongside brand safety, legal clearance, and accessibility review. The question was AI used, and if so, does it require a disclosure label? should be answered before every campaign goes live.

11. Frequently Asked Questions

It is an industry standard published by the Interactive Advertising Bureau in January 2026. It defines when AI-generated or AI-modified advertising content requires a consumer-facing disclosure label, what that label must say, and how C2PA metadata can signal compliance machine-readably. It covers five content types: images, video, audio, synthetic influencers, and text.
No. The IAB is an industry body, not a regulator. The framework is voluntary. However, it aligns closely with binding laws in New York (effective June 2026), California (effective August 2026), South Korea (already in effect), and the EU (phasing in through 2027). Complying with the IAB framework is the most practical path to compliance with those laws.
It depends on what was edited. Routine post-production — colour correction, lighting, background cleanup — does not require disclosure. Substantive changes to the appearance of the product or of a real person in the ad do require disclosure. The Materiality Checker takes you through the relevant questions in under 90 seconds.
No. The current IAB framework does not require disclosure of AI-generated text or copy. The framework notes this may evolve, but as of January 2026, text-only AI use is exempt.
A generic AI voice — one with no connection to any real person — does not require disclosure under the IAB framework. If the voice is a replica of a real person (living or deceased), different rules apply and disclosure is very likely required.
The framework defines synthetic influencer broadly: any AI-generated or AI-modified representation of a person appearing in advertising content. This includes digital twins of real people (living or deceased), fully synthetic AI personas, and AI-powered interactive chatbots appearing in ads.
A visible label is a consumer-facing disclosure (like 'AI-generated image') that a human consumer can see and understand. A C2PA signal is a hidden, machine-readable cryptographic manifest embedded in the file. While the visible label communicates with humans, the C2PA signal communicates with ad-tech platforms (DSPs, SSPs) to automate compliance. The IAB framework requires both for full compliance: the C2PA manifest acts as the 'proof' that the required visible disclosure has been applied.
Not necessarily. The disclosure requirement is determined by the content, not the tool. If you create AI imagery with Midjourney that meets the materiality threshold — realistic-looking imagery a consumer could mistake for real photography — you need to add a disclosure label. You also need to add C2PA metadata separately, since Midjourney does not write it natively. If your Midjourney output is clearly stylized and non-realistic, no disclosure is required.
At the IAB level: nothing enforceable. At the legal level: fines of $1,000–$5,000 per violation in New York (from June 2026), similar exposure in California (from August 2026), and enforcement actions under South Korean and EU law for campaigns running in those jurisdictions. The larger risk for most brands is reputational — consumer backlash when AI use is discovered without disclosure tends to be significantly more damaging than the regulatory fine.
Run it once per unique AI-generated or AI-modified creative asset. If you use the same asset across multiple placements or flights, one check covers all deployments of that asset. If you modify the asset (changing the AI-generated elements), run the check again.
The IAB-specified label text is: 'AI-generated video'. This exact wording must be used — do not paraphrase. The label must appear from the first frame and persist for the entire duration of the video. It cannot be displayed briefly at the start and then removed.
The required label is 'AI-generated person'. It must appear from the first frame and remain visible for the entire duration that the synthetic avatar appears on screen. If the avatar is present throughout the ad, the label must be present throughout the ad.
For all video content types, the IAB framework requires the label to persist for the entire duration of the video — not just a minimum number of seconds. There is no '2-second rule' in the January 2026 framework. The label must be visible for as long as the AI-generated content is on screen.
For audio-only ads (radio, podcast, smart speaker) longer than 60 seconds, the spoken verbal disclosure of AI involvement must be made at least twice — not just once. The disclosure must be delivered in the same language as the advertisement, at a normal conversational pace, positioned before or immediately after the AI-generated segment.
Yes, the IAB framework was designed to harmonize with the New York law. By using the specified IAB label texts and following its placement/duration rules (especially for audio and synthetic personas), you should meet the requirements of the New York Synthetic Performer Law effective June 2026.
SB 942 requires clear and conspicuous disclosure but does not mandate a specific icon. However, it explicitly recognizes industry standards. Following the IAB’s recommendation (either the text label, Sparkle ✨, or C2PA CR icon) is the most reliable way to satisfy the 'conspicuous' requirement of California's law.

If this guide raised questions you want to go deeper on, here are the articles in this series that cover each topic in detail.