Your best idea didn't disappear into thin air. It was scraped, tokenised, and fed into a system that now sells it back to the world without your name attached.
That's not a conspiracy theory. That's the documented reality of how most large language models and generative AI systems were built. Billions of web pages, creative portfolios, published articles, design files, forum threads, and yes your work were consumed at scale to train systems that now compete directly with the people who made the training data possible.
And if you're a woman who built a career on creative expertise, this hits differently.
Why Creatives Are the Canary in the Coal Mine
The artists saw it coming first. In 2022, digital illustrators began noticing something unsettling: AI image generators could reproduce their exact stylistic signatures the particular way they blended shadow, chose colour palettes, composed negative space after being prompted with their names. No consent. No contract. No credit.
Sarah Andersen, Karla Ortiz, and Kelly McKernan filed one of the first major class action lawsuits in the US against Stability AI, Midjourney, and DeviantArt. Ortiz, who creates richly textured fantasy illustration, described seeing her aesthetic reproduced by a system that had scraped her entire portfolio without asking.
This isn't a niche problem for graphic designers. It's the blueprint for what happens across every creative profession writing, music, photography, UX design, copywriting, brand strategy. The mechanism is the same: your publicly available creative output becomes unpaid training data, and the resulting model then undercuts your market rate.
The EU's own data shows the creative sector employs 8.7 million people and women are disproportionately concentrated in the roles most exposed to AI displacement: content creation, graphic design, and media production.
The Invisible Theft Nobody Explained to You
Here's what actually happens when you post your work online.
A crawler an automated scraping bot visits your website, your Behance portfolio, your published articles, your Instagram grid. It records everything. That data gets packaged into training datasets with names like LAION-5B or The Pile. Those datasets get used to train a generative AI model. The model learns your patterns, your voice, your visual logic. Then a company monetises that model.
You were never asked. You never agreed. You received nothing.
Is this legal? That's exactly what courts across Europe and the US are now trying to decide.
In the EU, the Copyright in the Digital Single Market Directive (DSM Directive) created a Text and Data Mining (TDM) exemption that allows AI companies to scrape publicly available content unless the rights holder has explicitly opted out. The opt-out mechanism exists but it's deliberately obscure, technically complex, and places the burden entirely on creators.
Fewer than 12% of independent creators in Europe are even aware this opt-out exists, according to a 2024 survey by the European Writers' Council.
The Salary Anchoring Problem Has a New Dimension
Women in creative fields already face a well-documented rate gap. Freelance female designers charge, on average, 1823% less than male counterparts for equivalent work, according to research from the European Institute for Gender Equality. This isn't because women underperform it's the accumulation of anchoring effects, assertiveness penalties, and the structural expectation that women will accept less.
Now add an AI model that can generate "good enough" creative work at near-zero cost.
The market doesn't just pressure women to charge less it now has a technological mechanism to replace them entirely with output derived, in part, from their own unpaid labour.
Think about that sequence carefully. A woman builds a distinctive design aesthetic over ten years. She posts her work publicly because visibility is how freelance creatives build a client base. That work gets scraped. Her aesthetic gets encoded into a model. The model gets licensed to the companies she pitches. She loses contracts because clients say the AI "does something similar."
She wasn't just underpaid. She was robbed of her competitive edge by the very act of building it visibly.
What European Law Actually Says Right Now
[Cost Lever]
The EU AI Act which became fully applicable in August 2024 requires providers of general-purpose AI models to publish summaries of training data used, so that rights holders can identify whether their work was included. It's a significant step. But "significant step" and "adequate protection" are not the same thing.
The practical reality: accessing these summaries requires legal literacy that most individual creators don't have. Enforcement requires filing complaints with national authorities, who are still building capacity to handle AI-related IP cases.
The average cost of IP litigation in Europe ranges from 15,000 to over 200,000 depending on jurisdiction and complexity. That's not a realistic option for a freelance illustrator in Warsaw or a copywriter in Lisbon.
The legal framework exists in embryonic form. The protection it offers right now is mostly theoretical.
The Opt-Out Mechanism: Technically Real, Practically Broken
[Risk Lever]
Under the DSM Directive's TDM exemption, creators can opt out of having their work scraped for AI training. Here's how it works in practice:
For websites, you add a tag to your robots.txt file specifying that AI training crawlers are excluded. Major AI companies including OpenAI and Google have stated they will respect these signals for future training runs. "Future training runs" is doing a lot of work in that sentence. Your work that was already scraped remains in existing models.
For individual pieces of work on platforms like Instagram, LinkedIn, or Behance platforms you don't control the opt-out is determined by the platform's terms of service, which most people agreed to without reading in 2016.
Instagram's terms, updated in 2024, allow Meta to use your publicly posted content to train AI models unless you opt out through a form that's buried four menus deep and only available in certain regions.
The risk here is compounding. Every day you remain opted in through inaction, through ignorance, through the sheer administrative friction of the process is another day your work trains systems designed to replace you.
The Human Edge That Can't Be Scraped
[Quality Lever]
Let's be precise about what AI cannot replicate, because vague reassurances like "human creativity will always matter" are not useful.
What AI systems lack isn't creativity in the narrow technical sense they can generate novel combinations with impressive surface fluency. What they lack is contextual stakes. A human creative working on a campaign for a women's health brand in Hungary understands, implicitly, the cultural weight of certain imagery, the political charge of specific language, the trust economy they're operating inside. That understanding comes from being embedded in a culture, from having a body, from having experienced the consequences of getting it wrong.
AI generates output. Humans generate meaning.
The distinction matters commercially, not just philosophically. Clients who are buying "good enough" creative for internal docs or quick social posts will use AI. Clients who are buying work where precision, cultural nuance, and accountability matter healthcare communications, legal content, politically sensitive campaigns, luxury brand identity still need humans who carry professional risk and professional judgment.
The strategic move isn't to compete on volume or speed. It's to position explicitly in the territory where stakes are high enough that "good enough" isn't acceptable.
Three Things You Can Do This Week
[Speed Lever]
Not in six months when the legal framework is clearer. Not after you've read every article about AI and IP. This week.
First, audit your digital footprint. Use a tool like HaveIBeenTrained.com to check whether your images appear in LAION datasets. For writers, search distinctive phrases from your published work on AI-generated content checkers. This isn't about what you can immediately do with that information it's about knowing the actual state of your situation rather than assuming.
Second, update your robots.txt and platform settings. If you have a personal website, add the relevant AI crawler exclusions to your robots.txt file. The specific tags for GPTBot (OpenAI), ClaudeBot (Anthropic), and Google-Extended are publicly documented and take minutes to implement. On platforms, navigate to privacy or data settings and locate AI training opt-outs they exist on Instagram, LinkedIn, and several others.
Third, document and timestamp your creative process. This is low-tech but legally significant. When you create original work, keep dated records of drafts, briefs, reference files, and development stages. In an IP dispute, provable origination matters enormously. A private folder on your device with timestamped versions of work-in-progress is genuinely useful evidence.
None of these steps are perfect. None of them return work already scraped. But they change your position from entirely passive to actively defended and that shift matters, practically and psychologically.
The Visibility Trap
[Leverage Lever]
Here's the structural cruelty that deserves more anger than it gets: creatives and particularly women building freelance careers were told that online visibility was the path to professional survival. Post consistently. Build a portfolio. Share your process. Grow an audience. Be findable.
That advice was correct for the decade it was given in. It's now the mechanism by which your competitive edge was extracted without compensation.
The women who followed the rules who built visible, distinctive, well-documented creative practices provided the richest training data. The lesson isn't to disappear. Visibility still matters. But visibility without IP strategy is professional self-sabotage in 2025.
The women who will come out of this period with their market position intact are the ones who treat their creative output as an asset with legal dimensions, not just a portfolio with aesthetic dimensions.
That means knowing what TDM opt-outs exist. It means watermarking work with metadata that survives platform compression. It means building client relationships where your judgment, your accountability, and your cultural intelligence are explicitly part of what's being contracted not just the output.
It means understanding that the ghost in the machine learned from you. And deciding you're not going to keep teaching it for free.
The legal battles being fought by artists and writers across Europe and the US will shape what creative IP means for the next generation. The EU AI Act creates accountability mechanisms that didn't exist three years ago. Courts are genuinely uncertain. This is a period where the rules are still being written and that means the creatives who understand the terrain have more leverage than they've been led to believe.
Your work built something. You deserve to know what it built, who's using it, and what tools exist to make that stop.

Checking account status...
Loading comments...