Last weekend I found my article being paraphrased paragraph by paragraph by AI — I froze.

I’d spent three days on an industry breakdown, only to have some chatbot “summarize” it, even keeping the fictional case study I invented. My first reaction was anger; my second was — what can I even do about this? I’ve been stuck here too, thinking I’m just a small blogger, Big Tech won’t notice me. But this Meta lawsuit made me realize: they’re not “noticing you,” they’re bulk-scanning. Your content is in my pile, and in their training data too.

What happened: Zuckerberg personally authorized using copyrighted content to train AI

Multiple publishers sued Meta, and filings show Zuckerberg personally approved using copyrighted content to train the Llama model. In plain terms: top brass at a giant company signed off on feeding other people’s writing to their own AI — no heads-up, no payment. Zhang Wei, a freelance writer in Hangzhou, was sitting in a Starbucks near West Lake last month, searched her paid newsletter on her phone, and found an AI tool nearly paraphrasing her arguments paragraph by paragraph. She screenshot it and sent it to me: “I feel robbed, but I don’t even know who to go after.” I completely understand that helplessness — how do we, as solopreneurs, take on a hundred-billion-dollar company? But at least there are things we can do today.

What you can do today: spend $0 and 10 minutes to reduce what gets scraped

Replication cost: $0 + 10 minutes + technical barrier: just be able to log into your website backend. First step: if you use WordPress, search for the plugin “Block AI Crawlers,” click “Install and Activate,” and it will automatically add rules to your site’s robots.txt (the config file that tells crawlers “which pages not to touch”) that block major AI crawlers. If you don’t use WordPress, manually add two lines to your robots.txt: User-agent: GPTBot and Disallow: /, and similarly add CCBot and Google-Extended. This isn’t a silver bullet — well-behaved crawlers will comply, rogue ones won’t — but it’s at least one more layer of protection. Not everyone needs this right now; it’s fine if you don’t try it today — your content won’t disappear tomorrow.

Advice by stage

If you’re just starting out and don’t have much original content yet: focus on getting things written first. Spend 2 minutes adding the robots.txt lines and don’t overthink copyright — your bigger challenge is getting seen.

If you have 1–2 clients and are starting to have paid content: I’d recommend seriously adding crawler blocking, and also adding a copyright line at the bottom of your posts. This is your asset; it should be properly marked.

If you’re scaling up and already have steady content output: consider registering a DMCA takedown service (about $100/year), and periodically Google your original paragraphs to check if they’ve been ripped. The larger your content library, the higher the chance of being systematically harvested — it’s worth spending that money to protect it.