Summary
– Method #1 involves using the “Google Extended” command in Robots.txt to prevent Google from using your content for AI training, but it may have drawbacks such as removing your site from AI training entirely and potentially harming your website’s visibility.
– Method #2 involves using the “Disallow” command in Robots.txt to block Googlebot from accessing your site, which completely removes your content from Google search results but also eliminates all organic search traffic from Google and other search engines.
– Both methods have pros and cons, with potential impacts on search rankings, AI Overviews, and the licensing of content by Google.
– It is recommended for bloggers and publishers affected by content theft to consider blocking Google’s AI training using the Google-Extended robots.txt command until incentives are realigned for participation.
Update: Google has stated they will train their AI on content even if websites opt out. Read more here: https://dig.watch/updates/google-admits-using-opted-out-content-for-ai-training
Sidebar: Hey, did you know you can remove sections of Google, Bing, and other websites you don’t like on desktop with a simple toggle switch? We built Ultimasaurus the Chrome extension that puts a ton of customizations and tools in your hands for free to help boost your productivity. Don’t like AI Overviews? Toggle them off, want to try a search with them? Toggle them back on and refresh the page easily and seamlessly. Tons of other features like “Focus Mode” where all social media sites are turned off, remove nagging YouTube popups, take screenshots, and eliminate stories on Facebook: Get it on the Chrome Store