Is ChatGPT Use Of Web Content Fair? - Search Engine Journal

Some are uncomfortable with how ChatGPT uses their web content to train and learn from

  • There is a way to block your content from being used to train large language models like ChatGPT
  • Intellectual Property Law specialist says technology has outstripped the ability of copyright laws to keep up
  • A search marketing expert questions if the use of Internet content by AI without permission is fair

Replace (not provided) with ALL of your organic keywords inside of Adobe & Google Analytics. Analyze performance by 400+ dimensions and metrics.

TRY FOR FREE

Large Language Models (LLMs) like ChatGPT train using multiple sources of information, including web content. This data forms the basis of summaries of that content in the form of articles that are produced without attribution or benefit to those who published the original content used for training ChatGPT.

Search engines download website content (called crawling and indexing) to provide answers in the form of links to the websites.

Website publishers have the ability to opt-out of having their content crawled and indexed by search engines through the Robots Exclusion Protocol, commonly referred to as Robots.txt.

The Robots Exclusions Protocol is not an official Internet standard but it’s one that legitimate web crawlers obey.

Should web publishers be able to use the Robots.txt protocol to prevent large language models from using their website content?

Large Language Models Use Website Content Without Attribution

Some who are...



Read Full Story: https://news.google.com/__i/rss/rd/articles/CBMiTmh0dHBzOi8vd3d3LnNlYXJjaGVuZ2luZWpvdXJuYWwuY29tL2lzLWNoYXRncHQtdXNlLW9mLXdlYi1jb250ZW50LWZhaXIvNDc3NTU4L9IBAA?oc=5

Your content is great. However, if any of the content contained herein violates any rights of yours, including those of copyright, please contact us immediately by e-mail at media[@]kissrpr.com.



Published by: Book Club