ChatGPT hides copyright training data, research finds

ChatGPT is trying to hide that it was trained on copyrighted material, according to new research published in a paper by a group of AI scientists from ByteDance.

The researchers found that ChatGPT now disrupts its outputs when users try to extract the next sentence from a prompt. This is a new behavior that was not present in previous versions of ChatGPT.

The researchers believe that ChatGPT developers have implemented a mechanism to detect if the prompts aim to extract copyright content. They also found that ChatGPT still responds to some prompts with copyrighted material, even with these new measures in place.

This is not the only LLM that has been found to contain copyrighted material. Other LLMs, such as OPT-1.3B from Meta and FLAN-T5 from Google, have also been found to respond to prompts with copyrighted text.

The researchers suggest that this is because LLMs are trained on massive amounts of data, including text from books, articles, and websites. This data often includes copyrighted material, which can then be inadvertently reproduced by the LLMs.

The sources for this piece include an article in BusinessInsider.

Top Stories

Related Articles

June 20, 2024 Target is introducing a new generative artificial intelligence tool aimed at enhancing the efficiency of its store employees more...

June 13, 2024 Generative AI tools are transforming the coding landscape, making both skilled and novice developers more efficient. However, the more...

May 16, 2024 Microsoft's ambitious strides in AI technology are now posing a significant challenge to its own climate goals, as more...

May 15, 2024 Ilya Sutskever, co-founder and chief scientist of OpenAI, has officially announced his departure from the company. This move more...

Jim Love

Jim Is and author and pud cast host with over 40 years in technology.