|
- Meta AI can reproduce half of Harry Potter book verbatim
A recent study has revealed that Meta's LLaMA 3 1 AI model can reproduce significant portions of copyrighted texts, including the first Harry Potter book The research was conducted by a group of
- Metas Llama 3. 1 can recall 42 percent of the first Harry . . .
Maybe Meta had trouble finding 15 trillion distinct tokens, so it trained on the Books3 dataset multiple times Or maybe Meta added third-party sources—such as online Harry Potter fan forums, consumer book reviews, or student book reports—that included quotes from Harry Potter and other popular books
- New study reveals how much Meta’s AI is copying famous books . . .
Meta’s Llama 3 1 stood out for remembering big parts of well-known books like The Hobbit, 1984, and Harry Potter and the Sorcerer’s Stone Older models, like Llama 1, only memorized about 4%
- Metas Llama 3. 1 Can Recall 42% of the First Harry Potter Book
Meta's Llama 3 1 can recall 42 percent of the first Harry Potter book New research could have big implications for copyright lawsuits against generative AI Timothy B Lee - Specifically, the paper estimates that Llama 3 1 70B has memorized 42 percent of the first Harry Potter book well enough to reproduce 50-token excerpts at least half the time
- Meta Llama 3. 1: New discoveries in AI memory and potential . . .
Recently, a study by Stanford, Cornell, and West Virginia University showed that Meta Llama 3 1 (70B), released in July 2024, consciously memorised 42% of the first Harry Potter book so that it can reproduce passages of 50 tokens with a probability of over 50% In other words, the model doesn't just remember the plot - it can reproduce entire chunks of text if given the first sentence For
- Metas AI memorised books verbatim – that could cost it billions
Many AI models were trained on the text of books, but a new test found at least one model has directly memorised nearly the entirety of some books, including Harry Potter and the Philosopher’s
- Metas Llama 3. 1 can recall 42 percent of the first Harry . . .
https: archive is OSQt6 If you've seen as many magnet links as I have, with your subconscious similarly primed with the foreknowledge of Meta having used torrents to download leech (and possibly upload seed) the dataset(s) to train their LLMs, you might scroll down to see the first picture in this article from the source paper, and find uncanny the resemblance of the chart depicted to a
|
|
|