So does any site that quotes the book. Just being trained on a work doesn’t give the model the ability to cite it word for word. For most of the books in this set you wouldn’t even be able to get a single accurate quote out of most models. The models gain the ability to cite passages from training on other sources citing these same passages.
It shares popular quotes from books, it can’t reproduce arbitrary content from a book. The content needs to be heavily duplicated in the training data to stick around (e.g. from book reviews), and even than half of it might still end up being made up on the spot.
Also request for copyrighted content will be blocked by ChatGPT and just receive the stock “I can’d do that” response anyway.
If you have some damning examples that show the opposite, show them.
Being blocked by ChatGPT just means that the interaction layer you see doesn’t show the output, not that the output wasn’t generated.
Everything you see that’s public facing and interfacing with an AI is an extreme filtering layer for what is output. There’s tons of checks that happen to ensure that they don’t output illegal content or any of a million other undesirable things.
I’m too lazy and care too little but you can basically get it to roleplay as a book expert or something and to “remind” you of certain passages. It gets around the filter pretty easily, that’s how jailbreaks work.
Don’t bother shouting into the AI misinformation void.
People aren’t going to put down their pitchforks and torches to brush up on basic ML principles and it’s just going to frustrate you engaging.
It’s going to be a non-issue within 24 months anyways.
No matter how the OpenAI court cases land, the writing is on the wall that the next generations of models are going to be built on the backs of synthetic data, which is inherently without copyright.
At best rulings against OpenAI mean a secondary market emerges in China for repackaging copyrighted data into synthetic data of equivalent value to help buffer SotA synthetic data in avoiding model collapse.
It’s not even going to end up amounting to a minor speedbump to progress by the time the court cases are finalized.
Let the armchair activists rant and rage and tire themselves out worrying about a fabricated version of reality, and just focus more on staying informed about actual reality for yourself when all this passes.
It will be years before people eventually drop the bias against AI we self-instilled from shortsighted Sci-Fi over the past few decades, and until then the average person online will be irrationally upset about something related to the tech. Might as well run themselves ragged over the misinformed “it just remixes copyright” in the meantime.
On the one hand, I agree with your estimation of how things will go overall.
On the other hand, though, I think there’s value to be had in pushing back against the misinformation whenever it comes up. I don’t think AI is going to be hindered by it in the long run, but it’s possible that in the short run it’s going to kill interesting projects and harm some of the people who are experimenting with it.
And I have seen technologies that have suffered from longer-term difficulties once the zeitgeist turned against them despite having technical merit. There are useful applications for NFTs to be had out there, for example, but just try mentioning them when the opportunity arises and see what sort of reaction you get.
Yes, though to be fair these things often move in pendulum swings back and forth and that’s a natural component of any system finding equilibrium.
Just as AI has many detractors raised to fear HAL or the Terminator and approaching any news of what’s perceived as the existential threat they were warned of with hyper caution, there’s also a ton of hypemasters packaging up snake oil with AI having become the new ‘quantum’ to slap on a pile of crap and claim is gold.
The two are going to balance out.
NFTs were hyped to shit on numerous get rich quick pyramid schemes, and a number of companies that were jumping on the bandwagon to try and catch a wave rightfully abandoned what was going to be a terrible idea (i.e. Square Enix).
That said, the technology isn’t going anywhere and I’m sure we’ll see peer to peer cryptography backed exchanges of goods and services continuing to work their way into future products where the technology makes sense on its own merits as opposed to hype cycles.
The utility of AI, and specifically LLMs, is so astounding right now even in its infancy that it’s not going anywhere no matter where public opinion sits. It just won’t necessarily be used as a selling point, like for a new Coke flavor. Which is ultimately going to be a good thing.
I agree that misinformation tends to be bad, and I do have legitimate concerns that the feverous anti-AI crowd is going to end up cutting off their nose to spite their face driving a technological revolution behind closed doors of international conglomerates rather than open access, but at a certain point pretty soon this ship is going to be out of anyone’s control, and just as the DCMA doesn’t actually prevent me from downloading The Matrix right now vs in 2000 outside of a few extra hoops, the likely eventual “let’s try to handicap AI Act” is probably not going to prevent me from running model weights published in Israel or Japan on a local GPU.
I used to get more stressed about the rhetoric online, but it’s reached a point where it’s clear 90% of people aren’t looking for facts or understanding, they are only seeking confirmation bias and down voting anything that doesn’t deliver it.
In that climate, why waste our time? Discussions where one stands to learn through contribution and formulating a comment are still probably worthwhile, but a lot of discussion of AI in more general forums have honestly just turned into tantrums where no one wants to have their outrage party rained on.
It’s become the equivalent of explaining the science of immunity to antivax crowds.
AI isn’t either. It’s selling statistical data about the books.
It literally shares passages verbatim
So does any site that quotes the book. Just being trained on a work doesn’t give the model the ability to cite it word for word. For most of the books in this set you wouldn’t even be able to get a single accurate quote out of most models. The models gain the ability to cite passages from training on other sources citing these same passages.
It shares popular quotes from books, it can’t reproduce arbitrary content from a book. The content needs to be heavily duplicated in the training data to stick around (e.g. from book reviews), and even than half of it might still end up being made up on the spot.
Also request for copyrighted content will be blocked by ChatGPT and just receive the stock “I can’d do that” response anyway.
If you have some damning examples that show the opposite, show them.
Being blocked by ChatGPT just means that the interaction layer you see doesn’t show the output, not that the output wasn’t generated.
Everything you see that’s public facing and interfacing with an AI is an extreme filtering layer for what is output. There’s tons of checks that happen to ensure that they don’t output illegal content or any of a million other undesirable things.
I’m too lazy and care too little but you can basically get it to roleplay as a book expert or something and to “remind” you of certain passages. It gets around the filter pretty easily, that’s how jailbreaks work.
That’s maybe an issue. I mirror speech a lot, though. How large are the passages?
That claim is disingenuous at best, and misinformed otherwise.
deleted by creator
LLMs can’t reprint their entire training data on demand. They rarely even remember quotes.
Don’t bother shouting into the AI misinformation void.
People aren’t going to put down their pitchforks and torches to brush up on basic ML principles and it’s just going to frustrate you engaging.
It’s going to be a non-issue within 24 months anyways.
No matter how the OpenAI court cases land, the writing is on the wall that the next generations of models are going to be built on the backs of synthetic data, which is inherently without copyright.
At best rulings against OpenAI mean a secondary market emerges in China for repackaging copyrighted data into synthetic data of equivalent value to help buffer SotA synthetic data in avoiding model collapse.
It’s not even going to end up amounting to a minor speedbump to progress by the time the court cases are finalized.
Let the armchair activists rant and rage and tire themselves out worrying about a fabricated version of reality, and just focus more on staying informed about actual reality for yourself when all this passes.
It will be years before people eventually drop the bias against AI we self-instilled from shortsighted Sci-Fi over the past few decades, and until then the average person online will be irrationally upset about something related to the tech. Might as well run themselves ragged over the misinformed “it just remixes copyright” in the meantime.
On the one hand, I agree with your estimation of how things will go overall.
On the other hand, though, I think there’s value to be had in pushing back against the misinformation whenever it comes up. I don’t think AI is going to be hindered by it in the long run, but it’s possible that in the short run it’s going to kill interesting projects and harm some of the people who are experimenting with it.
And I have seen technologies that have suffered from longer-term difficulties once the zeitgeist turned against them despite having technical merit. There are useful applications for NFTs to be had out there, for example, but just try mentioning them when the opportunity arises and see what sort of reaction you get.
Yes, though to be fair these things often move in pendulum swings back and forth and that’s a natural component of any system finding equilibrium.
Just as AI has many detractors raised to fear HAL or the Terminator and approaching any news of what’s perceived as the existential threat they were warned of with hyper caution, there’s also a ton of hypemasters packaging up snake oil with AI having become the new ‘quantum’ to slap on a pile of crap and claim is gold.
The two are going to balance out.
NFTs were hyped to shit on numerous get rich quick pyramid schemes, and a number of companies that were jumping on the bandwagon to try and catch a wave rightfully abandoned what was going to be a terrible idea (i.e. Square Enix).
That said, the technology isn’t going anywhere and I’m sure we’ll see peer to peer cryptography backed exchanges of goods and services continuing to work their way into future products where the technology makes sense on its own merits as opposed to hype cycles.
The utility of AI, and specifically LLMs, is so astounding right now even in its infancy that it’s not going anywhere no matter where public opinion sits. It just won’t necessarily be used as a selling point, like for a new Coke flavor. Which is ultimately going to be a good thing.
I agree that misinformation tends to be bad, and I do have legitimate concerns that the feverous anti-AI crowd is going to end up cutting off their nose to spite their face driving a technological revolution behind closed doors of international conglomerates rather than open access, but at a certain point pretty soon this ship is going to be out of anyone’s control, and just as the DCMA doesn’t actually prevent me from downloading The Matrix right now vs in 2000 outside of a few extra hoops, the likely eventual “let’s try to handicap AI Act” is probably not going to prevent me from running model weights published in Israel or Japan on a local GPU.
I used to get more stressed about the rhetoric online, but it’s reached a point where it’s clear 90% of people aren’t looking for facts or understanding, they are only seeking confirmation bias and down voting anything that doesn’t deliver it.
In that climate, why waste our time? Discussions where one stands to learn through contribution and formulating a comment are still probably worthwhile, but a lot of discussion of AI in more general forums have honestly just turned into tantrums where no one wants to have their outrage party rained on.
It’s become the equivalent of explaining the science of immunity to antivax crowds.
I mean, yeah? They were running to a concrete description. That is not valid. My brain has most of Terry Pratchett’s works.