• 0 Posts
  • 67 Comments
Joined 1 year ago
cake
Cake day: June 15th, 2023

help-circle
  • I never anthropomorphized the technology, unfortunately due to how language works it’s easy to misinterpret it as such. I was indeed trying to explain overfitting. You are forgetting the fact that current AI technology (artificial neural networks) are based on biological neural networks. There is a range of quirks that it exhibits that biological neural networks do as well. But it is not human, nor anything close. But that does not mean that there are no similarities that can be rightfully pointed out.

    Overfitting isn’t just what you describe though. It also occurs if the prompt guides the AI towards a very specific part of it’s training data. To the point where the calculations it will perform are extremely certain about what words come next. Overfitting here isn’t caused by an abundance of data, but rather a lack of it. The training data isn’t being produced from within the model, but as a statistical inevitability of the mathematical version of your prompt. Which is why it’s tricking the AI, because an AI doesn’t understand copyright - it just performs the calculations. But you do. And so using that as an example is like saying “Ha, stupid gun. I pulled the trigger and you shot this man in front of me, don’t you know murder is illegal buddy?”

    Nobody should be expecting a machine to use itself ethically. Ethics is a human thing.

    People that use AI have an ethical obligation to avoid overfitting. People that produce AI also have an ethical obligation to reduce overfitting. But a prompt quite literally has infinite combinations (within the token limits) to consider, so overfitting will happen in fringe situations. That’s not because that data is actually present in the model, but because the combination of the prompt with the model pushes the calculation towards a very specific prediction which can heavily resemble or be verbatim the original text. (Note: I do really dislike companies that try to hide the existence of overfitting to users though, and you can rightfully criticize them for claiming it doesn’t exist)

    This isn’t akin to anything human, people can’t repeat pages of text verbatim like this and no toddler can be tricked into repeating a random page from a random book as you say.

    This is incorrect. A toddler can and will verbatim repeat nursery rhymes that it hears. It’s literally one of their defining features, to the dismay of parents and grandparents around the world. I can also whistle pretty much my entire music collection exactly as it was produced because I’ve listened to each song hundreds if not thousands of times. And I’m quite certain you too have a situation like that. An AI’s mind does not decay or degrade (Nor does it change for the better like humans) and the data encoded in it is far greater, so it will present more of these situations in it’s fringes.

    but it isn’t crafting its own sentences, it’s using everyone else’s.

    How do you think toddlers learn to make their first own sentences? It’s why parents spend so much time saying “Papa” or “Mama” to their toddler. Exactly because they want them to copy them verbatim. Eventually the corpus of their knowledge grows big enough to the point where they start to experiment and eventually develop their own style of talking. But it’s still heavily based on the information they take it. It’s why we have dialects and languages. Take a look at what happens when children don’t learn from others: https://en.wikipedia.org/wiki/Feral_child So yes, the AI is using it’s training data, nobody’s arguing it doesn’t. But it’s trivial to see how it’s crafting it’s own sentences from that data for the vast majority of situations. It’s also why you can ask it to talk like a pirate, and then it will suddenly know how to mix in the essence of talking like a pirate into it’s responses. Or how it can remember names and mix those into sentences.

    Therefore it is factually wrong to state that it doesn’t keep the training data in a usable format

    If your arguments is that it can produce something that happens to align with it’s training data with the right prompt, well yeah that’s not incorrect. But it is so heavily misguided and borders bad faith to suggest that this tiny minority of cases where overfitting occurs is indicative of the rest of it. LLMs are a prediction machines, so if you know how to guide it towards what you want it to predict, and that is in the training data, it’s going to predict that most likely. Under normal circumstances where the prompt you give it is neutral and unique, you will basically never encounter overfitting. You really have to try for most AI models.

    But then again, you might be arguing this based on a specific AI model that is very prone to overfitting, while I am arguing this out of the technology as a whole.

    This isn’t originality, creativity or anything that it is marketed as. It is storing, encoding and copying information to reproduce in a slightly different format.

    It is originality, as these AI can easily produce material never seen before in the vast, vast majority of situations. Which is also what we often refer to as creativity, because it has to be able to mix information and still retain legibility. Humans also constantly reuse phrases, ideas, visions, ideals of other people. It is intellectually dishonest to not look at these similarities in human psychology and then treat AI as having to be perfect all the time, never once saying the same thing as someone else. To convey certain information, there are only finite ways to do so within the English language.





  • Your first point is misguided and incorrect. If you’ve ever learned something by ‘cramming’, a.k.a. just repeating ingesting material until you remember it completely. You don’t need the book in front of you anymore to write the material down verbatim in a test. You still discarded your training material despite you knowing the exact contents. If this was all the AI could do it would indeed be an infringement machine. But you said it yourself, you need to trick the AI to do this. It’s not made to do this, but certain sentences are indeed almost certain to show up with the right conditioning. Which is indeed something anyone using an AI should be aware of, and avoid that kind of conditioning. (Which in practice often just means, don’t ask the AI to make something infringing)


  • This would be a good point, if this is what the explicit purpose of the AI was. Which it isn’t. It can quote certain information verbatim despite not containing that data verbatim, through the process of learning, for the same reason we can.

    I can ask you to quote famous lines from books all day as well. That doesn’t mean that you knowing those lines means you infringed on copyright. Now, if you were to put those to paper and sell them, you might get a cease and desist or a lawsuit. Therein lies the difference. Your goal would be explicitly to infringe on the specific expression of those words. Any human that would explicitly try to get an AI to produce infringing material… would be infringing. And unknowing infringement… well there are countless court cases where both sides think they did nothing wrong.

    You don’t even need AI for that, if you followed the Infinite Monkey Theorem and just happened to stumble upon a work falling under copyright, you still could not sell it even if it was produced by a purely random process.

    Another great example is the Mona Lisa. Most people know what it looks like and if they had sufficient talent could mimic it 1:1. However, there are numerous adaptations of the Mona Lisa that are not infringing (by today’s standards), because they transform the work to the point where it’s no longer the original expression, but a re-expression of the same idea. Anything less than that is pretty much completely safe infringement wise.

    You’re right though that OpenAI tries to cover their ass by implementing safeguards. Which is to be expected because it’s a legal argument in court that once they became aware of situations they have to take steps to limit harm. They can indeed not prevent it completely, but it’s the effort that counts. Practically none of that kind of moderation is 100% effective. Otherwise we’d live in a pretty good world.



  • Exactly, thinking that is what I was getting pulled me over the edge, I sometimes remember a music video I want to listen to on my phone during my commute and I don’t want to spend 30 minutes either getting on my PC to download it with a tool, or using a third party downloader which can at times be shady. So upgrading that to a single click in the app seemed like a great deal. Crushingly disappointed when I found out how it actually was. Turns out the real answer was NewPipe, which I don’t even have to pay for.


  • Yeah, I thought it was a nice compromise. It seemed sensible that if Premium is the ‘compliant’ response to not wanting ads, the ‘compliant’ response to using third party tools to download videos was to just be able to do things more easily and with more options through Premium as well. But apparently they wanted to advertise something anyone who’s wanted to download a youtube video would not describe as ‘downloading’, which is easily out competed by free (but at times shady) tools.



  • I like those perks too, but if I pay more to be able to download videos (which again, I could’ve used a free tool for) I want to be able to do whatever I want with it. Download means getting a file I can watch using my own video player and store for later even if Youtube dies tomorrow, If I go on holiday without internet, or if my internet goes down for a week. Anything.

    If Google is going to be “Uhm aksually, you are technically downloading it, thats why we can advertise it like that”, then I’m already downloading literally every video I watch. And thats not the kind of bullshit you give to a paying customer. That is spitting in my face for paying you. Why does a non-Premium user get better service with free third party youtube downloaders?

    It’s a matter of principle.


  • ClamDrinker@lemmy.worldtoMemes@lemmy.mlMe but ublock origin
    link
    fedilink
    arrow-up
    96
    arrow-down
    3
    ·
    edit-2
    26 days ago

    I gave Premium a shot. Then the one time I wanted to use the feature Google said I was paying for - being able to download videos - I found out that it was just a glorified pre-buffer.

    • Can’t view the video outside the youtube app or the website, source video file encrypted ✅️
    • Can’t view the video if you havent connected to the internet in 3 days ✅️
    • Does less than your average youtube downloader that you can find for free with one search ✅️
    • Literally just saving Youtube bandwidth because they destroyed every benefit you would get if it was actually reasonable ✅️

    Enshittification isnt just limited to free users folks. Slammed that cancellation button right then and there. Good luck earning back my trust, I’m happy to pay if you didnt scream so loudly that even if I paid, you were going to treat me like shit anyways.


  • I am kind of afraid that if voting becomes more public than it already is, it will lead exactly to more of the kind of “zero-content downvote” accounts mentioned in the ticket. Because some people are just wildly irrational when it comes to touchy subjects, and aint nobody got time to spend an eternity with them dismantling their beliefs so they understand the nuance you see that they don’t (If they even let you). So it kind of incentivizes people to create an account like that to ensure a crazy person doesn’t latch on to the account you’re trying to have normal discussions with.

    But I understand that they can technically already do this if they wanted to. So perhaps it will be fine as long as we fight against vote viewing being weaponized as a community.


  • I respectfully disagree. Sure, it didn’t cure the world of ignorant people like we hoped, but they are not the average rational person. It massively increased the awareness of people about international issues like climate change, racism, injustice, and allowed people to forge bonds abroad far more easily. The discourse even among ignorant people is different from 20 years ago. However, the internet that did that might no longer be the same one it is today.

    But honestly, “more facts leads to more truth” wasn’t the point of my message. It was “more spread of falsehoods leads to higher standards of evidence to back up the actual truth”, which isn’t quite the same. Before DNA evidence and photographic / video evidence, people sometimes had to rely on testimony. Nowadays if someone tells you a story that screams false you might say “pics or it didn’t happen.”. That’s the kind of progress I’m referring to.

    Someone presenting you only a single photo of something damning is the hearsay of yesterday. (And honestly, it’s been that way since Photoshop came out, but AI will push that point even further)


  • I have a similar hesitancy, but unfortunately that’s why we can’t even really trust ourselves either. The statistics we can put to paper already paints such a different image of society than the one we experience. So even though it feels like these people are everywhere and such a mindset is growing, there are many signs that this is not the case. But I get it, that at times also feels like puffing some hopium. I’m fortunate to have met enough stubborn people that did end up changing their minds on their own personal irrationality, and as I grew older I caught myself doing the same a couple of times as well. That does give me hope.

    And well, if you look at history, the kind of shit people believed. Miasma, bloodletting, superstitious beliefs, to name a few. As time has moved on, the majority of people has grown. Even a century where not a lot changes in that regard (as long as it doesn’t regress) can be a speed bump in the mindset of the future.


  • While I share this sentiment, I think/hope the eventual conclusion will be a better relationship between more people and the truth. Maybe not for everyone, but more people than before. Truth is always more like 99.99% certain than absolute truth, and it’s the collection of evidence that should inform ‘truth’. The closest thing we have to achieving that is the court system (In theory).

    You don’t see the electric wiring in your home, yet you ‘know’ flipping the switch will cause electricity to create light. You ‘know’ there is not some other mechanism in your walls that just happens to produce the exact same result. But unless you check, you technically didn’t know for sure. Someone could have swapped it out while you weren’t looking, even if you built it yourself. (And even if you check, your eyes might deceive you).

    With Harris’ airport crowd, honestly if you weren’t there, you have to trust second hand accounts. So how do you do that? One video might not say a lot, and honestly if I saw the alleged image in a vacuum I might have been suspicious of AI as well.

    But here comes the context. There are many eye witness perspectives where details can be verified and corroborated. The organizer isn’t an habitual liar. It happened at a time that wasn’t impossible (eg. a sort of ‘counter’-alibi). It happened in a place that isn’t improbable (She’s on the campaign trail). If true, it would require a conspiracy level of secrecy to pull of. And I could list so many more things.

    Anything that could be disproven with ‘It might have been AI’, probably would have not stuck in court anyways. It’s why you take testimony, because even though that proves nothing on it’s own, if corroborated with other information it can make one situation more or less probable.