• 5 Posts
  • 469 Comments
Joined 2 years ago
cake
Cake day: August 27th, 2023

help-circle







  • All LLMs and Gen AI use data they don’t own. The Pile is all scraped or pirated info, which served as a starting point for most LLMs. Image gen is all scraped from the web. Speech to text and video gen mainly uses YouTube data.

    So either you put a price tag on that data, which means only a handful of companies can afford to build these tools (including Meta), or you understand that piracy is the only way for most to aquire this data but since it’s highly transformative, it isn’t breaching copyrights or directly stealing from them as piracy “normally” is.

    I’m being pragmatic.





  • Meta has open sourced every single one of their llms. They essentially gave birth to the whole open llm scene.

    If they start losing all these lawsuits, the whole scene dies and all those nifty models and their fine-tunes get removed from huggingface, to be repackaged and sold to us with a subscription fee. All the other domestic open source players will close down.

    The copyright crew aren’t the good guys here, even if it’s spearheaded by Sarah Silverman and Meta has traditionally played the part of the villain.







  • The context only mattered because you were talking about the bot missing the euphemism. It doesn’t matter if the bot is invested in the fantasy, that is what it’s suppose to do. It’s up to the user to understand it’s a fantasy and not reality.

    Many video games let you do violent things to innocent npcs. These games are invested in the fantasy, as well as trying to immerse you in it. Although It’s not exactly the same, it’s not up to the game or the chatbot to break character.

    Llms are quickly going to be included in video games and I would rather not have safeguards (censorship) because a very small percentage of people with clear mental issues can’t deal with them.




  • I think there’s place for regulation in case of gross negligence or purposefully training it to output bad behavior.

    When it comes to mistakes, I don’t really believe in it. These platforms always have warnings about not trusting what the AI says.

    I like to compare it to users on social media for example. If someone on lemmy told you to use peanut butter, he wouldn’t really be at fault, nor the instance owner.

    AI systems don’t present themselves as scientific papers. If you are taking for truth things random redditors and auto complete bots are saying, that’s on you so to speak.