• 0 Posts
  • 24 Comments
Joined 1 year ago
cake
Cake day: August 2nd, 2023

help-circle











  • There’s an old saying in Tennessee — I know it’s in Texas, probably in Tennessee — that says, fool me once, shame on — shame on you. Fool me — you can’t get fooled again.

    A few things:

    • Unity is still bleeding money. They have a product that could be the basis for a reasonably profitable company, but spending billions on a microtransaction company means it is not sufficient for their current leadership. It doesn’t seem wise to build your bussniess on the product of a company whose bussniess plan you fundamentally disagree with.

    • It would be the best for the long term health of bussniess-to-bussnies services if we as a community manages to send the message that it doesn’t matter what any contract says - just trying to introduce retroactive fees is unforgivable and a death sentence to the company that tries it.



  • I understand LLaMA and some other models come with instructions that say that they cannot be used commercially. But, unless the creators can show that you have formally accepted a license agreement to that effect, on what legal grounds can that be enforceable?

    If we look at the direction US law is moving, it seems the current legal theory is that AI generated works fall in the public domain. That means restricting their use commercially should be impossible regardless of other circumstances - public domain means that anyone can use them for anything. (But it also means that your commercial use isn’t protected from others likewise using the exact same output).

    If we instead look at what possible legal grounds restrictions on the output of these models could be based on if you didn’t agree to a license agreement to access the model. Copyright don’t restrict use, it restricts redistribution. The creators of LLMs cannot reasonably take the position that output created from their models is a derivative work of the model, when their model itself is created from copyrighted works, many of which they have no right to redistribute. The whole basis of LLMs rest on that “training data” -> “model” produces a model that isn’t encumbered by the copyright of the training data. How can one take that position and simultaneously belive “model” -> “inferred output” produces copyright encumbered output? That would be a fundamentally inconsistent view.

    (Note: the above is not legal advice, only free-form discussion.)


  • While a broad concept, in the context of your question, science is a metod to derive knowledge from observations.

    Alternatives to the scientific method is to guess or to obtain knowledge from others. (Most other ways I can come up with, e.g. “religion” can still be sorted under these two.)

    Obtaining knowledge from others is great, but may not always be available, and the quality of the knowledge derived this way depends on the reliability of the source.

    For the other alternative, every sensible metric shows how science is a better method than guessing to derive knowledge.





  • backgroundcow@lemmy.worldtoMemes@lemmy.mlI'd pick the person.
    link
    fedilink
    arrow-up
    32
    arrow-down
    1
    ·
    1 year ago

    Treating the question seriously, I think it is easy to think the one person is the right choice, because it just seems like less of a mess to deal with. However, I also think the end result is that you will never feel safe in your home again. You will always second guess if every squeak you hear is yet another person who somehow have made it into your house. This is the path towards slowly descending into madness.