• Alex@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 day ago

    No, that’s why the author asserts that with their signed-of-by. It’s what I do if I use any LLM content as the basis of my patches.

    • ell1e@leminal.space
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      7 hours ago

      So what does the signed-off-by magically solve here, that doesn’t require either you or the contributor to legally review every line by an LLM? If you’re not a lawyer, is your contributor going to be one?

      • Alex@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        6 hours ago

        They don’t have to be. They know what they asked the LLM to do. They know how much they adapted the output. You usually have to work to get the models to spit out significant chunks of memorised text.

        • ell1e@leminal.space
          link
          fedilink
          English
          arrow-up
          1
          ·
          6 hours ago

          I don’t have much more to say other than I doubt the data backs up what you’re saying at all.