• lugal@sopuli.xyz
        link
        fedilink
        arrow-up
        5
        ·
        7 months ago

        Not sure what would frighten me more: the fact that this is trainings data or if it was hallucinated

        • EpeeGnome@lemm.ee
          link
          fedilink
          English
          arrow-up
          4
          ·
          edit-2
          7 months ago

          Neither, in this case it’s an accurate summary of one of the results, which happens to be a shitpost on Quara. See, LLM search results can work as intended and authoritatively repeat search results with zero critical analysis!

      • xavier666@lemm.ee
        link
        fedilink
        English
        arrow-up
        4
        ·
        7 months ago

        Pretty sure AI will start telling us “You should not believe everything you see on the internet as told by Abraham Lincoln”

      • TheFriar@lemm.ee
        link
        fedilink
        arrow-up
        12
        arrow-down
        1
        ·
        7 months ago

        Sure we can. If it gives you bad information because it can’t differentiate between a joke a good information…well, seems like the blame falls exactly at the feet of the AI.

        • kate@lemmy.uhhoh.com
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          1
          ·
          7 months ago

          Should an LLM try to distinguish satire? Half of lemmy users can’t even do that

          • KevonLooney@lemm.ee
            link
            fedilink
            arrow-up
            9
            ·
            7 months ago

            Do you just take what people say on here as fact? That’s the problem, people are taking LLM results as fact.

          • BakerBagel
            link
            fedilink
            arrow-up
            4
            ·
            7 months ago

            It should if you are gonna feed it satire to learn from

          • ancap shark@lemmy.today
            link
            fedilink
            arrow-up
            2
            arrow-down
            1
            ·
            7 months ago

            If it’s being used to give the definite answer of a search, so it should. If it can, than it shouldn’t be used for that