• BigFig@lemmy.world
    link
    fedilink
    English
    arrow-up
    54
    arrow-down
    2
    ·
    1 year ago

    What do you do when ChatGPT just makes shit up or answers incorrectly to yes or no questions, you’d have no way of knowing it was wrong

    • gridleaf@lemmy.world
      link
      fedilink
      English
      arrow-up
      15
      ·
      1 year ago

      ChatGPT is most useful when you may not know the right answer, but you know a wrong answer when you see one. It’s very useful for technical issues. Much quicker for troubleshooting than searching page after page for a solution.

    • OtterA
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Not the other commenter:

      I usually have an idea about the thing I’m asking, and if not then I’ll look up the topics mentioned after some guided brainstorming

      I’ve also found that asking the same question again, after resetting the chat, can give you an idea of what is happening

    • IronKrill
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      1 year ago

      While this is an important thing to understand about AI, it’s an overstated issue once understood. For most questions I ask AI, it doesn’t matter if it’s correct as long as it pulls some half useful info to get me on track (i.e programming). For other questions, I only ask it if I need to figure out where to look next, which it will usually do just fine.

      The first page of my search results is all AI generated garbage articles anyway, at least I know what I am getting with GPT and can take it as such.

      • Womble@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        Yup, as long as you are aware that it could be wrong and look at it critically LLMs at GPT scale are very useful tools. The best way I’ve heard it described is having a lightning fast intern who often gets things wrong but will always give it a go.

        So long as you’re calibrated to “how might this be wrong” when looking at the results it is exceptionally useful.