• Rhaedas@fedia.io
    link
    fedilink
    arrow-up
    24
    ·
    7 months ago

    The narrow purpose models seem to be the most successful, so this would support the idea that a general AI isn’t going to happen from LLMs alone. It’s interesting that hallucinations are seen as a problem yet are probably part of why LLMs can be creative (much like humans). We shouldn’t want to stop them, but just control when they happen and be aware of when the AI is off the tracks. A group of different models working together and checking each other might work (and probably has already been tried, it’s hard to keep up).

    • dan1101@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      ·
      7 months ago

      Yeah the hallucinations could be very useful for art and creative stepping stones. But not as much for factual information.