• Simon 𐕣he 🪨 Johnson@lemmy.ml
    link
    fedilink
    English
    arrow-up
    4
    ·
    edit-2
    6 days ago

    From my other comment about o1 and o3/o4 potential issues:

    The other big difference between o1 and o3 and o4 that may explain the higher rate of hallucinations is that the o1’s reasoning is not user accessible, and it’s purposefully trained to not have safe guards on reasoning. Where o3 and o4 have public reasoning and reasoning safeguards. I think safeguards may be a significant source of hallucination because they change prompt intent, encoding and output. So on a non-o1 model that safeguard process is happening twice per turn once for reasoning and once for output, then being accumulated into the input. On an o1 model that’s happening once per turn only for output and then being accumulated.