• Clocks [They/Them]@lemmy.ml
    link
    fedilink
    English
    arrow-up
    4
    ·
    5 days ago

    So instead of recognizing characters…

    1. Compress page / text into a handful of pixels.
    2. Feed pixels into a generative AI.
    3. Hope for the best.

    I rather just use existing OCRs which can be easily backtracked in how they processed text.

    • FauxLiving@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      5 days ago

      They were able to efficiently encode visual information to be used by further networks. In this case the further network was a language model trained on an OCR task.

      The news is the technique, the OCR software is a demonstration of the technique. Encoding visual information efficiently is also key for robotics which use trained networks in their feedback control loops. Being able to process 10 times as much visual data with the same hardware is a very significant increase in capability.

    • Moidialectica [he/him, comrade/them]@hexbear.net
      link
      fedilink
      English
      arrow-up
      2
      ·
      4 days ago

      it doesn’t actually process text, which is why it’s more efficient, it can essentially take in ten times the text through images without suffering the penalties associated by having that many tokens