The creator of Nearby Glasses made the app after reading 404 Media’s coverage of how people are using Meta’s Ray-Bans smartglasses to film people without their knowledge or consent. “I consider it to be a tiny part of resistance against surveillance tech.”

more at: @feed@404media.co

  • VindictiveJudge@lemmy.world
    link
    fedilink
    English
    arrow-up
    9
    ·
    2 hours ago

    Drop the cameras and microphones and replace them with a couple accelerometers and gyros. Paired with your phone’s GPS tracking, the glasses can tell where you’re looking without actually seeing anything. You can get handy features like a floating ‘turn here’ sign over your exit while driving with GPS navigation without recording anyone or anything at any time.

    • ArcaneSlime@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      10
      ·
      2 hours ago

      Tbh I don’t even mind cameras that much if they were entirely controlled by the individuals themselves. I have a much bigger issue with it when you’re streaming my facial recognition data to Evil Megacorp 2™ servers that also feed directly to the “Not Spying… Again” agency, though.

    • Except that one cool thing with AR is being able to have it tell what you’re looking at is. Not just positioning things in space. A lot of cool shit that could be done with AR needs some kind of camera, even if it just sees IR light.

    • _haha_oh_wow_@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      2 hours ago

      I don’t think that would work particularly well with AR: People get sick if movement isn’t synced up properly, not having any sort of cameras or sensors at all would exacerbate that problem.

      If you are talking about a simple HUD, then that might be a lot more viable, but for AR and the tech we currently have, some sort of camera or sensor array is kind of a requirement practically speaking.