cross-posted from: https://lemmy.ca/post/37011397

!opensource@programming.dev

The popular open-source VLC video player was demonstrated on the floor of CES 2025 with automatic AI subtitling and translation, generated locally and offline in real time. Parent organization VideoLAN shared a video on Tuesday in which president Jean-Baptiste Kempf shows off the new feature, which uses open-source AI models to generate subtitles for videos in several languages.

    • shyguyblue@lemmy.world
      link
      fedilink
      English
      arrow-up
      83
      arrow-down
      1
      ·
      7 months ago

      I was just thinking, this is exactly what AI should be used for. Pattern recognition, full stop.

      • snooggums@lemmy.world
        link
        fedilink
        English
        arrow-up
        33
        arrow-down
        1
        ·
        7 months ago

        Yup, and if it isn’t perfect that is ok as long as it is close enough.

        Like getting name spellings wrong or mixing homophones is fine because it isn’t trying to be factually accurate.

        • vvv@programming.dev
          link
          fedilink
          English
          arrow-up
          7
          ·
          7 months ago

          I’d like to see this fix the most annoying part about subtitles, timing. find transcript/any subs on the Internet and have the AI align it with the audio properly.

          • Scrollone@feddit.it
            link
            fedilink
            English
            arrow-up
            2
            ·
            7 months ago

            YES! I can’t stand when subtitles are misaligned to the video. If this AI tool could help with that, it would be super useful.

  • m8052@lemmy.world
    link
    fedilink
    English
    arrow-up
    86
    arrow-down
    1
    ·
    7 months ago

    What’s important is that this is running on your machine locally, offline, without any cloud services. It runs directly inside the executable

    YES, thank you JB

  • renzev@lemmy.world
    link
    fedilink
    English
    arrow-up
    76
    arrow-down
    2
    ·
    7 months ago

    This sounds like a great thing for deaf people and just in general, but I don’t think AI will ever replace anime fansub makers who have no problem throwing a wall of text on screen for a split second just to explain an obscure untranslatable pun.

  • m-p{3}@lemmy.ca
    link
    fedilink
    English
    arrow-up
    33
    ·
    edit-2
    7 months ago

    Now I want some AR glasses that display subtitles above someone’s head when they talk à la Cyberpunk that also auto-translates. Of course, it has to be done entirely locally.

  • Phoenixz@lemmy.ca
    link
    fedilink
    English
    arrow-up
    29
    arrow-down
    2
    ·
    edit-2
    7 months ago

    As vlc is open source, can we expect this technology to also be available for, say, jellyfin, so that I can for once and for all have subtitles.done right?

    Edit: I think it’s great that vlc has this, but this sounds like something many other apps could benefit from

  • billwashere@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    ·
    7 months ago

    This might be one of the few times I’ve seen AI being useful and not just slapped on something for marketing purposes.

  • Doorbook@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    ·
    7 months ago

    The nice thing is, now at least this can be used with live tv from other countries and languages.

    Think you want to watch Japanese tv or Korean channels with out bothering about downloading, searching and syncing subtitles

  • Nalivai@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    1
    ·
    7 months ago

    The technology is nowhere near being good though. On synthetic tests, on the data it was trained and tweeked on, maybe, I don’t know.
    I corun an event when we invite speakers from all over the world, and we tried every way to generate subtitles, all of them run on the level of YouTube autogenerated ones. It’s better than nothing, but you can’t rely on it really.

    • Scrollone@feddit.it
      link
      fedilink
      English
      arrow-up
      5
      ·
      7 months ago

      No, but I think it would be super helpful to synchronize subtitles that are not aligned to the video.

      • Telodzrum@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        7 months ago

        This is already trivial. Bazarr has been doing it for all my subtitles for almost a decade.

  • TheRealKuni@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    3
    ·
    7 months ago

    And yet they turned down having thumbnails for seeking because it would be too resource intensive. 😐

    • DreamlandLividity@lemmy.world
      link
      fedilink
      English
      arrow-up
      11
      ·
      7 months ago

      I mean, it would. For example Jellyfin implements it, but it does so by extracting the pictures ahead of time and saving them. It takes days to do this for my library.

        • DreamlandLividity@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          ·
          7 months ago

          I get what you are saying, but I don’t think there is any standardized format for these trickplay images. The same images from Plex would likely not be usable in Jellyfin without converting the metadata (e.g. to which time in the video an image belongs to). So VLC probably does not have a good way to understand trickplay images not made by VLC.

    • cley_faye@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      ·
      7 months ago

      Video decoding is resource intensive. We’re used to it, we have hardware acceleration for some of it, but spewing something around 52 million pixels every second from a highly compressed data source is not cheap. I’m not sure how both compare, but small LLM models are not that costly to run if you don’t factor their creation in.

      • TheRealKuni@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        7 months ago

        All they’d need to do is generate thumbnails for every period on video load. Make that period adjustable. Might take a few extra seconds to load a video. Make it off by default if they’re worried about the performance hit.

        There are other desktop video players that make this work.