• AbouBenAdhem@lemmy.world
    link
    fedilink
    English
    arrow-up
    33
    arrow-down
    1
    ·
    edit-2
    15 days ago

    The basic idea behind the researchers’ data compression algorithm is that if an LLM knows what a user will be writing, it does not need to transmit any data, but can simply generate what the user wants them to transmit on the other end

    Great… but if that’s the case, maybe the user should reconsider the usefulness of transmitting that data in the first place.

  • andallthat@lemmy.world
    link
    fedilink
    English
    arrow-up
    16
    ·
    edit-2
    18 days ago

    I tried reading the paper. There is a free preprint version on arxiv. This page (from the article linked by OP) also links the code they used and the data they tried compressing, in the end.

    While most of the theory is above my head, the basic intuition is that compression improves if you have some level of “understanding” or higher-level context of the data you are compressing. And LLMs are generally better at doing that than numeric algorithms.

    As an example if you recognize a sequence of letters as the first chapter of the book Moby-Dick you’ll probably transmit that information more efficiently than a compression algorithm. “The first chapter of Moby-Dick”; there … I just did it.

  • besselj@lemmy.ca
    link
    fedilink
    English
    arrow-up
    9
    ·
    edit-2
    18 days ago

    So if I have two machines running the same local LLM and I pass a prompt between them, I’ve achieved data compression by transmitting the prompt rather than the LLM’s expected response to the prompt? That’s what I’m understanding from the article.

    Neat idea, but what if you want to transmit some information that an LLM can’t tokenize and generate accurately?

    • taladar@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      6
      ·
      18 days ago

      And how do I get the prompt that will reliably generate the data from the data? Usually for compression we do not start from an already compressed version.

  • Alphane Moon@lemmy.world
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    18 days ago

    I found the article to be rather confusing.

    One thing to point out is that the video codec used in this research (but for which results weren’t published for some reason), H264, is not at all state of the art.

    H265 is far newer and they are already working on H266. There are also other much higher quality codecs such as AV1. For what it’s worth, they do reference H265, but I don’t have access to the source research paper, so it’s difficult to say what they are comparing against.

    The performance relative to FLAC is interesting though.