• Tibert@compuverse.uk
      link
      fedilink
      English
      arrow-up
      13
      ·
      1 year ago

      Well from the article a dataset is required, but not always the heavier one.

      Tho it doesn’t solve the speed issue, where the llm will take a lot more time to do the compression.

      gzip can compress 1GB of text in less than a minute on a CPU, an LLM with 3.2 million parameters requires an hour to compress

      • rubikcuber@programming.dev
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        I imagine that the compression is linked to the dataset, so if you update or retrain then you maybe lose access to the compressed data.