Study Finds AI Language Models Surpass PNG and FLAC in Lossless Compression

1 min read
Source: Ars Technica
Study Finds AI Language Models Surpass PNG and FLAC in Lossless Compression
Photo: Ars Technica
TL;DR Summary

A research paper by DeepMind reveals that their large language model (LLM) called Chinchilla 70B can achieve better lossless compression than traditional algorithms designed for image and audio compression. The LLM compressed image patches from the ImageNet database to 43.4% of their original size, surpassing the PNG algorithm at 58.5%. For audio, Chinchilla compressed samples from the LibriSpeech dataset to 16.4% of their raw size, outperforming FLAC compression at 30.3%. The study suggests that language models like Chinchilla can be effective tools for compressing various types of data, opening up new possibilities for machine learning models beyond text prediction and writing. The relationship between compression and intelligence is an ongoing area of research and debate.

Share this article

Reading Insights

Total Reads

0

Unique Readers

1

Time Saved

3 min

vs 4 min read

Condensed

86%

798115 words

Want the full story? Read the original article

Read on Ars Technica