Study Finds AI Language Models Surpass PNG and FLAC in Lossless Compression

A research paper by DeepMind reveals that their large language model (LLM) called Chinchilla 70B can achieve better lossless compression than traditional algorithms designed for image and audio compression. The LLM compressed image patches from the ImageNet database to 43.4% of their original size, surpassing the PNG algorithm at 58.5%. For audio, Chinchilla compressed samples from the LibriSpeech dataset to 16.4% of their raw size, outperforming FLAC compression at 30.3%. The study suggests that language models like Chinchilla can be effective tools for compressing various types of data, opening up new possibilities for machine learning models beyond text prediction and writing. The relationship between compression and intelligence is an ongoing area of research and debate.
Reading Insights
0
1
3 min
vs 4 min read
86%
798 → 115 words
Want the full story? Read the original article
Read on Ars Technica