This Science News Wire page contains a press release issued by an organization and is provided to you "as is" with little or no review from Science X staff.

Studying the theoretical limit of learned source coding

August 12th, 2023 Rachel Rose

Each day, an estimated 330,000 billion bytes of data are generated in various forms. This data is shared in many ways—videos, images, music, gaming, streaming content and video calls. This immense amount of data requires compression to save on storage capacity, speed up file transfer and decrease costs for storage hardware and network bandwidth.

Dr. Zixiang Xiong, professor and associate department head in the Department of Electrical and Computer Engineering at Texas A&M University, recently received a National Science Foundation grant to research the fundamental limits of learned source coding—or data compression that uses machine learning—now that new machine learning methods have permeated the scene.

The project is a culmination of over 30 years of research conducted by Xiong. Since the late 1980s, he has studied the area of data compression and has seen the evolution of the process.

In the 1990s, successfully sharing an image file required converting the file into text and back into an image. The flip side is now possible; machine learning generative models such as ChatGPT create new content and images based on input text into the model. With the unprecedented tools now available through artificial intelligence, Xiong will work to create new parameters for the evolving process.

"We always ask ourselves before we begin any engineering project, 'What's the theoretical limit?'" said Xiong. "That's very fundamental now because AI is completely different. There's no current theory because we don't know the theoretical limit."

This project aims to understand what types of machine learning algorithms can compress data well and how many samples are needed to learn compression well. While gaining a fundamental understanding of data compression that utilizes machine learning, Xiong hopes to develop more powerful compression methods, leading to more efficient use of wireless communication and less energy consumption by mobile devices.

Traditional compression methods include the well-known JPEG compression for smartphone images; this is a lossy compression method, which means that some image quality is lost. Lossless compression, meaning no quality is lost, is typically used for compressing computer files, such as with Zip, and for music streaming. This project aims to develop boundaries for the performance of machine learning for both compression methods.

In 2020, Xiong worked on a project titled "Deep Learning based Scalable Compression Framework" with Amazon Prime Video, which was preliminary work that led to this new project.

Collaborators for this project include Dr. Anders Høst-Madsen and Dr. Narayana Santhanam, both professors at the University of Hawai'i at Mānoa.

More information:
www.nsf.gov/awardsearch/showAward?AWD_ID=2324397

Provided by Texas A&M University College of Engineering

Citation: Studying the theoretical limit of learned source coding (2023, August 12) retrieved 29 November 2024 from https://sciencex.com/wire-news/453275153/studying-the-theoretical-limit-of-learned-source-coding.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.