This paper proposes a method that enhances the compression performance of the current model under development for the upcoming MPEG standard on Feature Compression for Machines (FCM).
This standard aims at providing inter-operable compressed bitstreams of features in the context of split computing, i.e., when the inference of a large computer vision Neural-Network-based model is split between two devices.
Intermediate features correspond to tensors that can be reduced and entropy coded to limit the required bandwidth of such transmission.
In the envisioned design for the MPEG-FCM standard, feature tensors may be reduced using NN layers before being converted into 2D video frame to be compressed using existing video compression standards.
This paper introduces an additional channel truncation and packing method which enables the system to preserve only the relevant channels for a given content, while preserving the computer vision task performance at the receiver.
Implemented within the MPEG-FCM test model, the proposed method yields compression gains of 10.59\% on average in terms of bitrate vs. accuracy on multiple computer vision tasks and datasets.
Feature Compression for Machines with Range-Based Channel Truncation and Frame Packing
Related INSIGHTS
Explore the latest research and innovations in wireless, video, and AI technologies.

BLOG POST
InterDigital Honored with Produit en Bretagne Award: A Tes...

BLOG POST
Touching the Future: Revealing the Magic of Haptics
Touch is the next digital frontier. Explore the future of haptics from today’s XR to new opportunities for haptographers...

BLOG POST
Bridging Consistency and Creativity: How InterDigital’s HDR...

BLOG POST
InterDigital’s Insights for 2026

BLOG POST
2025: InterDigital’s Year In Review

BLOG POST