[June 23] One EUSIPCO accepted

Title:A Water-filling Algorithm Maximizing the Volume of Submatrices Above the Rank

Authors: C. Petit, A. Roumy, T. Maugey

Abstract: In this paper, we propose an algorithm to extract, from a given rectangular matrix, a submatrix with maximum volume, whose number of extracted columns is greater than the initial matrix rank. This problem arises in compression and summarization of databases, recommender systems, learning, numerical analysis or applied linear algebra. We use a continuous relaxation of the maximum volume matrix extraction problem, which admits a simple and closed form solution: the nonzero singular values of the extracted matrix must be equal. The proposed algorithm extracts matrices with singular values, which are close to be equal. It is inspired by a water-filling technique, traditionally dedicated to equalization strategies in communication channels. Simulations show that the proposed algorithm performs better than sampling methods based on determinantal point processes (DPPs) and achieves similar performance as the best known algorithm, but with a lower complexity.

[Feb 23] our ICASSP paper accepted

Title: Learning on entropy coded data using CNN

Authors: R. Piau, T. Maugey, A. Roumy

Abstract: We propose an empirical study to see whether learning with convolutional neural networks (CNNs) on entropy coded data is possible. First, we define spatial and semantic closeness, two key properties that we experimentally show to be necessary to guarantee the efficiency of the convolution. Then, we show that these properties are not satisfied by the data processed by an entropy coder. Despite this, our experimental results show that learning in such difficult conditions is still possible, and that the performance are far from a random guess. These results have been obtained thanks to the construction of CNN architectures designed for 1D data (one based on VGG, the other on ResNet). Finally we propose some experiments that explains why CNN are still performing reasonably well on entropy coded data.

[Jan 22] New journal article accepted in MDPI Sensors

Title: Machine learning for multimedia communications

Authors: Nikolaos Thomos, Thomas Maugey, Laura Toni

Abstract: Machine learning is revolutionizing the way multimedia information is processed and transmitted to users. After intensive and powerful training, some impressive efficiency/accuracy improvements have been reached all over the transmission pipeline. For example, the high model capacity of the learning-based architectures enables to accurately model the image and video behavior such that tremendous compression gains can be achieved. Similarly, error concealment, streaming strategy or even user’s perception modeling have widely benefited from the recent
learning-oriented developments. However, learning-based algorithms often imply drastic changes on the way data is represented or consumed, meaning that the overall pipeline can be affected even though a subpart of it is optimized. In this paper, we review the recent major advances that have been proposed all over the transmission chain, and we discuss their potential impact and the research challenges that they raise.

[Sep 20] New IEEE SP Letter accepted

Title: Large Database Compression Based on Perceived Information

Authors: Thomas Maugey and Laura Toni

Abstract: Lossy compression algorithms trade bits for quality, aiming at reducing as much as possible the bitrate needed to represent the original source (or set of sources), while preserving the source quality. In this letter, we propose a novel paradigm of compression algorithms, aimed at minimizing the information loss perceived by the final user instead of the actual source quality loss, under compression rate constraints.
As main contributions, we first introduce the concept of perceived information (PI), which reflects the information perceived by a given user experiencing a data collection, and which is evaluated as the volume spanned by the sources features in a personalized latent space.
We then formalize the rate-PI optimization problem and propose an algorithm to solve this compression problem. Finally, we validate our algorithm against benchmark solutions with simulation results, showing the gain in taking into account users’ preferences while also maximizing the perceived information in the feature domain.


[Aug 20] New journal article accepted in Annals of Telecommunications

Title: Excess rate for model selection in interactive compression using Belief-propagation decoding

Authors: Navid Mahmoudian-Bidgoli, Thomas Maugey, Aline Roumy

Abstract: Interactive compression refers to the problem of compressing data while sending only the part requested by the user. In this context, the challenge is to perform the extraction in the compressed domain directly. Theoretical results exist, but they assume that the true distribution is known. In practical scenarios instead, the distribution must be estimated. In this paper, we first formulate the model selection problem for interactive compression and show that it requires to estimate the excess rate incurred by mismatched decoding. Then, we propose a new expression to evaluate the excess rate of mismatched decoding in a practical case of interest: when the decoder is the belief-propagation algorithm. We also propose a novel experimental setup to validate this closed-form formula. We show a good match for practical interactive compression schemes based on fixed-length Low-Density Parity-Check (LDPC) codes. This new formula is of great importance to perform model and rate selection.

[Sep 19] New journal article accepted in Physical Communication

Title: Rate-Storage Regions for Extractable Source Coding with Side Information

Authors: E. Dupraz, T. Maugey, A. Roumy, M. Kieffer

Abstract:This papers considers the coding of a source with decoders, each having access to a different side information . We define a new source coding problem called Extractable Source Coding with Side Information (ESC-SI). In this problem, the server stores one single coded description of the source, from which descriptions can be extracted without re-encoding, depending on the side information available at the decoder. We want to minimize both the storage rate of the source on the server, and the transmission rates from the server to the decoders. We provide the achievable storage-transmission rate regions for lossless source coding of general, non i.i.d., non-ergodic sources, and the achievable storage-transmission rate–distortion regions for lossy source coding for non i.i.d. Gaussian sources. The regions obtained for such general source models provide insightful design guidelines for practical applications.

[Aug 19] three papers accepted at PCS 2019

The three following papers are accepted and will be presented at the Picture Coding Symposium (PCS) 2019, in China:

N. Mahmoudian Bidgoli, T. Maugey , A. Roumy, Intra-coding of 360-degree images on the sphere
Picture Coding Symposium (PCS), Ningbo, China, Nov. 2019

N. Mahmoudian Bidgoli, T. Maugey, A. Roumy, F. Nasiri and F. Payan, A geometry-aware compression of 3D mesh texture with random access
Picture Coding Symposium (PCS), Ningbo, China, Nov. 2019

P. Garus, J. Jung, T. Maugey and C. Guillemot, Bypassing Depth Maps Transmission For Immersive Video Coding
Picture Coding Symposium (PCS), Ningbo, China, Nov. 2019

[June 19] New journal article accepted in IEEE TIP

Title: Geometry-Aware Graph Transforms for Light Field Compact Representation

Authors: Mira Rizkallah, Xin Su, Thomas Maugey, Christine Guillemot

Abstract: The paper addresses the problem of energy compaction of dense 4D light fields by designing geometry-aware local graph-based transforms. Local graphs are constructed on super-rays that can be seen as a grouping of spatially and geometry-dependent angularly correlated pixels. Both non separable and separable transforms are considered. Despite the local support of limited size defined by the super-rays, the Laplacian matrix of the non separable graph remains of high dimension and its diagonalization to compute the transform eigen vectors remains computationally expensive. To solve this problem, we then perform the local spatio-angular transform in a separable manner. We show that when the shape of corresponding super-pixels in the different views is not isometric, the basis
functions of the spatial transforms are not coherent, resulting in decreased correlation between spatial transform coefficients. We hence propose a novel transform optimization method that aims at preserving angular correlation even when the shapes of the super-pixels are not isometric. Experimental results show the benefit of the approach in terms of energy compaction. A coding scheme is also described to assess the rate-distortion perfomances of the proposed transforms and is compared to state of the art
encoders namely HEVC-lozenge [1], JPEG pleno 1.1 [2], HEVC- pseudo [3] and HLRA [4] .

[May 2019] One paper accepted in ICIP

Title: Evaluation framework for 360-degree visual content compression with user-dependent transmission

Authors: Navid MAHMOUDIAN BIDGOLI, Thomas MAUGEY, Aline ROUMY

Abstract: Immersive visual experience can be obtained by allowing the user to navigate in a 360-degree visual content. These contents are stored in high resolution and need a lot of space on the server to store them. The transmission depends on the user’s request and only the spatial region which is requested by the user is transmitted to avoid wasting network bandwidth. Therefore, storage and transmission rates are both critical.
%The former is important to reduce the space for storage on the server and the latter reduces the bitrate for the available network bandwidth.
Splitting the rates into storage and transmission has not been formally considered in the literature for evaluating 360-degree content compression algorithms. In this paper, we propose a framework to evaluate the coding efficiency of 360-degree content while discriminating between storage and transmission rate and taking into account user dependency. This brings the flexibility to compare different coding methods based on the storage capacity on the server and network bandwidth of users.