Sorry, you need to enable JavaScript to visit this website.

We present a tandem scheme for Gaussian source compression, where a dead-zone quantizer is concatenated with a ternary low density generator matrix (LDGM) code. Both theoretical analysis and simulation results show that the LDGM codes can be universally optimal for near-lossless compression of ternary sources. Consequently, the distortion with the tandem scheme is mainly caused by the quantization, which can be negligible for high-rate quantizer. The most distinguished feature of the proposed scheme is its flexibility.


In the paper Huffman codes that mix different r-nary code elements in one code, the mixed Huffman codes, are analyzed. The Huffman code generalization usually leads to shortening of average codeword length: a statistical test shows that for source alphabets longer than 8-12 elements more than 99% of the best compact codes are mixed Huffman ones. This is also true for practical mixed Huffman codes, which is demonstrated in experiments with data files containing up to milion elements for sources of size 12-17 symbols.


Reconfigurable intelligent surfaces (RISs), which enable tunable anomalous reflection, have appeared as a promising method to enhance wireless systems. In this paper, we propose to use an RIS as a spatial equalizer to address the well-known multi-path fading phenomenon. By introducing some controllable paths artificially against the multi-path fading through the RIS, we can perform equalization during the transmission process instead of at the receiver, and thus all the users can share the same equalizer.


Wireless edge caching is an important strategy to fulfill the demands in the next generation wireless systems. Recent studies have indicated that among a network of small base stations (SBSs), joint content placement improves the cache hit performance via reinforcement learning, since content requests are correlated across SBSs and files. In this paper, we investigate multi-agent reinforcement learning (MARL), and identify four scenarios for cooperation.


PDF icon PPT (96)

Exchanging large amounts of floating-point data is common in distributed scientific computing applications. Data compression, when fast enough, can speed up such workloads by reducing the time spent waiting for data transfers. We propose ndzip, a high-throughput, lossless compression algorithm for multi-dimensional univariate regular grids of single- and double-precision floating point data.


In this paper, we present a new coding approach to near-lossless compression for binary sparse sources by using a special class of low density generator matrix (LDGM) codes. On the theoretical side, we proved that such a class of block LDGM codes are universal in the sense that any source with an entropy less than the coding rate can be compressed and reconstructed with an arbitrarily low bit-error rate (BER). On the practical side, we employ spatially coupled LDGM codes to reduce the complexity of reconstruction by implementing an iterative sliding window decoding algorithm.