Deep Learning for Channel Coding via Neural Mutual Information Estimation
Research output: Contribution to book/Conference proceedings/Anthology/Report › Conference contribution › Contributed › peer-review
Contributors
Abstract
End-to-end deep learning for communication systems, i.e., systems whose encoder and decoder are learned, has attracted significant interest recently, due to its performance which comes close to well-developed classical encoder-decoder designs. However, one of the drawbacks of current learning approaches is that a differentiable channel model is needed for the training of the underlying neural networks. In real-world scenarios, such a channel model is hardly available and often the channel density is not even known at all. Some works, therefore, focus on a generative approach, i.e., generating the channel from samples, or rely on reinforcement learning to circumvent this problem. We present a novel approach which utilizes a recently proposed neural estimator of mutual information. We use this estimator to optimize the encoder for a maximized mutual information, only relying on channel samples. Moreover, we show that our approach achieves the same performance as state-of-the-art end-to-end learning with perfect channel model knowledge.
Details
Original language | English |
---|---|
Title of host publication | 2019 IEEE 20th International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2019 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (electronic) | 978-1-5386-6528-2 |
Publication status | Published - Jul 2019 |
Peer-reviewed | Yes |
Externally published | Yes |
Publication series
Series | IEEE Workshop on Signal Processing Advances in Wireless Communications (SPAWC) |
---|---|
ISSN | 1948-3244 |
Conference
Title | 20th IEEE International Workshop on Signal Processing Advances in Wireless Communications, SPAWC 2019 |
---|---|
Duration | 2 - 5 July 2019 |
City | Cannes |
Country | France |
External IDs
ORCID | /0000-0002-1702-9075/work/165878324 |
---|