Neural Joint Source-Channel Coding Kristy Choi , Kedar Tatwawadi, - - PowerPoint PPT Presentation

neural joint source channel coding
SMART_READER_LITE
LIVE PREVIEW

Neural Joint Source-Channel Coding Kristy Choi , Kedar Tatwawadi, - - PowerPoint PPT Presentation

Neural Joint Source-Channel Coding Kristy Choi , Kedar Tatwawadi, Aditya Grover, Tsachy Weissman, Stefano Ermon Computer Science Department, Stanford University Motivation Reliable, robust, and efficient information transmission is key for


slide-1
SLIDE 1

Neural Joint Source-Channel Coding

Kristy Choi, Kedar Tatwawadi, Aditya Grover, Tsachy Weissman, Stefano Ermon Computer Science Department, Stanford University

slide-2
SLIDE 2

Motivation

Reliable, robust, and efficient information transmission is key for everyday communication

slide-3
SLIDE 3

Problem Statement

1 1 1 1 1 Compression (Source coding) Decompression Reliable communication across noisy channel 1 1 Channel coding Channel decoding Separation Theorem [Shannon 1948] Assumes infinite blocklength & compute channel model

slide-4
SLIDE 4

Neural Joint Source-Channel Coding

1 1 1 1 1 channel model Neural joint source-channel encoder Neural joint source-channel decoder Learn to jointly compress and channel code

slide-5
SLIDE 5

NECST Model

1 1 1 1 1 encoder decoder Maximize mutual information [MacKay 2003] channel model

slide-6
SLIDE 6

Coding Process

1 1 1 1 1

slide-7
SLIDE 7

Learning Objective

  • Mutual information maximization
  • Y should capture as much information about X as

possible, even after corruption!

  • Estimation is hard ☹ [Barber & Agakov 2004]
  • Variational lower bound is nicer:

Reconstruction loss! [Kingma & Welling 2014] [Vincent 2008]

slide-8
SLIDE 8

Optimization Procedure

  • Our latent variables y are discrete ☹
  • Use VIMCO: [Mnih and Rezende 2016]
  • Draw multiple (K) samples from inference

network, get tighter lower bound

Multiple reconstruction loss terms Multiple samples of y

slide-9
SLIDE 9

Fixed Rate: Comparison vs. Ideal Codes

We need a much smaller number of bits to get the same level of distortion, even vs. WebP [Google 2010] + ideal channel code

slide-10
SLIDE 10

Extremely Fast Decoding

Up to 2x orders of magnitude in speedup on GPU vs. LDPC decoder [Gallager 1963]

slide-11
SLIDE 11

Learning the Data Distribution

1 1 1

Theorem (informal): NECST learns an implicit model of

1 1

slide-12
SLIDE 12

Robust Representation Learning

1 1 1 1 1

1) Encoded redundancies: interpolation in latent space by bit-flip 2) Improved downstream classification: improves accuracy by as much as 29% across variety of classifiers when inputs are corrupted by noise!

1 bit flip 0 bit flips 45/100 bits

slide-13
SLIDE 13

Summary

  • End-to-end deep generative modeling framework

for the JSCC problem

  • Better bitlength efficiency than separation scheme
  • n CIFAR10, CelebA, SVHN
  • Another way to learn robust latent representations
  • Get an extremely fast decoder for free
slide-14
SLIDE 14

Thanks!

Contact: kechoi@stanford.edu Code: https://github.com/ermongroup/necst Poster #165: Tuesday, June 11th @ Pacific Ballroom

Kedar Tatwawadi Tsachy Weissman Stefano Ermon Aditya Grover