An Information-Theoretic Approach to Optimal Neural-Network-Based Compression
dc.contributor.author | Bhadane, Sourbh | |
dc.contributor.chair | Acharya, Jayadev | en_US |
dc.contributor.committeeMember | Wagner, Aaron | en_US |
dc.contributor.committeeMember | Weinberger, Kilian | en_US |
dc.contributor.committeeMember | Goldfeld, Ziv | en_US |
dc.date.accessioned | 2024-04-05T18:46:10Z | |
dc.date.issued | 2023-08 | |
dc.description | 125 pages | en_US |
dc.description.abstract | Modern artificial-neural-network-based (ANN-based) compressors have recently achieved notable successes on compressing multimedia formats such as images. This is despite information-theoretic near-optimality results of the linear transform coding paradigm, which forms the basis of existing standard lossy compressors such as JPEG, AAC etc., for stationary Gaussian sources with respect to mean-squared error distortion (at high rate). This thesis attempts to fill in some of the gaps in our theoretical understanding of modern ANN-based compressors. We list our contributions below. We propose a set of sources that obey the manifold hypothesis, i.e., that are high-dimensional in input space but lie on a low-dimensional manifold. We analytically derive optimal entropy-distortion tradeoffs for such sources and test the performance of ANN-based compressors on them. We find that for some sources that exhibit circular symmetry, ANN-based compressors are suboptimal. Our fix to this issue involves embedding Random Fourier Features (RFFs) before passing the input through either encoding or decoding nonlinear transforms. As the set of manifold sources gets more sophisticated, exact characterization of entropy-distortion tradeoffs can get challenging. We focus on the low-rate regime and develop general methods for one-bit quantization of sources in an arbitrary Hilbert space. Using these methods, we derive optimal one-bit quantizers for several examples including elliptical distributions and a manifold source that we proposed. We also study the low-rate asymptotics for variable-rate dithered quantization for vector Gaussian sources. We revisit the ubiquitous autoencoder architecture and analyze dimensionality-reducing linear autoencoders that are often used for general-purpose lossy compression. We propose an alternate autoencoder formulation that embraces the compression point of view by constraining the number of bits required to represent the output of the encoder. Our characterization of the optimal solution to this non-convex constrained linear autoencoder involves generalizing to any Schur-concave constraint on the variances of the encoder output. We provide experimental validation of our autoencoder-based variable-rate compressor. | en_US |
dc.identifier.doi | https://doi.org/10.7298/0kg8-8787 | |
dc.identifier.other | Bhadane_cornellgrad_0058F_13983 | |
dc.identifier.other | http://dissertations.umi.com/cornellgrad:13983 | |
dc.identifier.uri | https://hdl.handle.net/1813/114578 | |
dc.language.iso | en | |
dc.rights | Attribution 4.0 International | * |
dc.rights.uri | https://creativecommons.org/licenses/by/4.0/ | * |
dc.subject | autoencoders | en_US |
dc.subject | data compression | en_US |
dc.subject | information theory | en_US |
dc.subject | machine learning | en_US |
dc.subject | neural networks | en_US |
dc.subject | neural-network-compression | en_US |
dc.title | An Information-Theoretic Approach to Optimal Neural-Network-Based Compression | en_US |
dc.type | dissertation or thesis | en_US |
dcterms.license | https://hdl.handle.net/1813/59810.2 | |
thesis.degree.discipline | Electrical and Computer Engineering | |
thesis.degree.grantor | Cornell University | |
thesis.degree.level | Doctor of Philosophy | |
thesis.degree.name | Ph. D., Electrical and Computer Engineering |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Bhadane_cornellgrad_0058F_13983.pdf
- Size:
- 1.36 MB
- Format:
- Adobe Portable Document Format