eCommons

 

An Information-Theoretic Approach to Optimal Neural-Network-Based Compression

dc.contributor.authorBhadane, Sourbh
dc.contributor.chairAcharya, Jayadeven_US
dc.contributor.committeeMemberWagner, Aaronen_US
dc.contributor.committeeMemberWeinberger, Kilianen_US
dc.contributor.committeeMemberGoldfeld, Ziven_US
dc.date.accessioned2024-04-05T18:46:10Z
dc.date.issued2023-08
dc.description125 pagesen_US
dc.description.abstractModern artificial-neural-network-based (ANN-based) compressors have recently achieved notable successes on compressing multimedia formats such as images. This is despite information-theoretic near-optimality results of the linear transform coding paradigm, which forms the basis of existing standard lossy compressors such as JPEG, AAC etc., for stationary Gaussian sources with respect to mean-squared error distortion (at high rate). This thesis attempts to fill in some of the gaps in our theoretical understanding of modern ANN-based compressors. We list our contributions below. We propose a set of sources that obey the manifold hypothesis, i.e., that are high-dimensional in input space but lie on a low-dimensional manifold. We analytically derive optimal entropy-distortion tradeoffs for such sources and test the performance of ANN-based compressors on them. We find that for some sources that exhibit circular symmetry, ANN-based compressors are suboptimal. Our fix to this issue involves embedding Random Fourier Features (RFFs) before passing the input through either encoding or decoding nonlinear transforms. As the set of manifold sources gets more sophisticated, exact characterization of entropy-distortion tradeoffs can get challenging. We focus on the low-rate regime and develop general methods for one-bit quantization of sources in an arbitrary Hilbert space. Using these methods, we derive optimal one-bit quantizers for several examples including elliptical distributions and a manifold source that we proposed. We also study the low-rate asymptotics for variable-rate dithered quantization for vector Gaussian sources. We revisit the ubiquitous autoencoder architecture and analyze dimensionality-reducing linear autoencoders that are often used for general-purpose lossy compression. We propose an alternate autoencoder formulation that embraces the compression point of view by constraining the number of bits required to represent the output of the encoder. Our characterization of the optimal solution to this non-convex constrained linear autoencoder involves generalizing to any Schur-concave constraint on the variances of the encoder output. We provide experimental validation of our autoencoder-based variable-rate compressor.en_US
dc.identifier.doihttps://doi.org/10.7298/0kg8-8787
dc.identifier.otherBhadane_cornellgrad_0058F_13983
dc.identifier.otherhttp://dissertations.umi.com/cornellgrad:13983
dc.identifier.urihttps://hdl.handle.net/1813/114578
dc.language.isoen
dc.rightsAttribution 4.0 International*
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/*
dc.subjectautoencodersen_US
dc.subjectdata compressionen_US
dc.subjectinformation theoryen_US
dc.subjectmachine learningen_US
dc.subjectneural networksen_US
dc.subjectneural-network-compressionen_US
dc.titleAn Information-Theoretic Approach to Optimal Neural-Network-Based Compressionen_US
dc.typedissertation or thesisen_US
dcterms.licensehttps://hdl.handle.net/1813/59810.2
thesis.degree.disciplineElectrical and Computer Engineering
thesis.degree.grantorCornell University
thesis.degree.levelDoctor of Philosophy
thesis.degree.namePh. D., Electrical and Computer Engineering

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Bhadane_cornellgrad_0058F_13983.pdf
Size:
1.36 MB
Format:
Adobe Portable Document Format