eCommons

 

Learning from Less: Improving and Understanding Model Selection in Penalized Machine Learning Problems

Other Titles

Abstract

Model selection is the task of selecting a "good" model from a set of candidate models given data. In machine learning, it is important that models fit the training data well, however it is more important for a model to generalize to unseen data. Additionally, it is desirable for a model to be as small as possible allowing for deployment in low-resource settings. In this thesis, we focus on three problems where the structure of the problem benefits from reducing the size of the models. In the first problem setting, we explore word embedding models and propose a framework in which we connect popular word embedding methods to low-rank matrix models and suggest new models for computing word vectors. In the second problem setting, we explore sparsity-inducing penalties for deep neural networks in order to obtain highly sparse networks which perform competitively with their over-parametrized counterparts. Finally, we explore the task of robot planar pushing and propose a novel penalty which adapts the parameters of a neural network to unseen examples allowing for robots to better interact in unseen environments. Our results on simulations and real-world data applications indicate that penalization is effective for learning models which perform well in settings with less computational budget, storage, or labeled data.

Journal / Series

Volume & Issue

Description

142 pages

Sponsorship

Date Issued

2020-08

Publisher

Keywords

Deep Learning; Machine Learning; Matrix Factorization; Model Selection; Penalization; Regularization

Location

Effective Date

Expiration Date

Sector

Employer

Union

Union Local

NAICS

Number of Workers

Committee Chair

Wells, Martin Timothy

Committee Co-Chair

Committee Member

Wilson, Andrew Gordon
Joachims, Thorsten

Degree Discipline

Statistics

Degree Name

Ph. D., Statistics

Degree Level

Doctor of Philosophy

Related Version

Related DOI

Related To

Related Part

Based on Related Item

Has Other Format(s)

Part of Related Item

Related To

Related Publication(s)

Link(s) to Related Publication(s)

References

Link(s) to Reference(s)

Previously Published As

Government Document

ISBN

ISMN

ISSN

Other Identifiers

Rights

Attribution 4.0 International

Types

dissertation or thesis

Accessibility Feature

Accessibility Hazard

Accessibility Summary

Link(s) to Catalog Record