eCommons will be completely unavailable from 8:00am April 4 until 5:00pm April 5, 2018, for software upgrades. Thank you for your patience during this planned service interruption. Please contact us at firstname.lastname@example.org if you have questions or concerns.
Representation Learning For Sequence And Comparison Data
The core idea of representation learning is to learn semantically more meaningful features (usually represented by a vector or vectors for each data point) from the dataset, so that they contain more discriminative information and make the given prediction task easier. It often provides better generalization performance and data visualization. In this thesis work, we improve the foundation and practice of representation learning methods for two types of data, namely sequences and comparisons: 1. Using music playlist data as an example, we propose Logistic Markov Embedding method that learns from sequence of songs and yields vectorized representations of songs. We demonstrate its better generalization performance in predicting the next song to play in a coherent playlist, as well as its capability in producing meaningful visualization for songs. We also propose an accompanying scalable training method that can be easily parallelized for learning representations on sequences. 2. Motivated by modeling intransitivity (rock-paper-scissors relation) in competitive matchup (two-player games or sports) data, we propose the blade-chest model for learning vectorized representations of players. It is then extended to a general framework that predicts the outcome of pairwise comparisons, making use of both object and context features. We see its successful application in matchup and preference prediction. The two lines of works have the same underlying theme: the object we study is first represented by a parameter vector or vectors, which are used to explain the interac- tions in the proposed models. These parameter vectors are learned by training on the datasets that contain interactions. The learned vectors can be used to predict any future interaction by simply plugging them back into the proposed models. Also, when the dimensionality of the vector is small (e.g. 2), plotting them gives interesting insight into the data.
Representation learning; Sequence; Pairwise comparison
Van Loan,Charles Francis; Bindel,David S.
Ph.D. of Computer Science
Doctor of Philosophy
dissertation or thesis