Sung, Jaeyong2017-07-072017-12-082017-05-30Sung_cornellgrad_0058F_10207http://dissertations.umi.com/cornellgrad:10207bibid: 9948842https://hdl.handle.net/1813/51619The ability to reason about different modalities of information, for the purpose of physical interaction with objects, is a critical skill for assistive robots. For a robot to be able to assist us in our daily lives, it is not feasible to train each robot for a large number of tasks with all instances of objects that exist in human environments. Robots will have to generalize their skills by jointly reasoning with various sensor modalities such as vision, language and haptic feedback. This is an extremely challenging problem because each modality has intrinsically different statistical properties. Moreover, even with expert knowledge, manually designing joint features between such disparate modalities is difficult. In this dissertation, we focus on developing learning algorithms for robots that model tasks involving interactions with various objects in unstructured human environments --- especially on novel objects and scenarios that involve sequences of complicated manipulation. To this end, we develop algorithms that learn shared representations of multimodal data and model full sequences of complex motions. We demonstrate our approach on several different applications: understanding human activities in unstructured environment, synthesizing manipulation sequences for under-specified tasks, manipulating novel appliances, and manipulating objects with haptic feedback.en-USmachine learningMultimodal DataRobotic ManipulationRobot LearningArtificial intelligenceDeep LearningComputer scienceRoboticsLearning to Manipulate Novel Objects for Assistive Robotsdissertation or thesishttps://doi.org/10.7298/X43R0R0W