Projects

A summary of some recent projects.

Information Maximizing Exploration with a Latent Dynamics Model

This work extended a previous exploration method, Variational Information Maximizing Exploration, that is based on the idea of optimism in the face of uncertainty. Essentially, we incentivize exploring parts of the environment about which little is known.

In VIME, a separate Bayesian dynamics model is learned and intrinsic reward bonuses are derived from the amount of information gained by visiting a state. Our key observation is based on a relationship between linear one-step forward models and linear value estimates. We propose fitting the dynamics model in the features found by the final hidden layer of the neural network value estimator. This has two significant advantages: (1) the dimensionality of the inputs are untied from the dimensionality of the Bayesian model and (2) the theory suggests that since the value estimate is linear in these features, the model can be as well. This simplifies the Bayesian modeling. We find this to be the case and observe that information maximizing exploration in the latent space performs comparable or better than VIME.

This work was presented at Deep Reinforcement Learning Symposium at the 2017 Conference on Neural Information Processing Systems.

Mutual Learning and Adaptation for Robot to Human Handover Tasks

We focus on sample-efficient reinforcement learning in the context of a human with a disability. An initial policy is defined by imitation, which is then adapted to someone which a particular impairment according to some objective (distance to the human's hand, for example). We are able to learn efficient trajectories for object handovers from robot to human in as few as 50 trials.

This work was presented at the 2017 Interdisciplinary Conference on Reinforcement Learning and Decision Making.

Reinforcement learning for tasks in a 3D environment using only visual input

Deep reinforcement learning is an effective method for training autonomous agents to a high level of performance on visual tasks. This work explores how reinforcement learning agents using deep Q-networks perform when visually processing 3-D virtual environments and how deeper network architectures can improve performance given the added difficulties of more complex environments.

We explored ways to handle 3D environments such as providing a point-cloud representation of the agent's environment or a depth map as input to the network. We found, however, that depth can accurately be inferred from 2D inputs in structured environments. We provide results for tests on a variety of tasks in a virtual 3-D world and show that deeper convolutional neural networks lead to increased performance.

This work was presented at Deep Reinforcement Learning Workshop at the 2016 International Joint Conference in Artifical Intelligence.

Visualizing Natural Language Features in Stacked Autoencoders

Visualizing the features of unsupervised deep networks is an important part of understanding what a network has learned. In this work, we present a method for visualizing a deep autoencoder's hidden layers when trained on natural language data with t-SNE plots and wordclouds. Our method is complementary to training error analysis and it can help determine an appropriate stopping point in the training process. It can also provide researchers insight into the semantic language features the network has extracted from the dataset. Finally, it can show a big picture view of what a network has learned and how the various features the network has extracted relate to one another in semantic hierarchies. We hope that these benefits will aid human understanding of deep networks and can help guide future experiments.

This work was presented at the 2016 European Symposium on Artificial Neural Networks.

Recent News

  • : I am an RA funded by Intel tasked with creating educational material in machine learning and reinforcement learning.
  • : I presented my work on Information Maximizing Exploration with a Latent Dynamics Model at the NIPS 2017 Deep RL Symposium.
  • : I am now a research assistant funded by Honda Research Institute studying the use of memory-augmented policies for long-term dependencies in reinforcement learning.
  • : I was a teaching assistant for the fall '17 iteration of CSE 571 Artificial Intelligence at ASU.
  • : I reviewed for Humanoids 2017.
  • : I presented my work on Mutual Learning and Adaptation for Robot to Human Handover Tasks at the 2017 Interdisciplinary Conference on Reinforcement Learning and Decision Making.
  • : I spent the summer as an intern at Honda Research in Mountain View, CA.

Profile - CV Link

Industry Experience

  • 5/2017–8/2017: Summer intern at Honda Research Institute. Mountain View, CA. Focused on RL applications to autonomous driving.
  • 6/2014–1/2015: Software Engineer Associate at Lockheed Martin. Colorado Springs, CO. Software engineer on the DIAMONDShield project.
  • 5/2012–8/2012: Google Summer of Code Intern with Benetech. Implemented text-to-speech for mathematical expressions.

Academic Experience

  • 8/2017–Present: Research Assistant and Teaching Assistant under Dr. Ben Amor; Interactive Robotics Lab, ASU. Tempe, AZ. Working on project to improve RL algorithms with memory-augmented policies for partially observed domains with long temporal dependencies. I was also a TA for the graduate level AI course in Fall 2017.
  • 8/2016–5/2017: Research Assistant under Dr. Ben Amor; Interactive Robotics Lab, ASU. Tempe, AZ. Worked on a Toyota-funded project to optimize human-robot handovers with applications to both manufacturing and assisting those with physical impairments.
  • 5/2015–5/2016: Computer Science Research Assistant and Paraprofessional, Colorado College. Colorado Springs, CO. Tutor for computer science courses and research assistant to my undergraduate adviser.

Education

  • 2016-present: M.S. in Computer Science. Arizona State University, Tempe, AZ.
  • 2011–2014: B.A. in Computer Science. Colorado College, Colorado Springs, CO. 3.89 GPA; 3.975 Major GPA. Distinction in Computer Science.

Awards

  • April 2013: Won the Colorado College Big Idea Entrepreneurship Competition. Mobile EEG brain monitoring application for epilepsy patients. Earned product development grant of $38,000. Assisted my partner's further development of the product.
  • August 2012: London Olympian in the 20km race walk. Finished 26th with the fastest time ever by an American in the event; U.S. national champion in the 20km race walk, 2011 and 2012; youngest U.S. competitor at the 2011 World Track and Field Championships in Daegu, Korea, finishing 23rd of 46 entrants.
  • Spring 2012: Euclid Scholarship Recipient. Awarded by the Colorado College Math and Computer Science department in recognition of exceptional promise in math and computer science.
  • Fall 2012–Spring 2014: Colorado College President's Council. Served as an ambassador for Colorado College both internally and externally.
  • Spring 2010: National Engineering Design Challenge finalist. Member of Stanford Online High School team that qualified for national finals in Washington, D.C. Developed a device that translated the lights on the phone indicating which line is in use or ringing to a tactile format that users with visual impairment could understand.
  • Summer 2008: University of Pittsburgh Gene Team Member. Selected through competitive application process to participate in National Science Foundation-funded biomedical summer research program.

Publications

. , .