Joey Hejna

(Donald Joseph Hejna III)

I'm a second year PhD student in the computer science department at Stanford University advised by Dorsa Sadigh. My research is supported by an NDSEG Fellowship. I completed my undergrad at UC Berkeley where I worked with Professors Pieter Abbeel and Lerrel Pinto.

jhejna @ cs.stanford.edu  /  Resume  /  Github  /  Scholar  /  LinkedIn

profile photo
Research

I'm broadly interested in learning for decision making. Papers (and preprints) are ordered by recency.

Improving Long-Horizon Imitation through Instruction Prediction
Donald J. Hejna III, Pieter Abbeel, Lerrel Pinto
Preprint, under review
paper

We show that predicting instructions along with actions drastically improves performance in combinatorially complex long-horizon imitation settings.

Task-Agnostic Morphology Evolution
Donald J. Hejna III, Pieter Abbeel, Lerrel Pinto
Accepted to ICLR 2021
paper / website / code

Better robot strucutres hold the promise of better performance. We propose a new algorithm, TAME, that is able to evolve morphologies without any task specification. This is accomplished using an information theoretic objective that efficiently ranks morphologies based on their ability to explore and control their environment.

Hierarchically Decoupled Imitation for Morphological Transfer
Donald J. Hejna III, Pieter Abbeel, Lerrel Pinto
Accepted to ICML 2020
paper / website / code / talk

We propose transferring RL policies across agents using a hierarchical framework. Then, to remedy poor zero-shot transfer performance we introduce two additional imitation objectives.

Projects
Research Lightning
Donald J. Hejna III
Open source project
code

A lightweight framework for general deep-learning research in pytorch.

Improving Latent Representations via Explicit Disentanglement
Donald J. Hejna III*, Ashwin Vangipuram*, Kara Liu*
Course Project, CS 294-158 Unsupervised Learning, Spring 2020
paper

We examine and compare three methods for explicitly disentangling learned latent representations in VAE models.

Awards
  • National Defense Science and Engineering Graduate Scholarship (NDSEG) 2021, roughly 2% selection rate.
  • Honorable mention for the 2021 CRA Outstanding Undergraduate Researcher Award
  • Highest Degree Honors in Engineering at UC Berkeley Spring 2021, top 3% of the graduating class.
  • UC Berkeley Regents and Chancellors Scholarship
  • Rambus Innovator of the Future 2017
Industry
Intern, Citadel Global Quantitative Strategies
Summer 2019

Developed C++ systems for trading APIs and monitoring systems. Worked on optimizing memory usage of large model training.

Intern, Intel Artificial Intelligence Group
Summer 2018
blog post

Worked on demo systems for Intel's OpenVino model optimization system in the AWS DeepLens. Explored systems for gradient based explanations of deep networks.

Teaching
berkeley UC Berkeley EECS Department

Teaching Assistant, EECS 127: Optimization Models, Fall 2020

Teaching Assistant, EECS 189: Machine Learning, Spring 2020

Teaching Assistant, CS 70: Discrete Math and Probability Theory, Fall 2019
teaching resources Public Resources

Introductory ML Notes

Deep Learning Workshop

Reinforcement Learning Workshop

Website source taken from here.