Recent/Active Projects
- Time Series Benchmarking
- ARC: Abstraction and Reasoning Challenge
- Catan: Deep Reinforcement Learning
This project is about developing some standard benchmarking for temporal and spatiotemporal datasets which are increasingly riding the wave of deep learning performance boosts. There is a vast number of important problems which can be cast as a temporally structured problem, and being able to benchmark the growing number of specially designed architectures of increasing importance.
This project works on the ARC Challenge created by Francois Chollet designed to push the boundaries of current methods for artificial intelligence towards the generalizability and reasoning of human level intelligence, highlighting the shortcomings of current deep learning approaches. I am exploring the limitations of data-hungry methods through a systematic set of experiments in this simplistic domain.
This project implements the board game CATAN which is a complex game involving planning, long-term decision-making, probabilistic value evaluation, and social trading amongst players. Next, this project will train reinforcement learning agents to play this complex board game. Currently, I have finished implementing the gameplay and am exploring different deep RL methods to solve this difficult problem.
Previous Projects
- League of Legends: Skill Prediction
- CNN Evalutation of Vector Field Visualizations
- Dataset Distillation for Privacy using GANs
- Classifying American Sign Language Letters from Images
- Exploring the Double Descent Curve in Deep Neural Networks
- The Effects of Size and Compression of Neural Networks
- Identifying Song Sentiment from Song Lyrics
This project attempts to evaluate the skill of the players in a given League of Legends video game. The network is trained on previous games from all players and is able to predict the winning team of future games with 60% accuracy while only knowing the players and champions at the beginning of the match.
The crux of data visualization research is the "user-study" to understand how effective a particular DV method is at conveying the information/ data it is trying to show. This work combines scientific visualization with the work of Haehn in the context of vector field visualization to see how CNN-vision compares to human-vision. In particular, we try to identify the 'type' of critical point in a 2D vector field visualization. This work has the potential to influence computer vision by increasing our understanding of CNN models as well as the potential to influence data visualization by creating a much quicker version of the "user-study" by replacing human surveyees with CNN models.
The medical domain is full of sensitive user data which companies are consequently reluctant to share. This project is about training a GAN to accurately represent a given distribution while maintaining the privacy of the data points used to train the model (so the GAN can thereafter be shared with the public and can generate new private samples for anyone to train their model on.) There is a language of differential privacy already in the literature to address the privacy concerns for each data sample (user) and in this project we evaluate both the sample quality and the sample privacy as a proof of concept for this line of work.