Data Analysis
My work has focused on analyzing structured and unstructured data to extract actionable results and business insights. I love to craft clever visualizations to illustrate results in a way that helps us see around obvious findings and discover unique ones. I've recently really enjoyed incorporating machine learning and AI into my work to see if we can improve on the high-quality discoveries we find. However, I do believe that these tools aren't necessary in every use case, so I'm a big proponent of using them strategically and not just for the sake of using them.
Programming & Software
I have used many tools to extract, clean, analyze, and visualize large datasets. My work during my PhD taught me how to know what questions to ask, and when to ask them. This has made learning new tools a fun challenge for me because I have the opportunity to learn as I go to solve challenging problems. A running list of the current tech stack that I work with includes: Python (numpy, pandas, scipy, matplotlib, seaborn, scikit-learn, umap, Jupyter notebooks), R, SQL, MATLAB, Power BI, Docker, Bash/Zsh, GitHub, and LaTeX.
Science Communication
My experiences delivering talks at international scientific conferences and pitching those same ideas to entrepreneurs and investors has helped me hone my skills in science communcation. I find it really fun to discuss complex science with non-experts, as it's super rewarding for me to be able to have conversations about life-changing scientific discoveries with people that rarely are given the opportunity to engage in those conversations. I have received great ideas through these discussions, and find them highly important for driving science forward. These experiences have helped me develop the skills to confidently present complex results to non-technical audiences.
Featured Projects - More Coming Soon!
SPOTs for low-cost, high-throughput biological and chemical reactions
The SPOTs platform is a new generation of tools for scientist and researchers to perform more experiments for less cost. This platform allows scientists run experiments and answer questions previously thought to be impossible to test. During my PhD, I used this tool to generate more data across life sciences and medicine that exceeded what was previously able to be collected over the past 100 years combined!
Watch us use SPOTs to make 1000 combinations of 4 liquids in minutes! Check out our paper on SPOTs!
Machine Learning + SPOTs = new drugs to treat bacterial infections
Once we were able to run 10,000+ experinments with SPOTs, we pulled all of the existing drug combination data from previous research studies. We trained a machine learning model on this data, and used it to predict what new drug combinations might be effective at killing bacteria. This project was a collaboration with a group at University of Michigan Medicine, where we doubled all of the data available from research studies within the past 30 years in a single day with SPOTs. We discovered 10 novel drug combinations that are being explored for use in the clinic to treat humans!
Decoding large datasets created through SPOTs
My most ambitious project stemmed from the idea of pushing the limits of SPOTs and questioning how we can discover new drug and chemical formulations. Here, we ran 150,000 experiments in a month, 10,000 times more than what would commonly be done in a chemistry lab. We discovered thousands of new reactions from a chemical process that's been known since 1890. This chemical process was previously thought to be exhausted of all potential reactions that could reasonably be produced, until we levaraged SPOTs to prove otherwise. The gif here represents just ~1/50th of the total data we collected, where we leverage changes in colors to represent complex chemical reactions. We then uncover the chemistry behind these color changes to discover new reactions. A thoroughly described data workflow and written manuscript is in the works to share here!