Hello! I am a PhD student at JHU CLSP, advised by Mark Dredze.
My interests are protean due to my restless curiosity. I am generally interested in understanding how language models function and exploring how we can change them. Terminologically, I am interested in studying interpretability (broadly), training dynamics, evaluation, and reasoning of language models.
Before my PhD studies, I was a resident at FAIR Labs, working with Adina Williams and Dieuwke Hupkes. I obtained my Masterโs and Bachelorโs degrees of computer science (B.S., M.S.) and mathematics (B.A.) with a mnior in classical studies at University of Washington. My advisor was Noah A. Smith, and I was supervised by Ana Marasoviฤ. I have also interned at the AWS AI Labs for two times, mentored by Peng Qi, Yuhao Zhang, Jifan Chen, and Danilo Ribeiro. During my undergraduate years, I also worked with Christopher Hoffman on dual random-walk systems. Thanks a lot to the support of my advisors and my donors, I was able to conduct research and keep learning.
I have a limited amount of life experiences, but if you are having questions and you think I can help, feel free to email me.
News
[August 2024] New preprint ๐: Amuro and Char: Analyzing the Relationship between Pre-Training and Fine-Tuning of Large Language Models. Our system achieve great performance ๐ in almost all languages in the IWSLT low-resource speech translation shared task.
[December 2023] Our paper, The Validity of Evaluation Results: Assessing Concurrence Across Compositionality Benchmarks, received an Honorable Mention ๐ in CoNLL2023.
[October 2023] New papers ๐: Tokenization Consistency Matters for Generative Models on Extractive Tasks is accepted into Findings of EMNLP2023; The Validity of Evaluation Results: Assessing Concurrence Across Compositionality Benchmarks is accepted into CoNLL2023; <- I will be presenting them in Singapore. A taxonomy and review of generalization research in NLP is accepted into Nature Machine Intelligence.
Older News
[July 2023] Our work A Replication Study of Compositional Generalization Works on Semantic Parsing was recognized as an ๐ Outstanding Paper ๐ in the ML Reproducibility challenge.[August 2022] I finished my master's degree, moved to New York City, and will spend a year here for my residency at Meta AI, working with Adina Williams and Dieuwke Hupkes.
[June 2022] I moved to Santa Clara for my internship at AWS AI, working with Peng Qi and Yuhao Zhang.
[December 2021] I began as a teaching assistant in Natural Language Processing at UW for winter and spring quarter.
[September 2021] I began as a teaching assistant in Machine Learning at UW for fall quarter.
[July 2021] I began my software engineering internship in AuthService team at Amazon.
[May 2021] Our paper Effective Attention Sheds Light On Interpretability was accepted into Findings of ACL2021. Big thanks to Ana! ๐ป
[March 2021] Finished my Bachelor's degrees ๐ค - B.S. in CS and B.A. in math, as well as a minor in classical studies; I began as a teaching assistant in Enterprise Chatbots at UW for spring quarter.
[November 2020] Joined Noonum as a data scientist intern.
[July 2020] I began my software engineering internship in AuthEngine team at Amazon.
[September 2019] Began a project on dual random-walk systems with Professor Christopher Hoffman at Washington Experimental Mathematics Lab.
[July 2019] Began my internship at National Oceanic & Atmospheric Administration (NOAA) mentored by Dr. Jason Cope.
Cat Warning
I have two very cute cats: Bert (white collar with white paws) and Roberta (grey and looks like a little leopard). While you are browsing my website, I hope pictures of Bert and Roberta can make you feel happy and relaxed for a second.