I am an assistant professor in Computer Science (Courant Institute) and Data Science at New York University. I was an assistant professor in the Computer Science department at the University of Texas at Austin from 2020. Before UT, I was a researcher at Google AI in NYC and a Ph.D. student at UW, advised by Luke Zettlemoyer and Yejin Choi.
I enjoy studying real world language usages with simple, efficient and generalizable models. I also build benchmarks that allows us to evaluate NLP models and conduct model analysis. Here are research topics that I am currently interested in:
- Continual Learning and Knowledge Editing: While LMs retain vast amounts of world knowledge seen during pretraining, such knowledge can get outdated. I am interested in retrieval augmentation and updating parametric knowledge in LMs.
- LMs and Retrieval: I design how LMs interact with retrieval models, as well as study the retrievers themselves. I build training and inference algorithms, as well as evaluation framework for systems answering complex questions in realistic scenarios (e.g., retrievers outputs conflicting documents).
- Human-LM Interaction: NLP systems are getting deployed fast and widely. I am interested in improving human interactions with LM, for example, how should we present information such that users will not be misled by plausible yet imperfect model predictions? The deployment of models also creates opportunities to learn from interaction with users. We study how to leverage implicit user feedback to improve the models. We also investigate how to engage models to be active collaborator rather than passive worker. For instance, we train models to ask clarification questions and follow-up questions.
- Spoken Language Processing: Spoken language exhibits richer prosodic features that are absent in written text. Can we build textless NLP system which can work on speech signals, opening doors to handle languages without written scripts?