My research primarily focuses on MultiModal Learning and Trustworthy & Interpretability AI, with a emphasis on the following areas:
Vision + Language:
Bridging the gap between our knowledge of language and vision representation, providing a foundation for more interpretable and controllable multimodal systems.
Unify Multimodal Understanding and Generation:
Bridging architectural differences between autoregressive and diffusion models via mutual promotion of understanding and generation.
Mechanistic Interpretability:
Understanding the internal mechanisms of LLMs and MLLMs through techniques like sparse autoencoder, circuit analysis, causal tracing, and neuron analysis.
Hallucination, Factuality and Safety:
Using the interpretability findings to help downstream tasks (e.g. factual knowledge, enhancing reasoning, reducing hallucinations, model editing), and design safer models.
Representative papers are highlighted, * = Equal Contribution