Welcome to Yu Yu's homepage!

Google Scholar | CV | Contact

I'm a fourth year PhD student advised by Jia Xu at Stevens Institute of Technology.

My research interest is robustness and generalization of Natural language processing (NLP) systems such as Machine Translation and Dialogue Systems.

I served as the review committee member for EMNLP22, NIPS22, EMNLP 2023, ACL 2023, NIPS 2023, ICML 2023, ICML 2024.

Publications

We propose a neural language modeling system based on low-rank adaptation (LoRA) for speech recognition output rescoring. 

Conventional data selection methods select training samples based on the test domain knowledge and not on real life data, thus they frequently fail in unknown domains like patent and Twitter. We propose to select training samples that maximize information uncertainty measured by observation entropy like empirical Shannon entropy, and prediction entropy using mutual information, to cover more possible queries that may appear in unknown worlds.

This paper introduces our Diversity Advanced Actor-Critic reinforcement learning (A2C) framework (DAAC) to improve the generalization and accuracy of Natural Language Processing (NLP). We quantify diversity on a set of samples using the max dispersion, convex hull volume, and graph entropy based on sentence embeddings in high dimensional metric space. 

Unlike previous work that has defined robustness using Minimax to bound worst cases, we measure robustness based on the consistency of cross-domain accuracy and introduce the coefficient of variation and (ε, γ)-Robustness.

Internship

Google DeepMind 

Research Intern, Fall 2023

Topics: Large Language Model, Low-resource code generation, In-context Transfer Learning

Host: Sheena Panthaplackel, Christian Walder

Amazon

Applied Scientist Intern, Summer 2023

Topics: Speech Recognition Rescoring, Parameter-efficient Fine-tuning, Low-rank Adaptation

Host: Huck Yang, Ivan Bulyko