
Bio
I am a Senior Scientist at Bloomberg. At Bloomberg, I have been building an interactive code agent to help analysts retrieve and analyze the various data provided by Bloomberg. This involves implementing various training procedures to improve large language models (LLMs) on capabilities such as tool calling, and training hybrid retrieval models to improve contextual accuracy in code generation.
Previously, I obtained my PhD from Johns Hopkins University, where I was advised by Prof. Benjamin Van Durme. I worked on large language model (LLM) projects including using reinforcement learning (RL) to better orchestrate between LLMs and retrieval systems for code generation [paper], faithfullness evaluation [paper], and preference alignment algorithms [paper, paper]. My research also explored sequence modeling [paper], particularly in the context of streaming, and developed evaluation metrics for code and other structures [paper].
I was awarded outstanding paper awards for Learning to Retrieve Iteratively for In-Context Learning at EMNLP 2024 and Iterative Document-level Information Extraction via Imitation Learning at EACL 2023.
Publications
Streaming Sequence Transduction through Dynamic Compression
Learning to Retrieve Iteratively for In-Context Learning
FaithScore: Evaluating Hallucinations in Large Vision-Language Models
The Language Barrier: Dissecting Safety Challenges of LLMs in Multilingual Contexts
Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation
Narrowing the Gap between Zero- and Few-shot Machine Translation by Matching Styles
MultiMUC: Multilingual Template Filling on MUC-4
A Unified View of Evaluation Metrics for Structured Prediction
Condensing Multilingual Knowledge with Lightweight Language-Specific Modules
On Event Individuation for Document-Level Information Extraction
When Do Decompositions Help for Machine Reading?
Iterative Document-level Information Extraction via Imitation Learning
Differentiable Tree Operations Promote Compositional Generalization
An Empirical Study on Finding Spans
Everything Is All It Takes: A Multipronged Strategy for Zero-Shot Cross-Lingual Information Extraction
LOME: Large Ontology Multilingual Extraction
Pattern-aware Data Augmentation for Query Rewriting in Voice Assistant Systems
Hierarchical Entity Typing via Multi-level Learning to Rank
Joint Modeling of Arguments for Event Understanding
Reading the Manual: Event Extraction as Definition Comprehension