profile photo

  • Yilun Zhao

I am a CS PhD student at Yale working with Professor Arman Cohan on Natural Language Processing and Large Language Model. My current focus areas include: (1) improving the reasoning capabilities, robustness, and interpretability of NLP models; (2) applying LLMs to expert domains; and (3) developing multimodal NLP systems. If you are also passionate about these topics, feel free to connect!

News

[5/2023] 4/5 papers (summarization, text-to-sql, summarization evaluation, and an industry-track paper for table-to-text) accepted to EMNLP 2023.
[5/2023] 3/4 papers (robust QA, summmarization evaluation, and a demo paper for table reasoning) accepted to ACL 2023.
[3/2023] Accepted Yale CS PhD offer. Look forward to starting a new journey in NLP!
[1/2023] 1/1 paper accepted to EACL 2023.
[10/2022] 2/2 papers on Table Reasoning and Faithful Text Generation accepted to EMNLP 2022.
[3/2022] 1/1 paper on Numerical Reasoning in Tables and Texts accepted to ACL 2022.

2023

KnowledgeMath: Knowledge-Intensive Math Word Problem Solving in Finance Domains
Yilun Zhao*, Hongjun Liu*, Yitao Long, Rui Zhang, Chen Zhao, Arman Cohan
Preprint [code/data]

Investigate LLMs' capabilities in solving knowledge-intensive math word problems.

DocMath-Eval: Evaluating Numerical Reasoning Capabilities of LLMs in Understanding Long Documents with Tabular Data
Yilun Zhao*, Yitao Long*, Hongjun Liu, Linyong Nan, Lyuhao Chen, Ryo Kamoi, Yixin Liu, Xiangru Tang, Rui Zhang, Arman Cohan
Preprint

Propose a new evaluation benchmark for numerical reasoning in financial documents containing both textual and tabular data.

QTSumm: Query-Focused Summarization over Tabular Data
Yilun Zhao, Zhenting Qi, Linyong Nan, Boyu Mi, Yixin Liu, Weijin Zou, Simeng Han, Ruizhe Chen, Xiangru Tang, Yumo Xu, Dragomir Radev, Arman Cohan
EMNLP 2023 long [code/data]

Propose a new query-focused table summarization task where text generation models have to perform human-like reasoning and analysis over the given table to generate a tailored summary.

Investigating Table-to-Text Generation Capabilities of LLMs in Real-World Information Seeking Scenarios
Yilun Zhao*, Haowei Zhang*, Shengyun Si*, Linyong Nan, Xiangru Tang, Arman Cohan
EMNLP 2023 industry [code/data]

LLM outperforms fine-tuned systems in table-to-text generation, evaluation, and feedback generation.

A Systematic Study of Table QA Robustness Against Human-Annotated Adversarial Perturbations
Yilun Zhao, Chen Zhao, Linyong Nan, Zhenting Qi, Wenlin Zhang, Xiangru Tang, Boyu Mi, Dragomir Radev
ACL 2023 long [code/data]

Analyze robustness of TableQA systems using a human labeled evaluation set of targeted adversarial perturbations.

LOFT: Enhancing Faithfulness and Diversity for Table-to-Text Generation via Logic Form Control
Yilun Zhao*, Zhenting Qi*, Linyong Nan, LJ Yu Flores, Dragomir Radev
EACL 2023 short (Oral) [code/data]

Apply logical form as fact checker and content planner to improve faithfulness and text-diversity of logical table-to-text generation.

2022

ReasTAP: Injecting Table Reasoning Skills During Pre-training via Synthetic Reasoning Examples
Yilun Zhao, Linyong Nan, Zhenting Qi, Rui Zhang, Dragomir Radev
EMNLP 2022 long [code/data]

Inject table reasoning skills into large language models by synthesizing corresponding QA examples as pre-training corpus.

MultiHiertt: Numerical Reasoning over Multi Hierarchical Tabular and Textual Data
Yilun Zhao, Yunxiang Li, Chenying Li, Rui Zhang
ACL 2022 long [code/data]

Construct QA benchmark that requires model to aggregate numerical information from text and hierarchical tables to answer complex questions about financial reports.



People Empowering Me

During my graduate study, I really appreciate the support from:
  • Prof Arman Cohan
  • Prof Dragomir Radev
  • Prof Rui Zhang (PSU)
  • Prof Chen Zhao (NYU Shanghai)
  • My family

Professional Services

Program Committee / Reviewer: ACL Rolling Review 2023 & 2022 & 2021, ICLR 2024, NeurIPS 2023, EACL 2023, EMNLP 2022



Updated at Jan 2024

Thanks Jon Barron for this amazing template!