Abby(Rong) Xu

I am a Ph.D. student at Stevens Institute of Technology ECE department, advising by Prof. Xiaojiang Du. My research interests include IoT Security, AI for security and privacy, and LLM security.

Curriculum Vitae

Education
  • Stevens Institute of Technology
    Stevens Institute of Technology
    Department of Electrical Engineering and Computer Science
    Ph.D. in Computer Engineering
    Expected Dec. 2028
  • University of Texas at Austin
    University of Texas at Austin
    M.S. in Computer Science
    Expected Dec. 2026
  • Texas A&M University
    Texas A&M University
    B.S. in Computer Science, Minor in Mathematics & Cybersecurity
    Dec. 2023
Experience
  • Amazon
    Amazon
    Software Development Engineer Intern
    May 2022 - Aug. 2022
  • Splunk
    Splunk
    Technical Marketing Engineer Intern - Security
    May 2021 - Aug. 2021
News
2024
One paper is accepted to AAAI 2025!
Dec 09
I'm beginning my PhD studies at Stevens Institute of Technology.
Sep 01
2023
I graduated from Texas A&M University with B.S. of Computer Science.
Dec 01
Selected Publications (view all )
Mapping from Meaning: Addressing the Miscalibration of Prompt-Sensitive Language Models
Mapping from Meaning: Addressing the Miscalibration of Prompt-Sensitive Language Models

Kyle Cox, Jiawei Xu, Yikun Han, Rong Xu, Tianhao Li, Chi-Yang Hsu, Tianlong Chen, Walter Gerych, Ying Ding# (# corresponding author)

Annual AAAI Conference on Artificial Intelligence (AAAI) 2024 Poster

An interesting behavior in large language models (LLMs) is prompt sensitivity. When provided with different but semantically equivalent versions of the same prompt, models may produce very different distributions of answers. This suggests that the uncertainty reflected in a model's output distribution for one prompt may not reflect the model's uncertainty about the meaning of the prompt. We model prompt sensitivity as a type of generalization error, and show that sampling across the semantic concept space with paraphrasing perturbations improves uncertainty calibration without compromising accuracy. Additionally, we introduce a new metric for uncertainty decomposition in black-box LLMs that improves upon entropy-based decomposition by modeling semantic continuities in natural language generation. We show that this decomposition metric can be used to quantify how much LLM uncertainty is attributed to prompt sensitivity. Our work introduces a new way to improve uncertainty calibration in prompt-sensitive language models, and provides evidence that some LLMs fail to exhibit consistent general reasoning about the meanings of their inputs.

Mapping from Meaning: Addressing the Miscalibration of Prompt-Sensitive Language Models

Kyle Cox, Jiawei Xu, Yikun Han, Rong Xu, Tianhao Li, Chi-Yang Hsu, Tianlong Chen, Walter Gerych, Ying Ding# (# corresponding author)

Annual AAAI Conference on Artificial Intelligence (AAAI) 2024 Poster

An interesting behavior in large language models (LLMs) is prompt sensitivity. When provided with different but semantically equivalent versions of the same prompt, models may produce very different distributions of answers. This suggests that the uncertainty reflected in a model's output distribution for one prompt may not reflect the model's uncertainty about the meaning of the prompt. We model prompt sensitivity as a type of generalization error, and show that sampling across the semantic concept space with paraphrasing perturbations improves uncertainty calibration without compromising accuracy. Additionally, we introduce a new metric for uncertainty decomposition in black-box LLMs that improves upon entropy-based decomposition by modeling semantic continuities in natural language generation. We show that this decomposition metric can be used to quantify how much LLM uncertainty is attributed to prompt sensitivity. Our work introduces a new way to improve uncertainty calibration in prompt-sensitive language models, and provides evidence that some LLMs fail to exhibit consistent general reasoning about the meanings of their inputs.

All publications