Prospective Students

Thank you for your interest in working with me! I am currently receiving a high volume of emails regarding prospective students. I have written this page to address any questions you may have. If you have additional questions that I have not answered below, feel free to email me with the header [Prospective Student].

What do I work on?

My area of research is Natural Language Processing (NLP). My main focus is on Responsible NLP, which encompasses several subfields in NLP. My work is very interdisciplinary and as a result, I have collaborated with researchers in public health, sociology, social work, and HCI. I have listed my research interests below:

  1. Fairness: This focuses on detecting and mitigating social biases in natural language systems. I am especially interested in cross-cultural fairness issues, which may include biases in non-English and/or non-American settings. Some related papers include:
    • Sharon Levy, Neha Anna John, Ling Liu, Yogarshi Vyas, Jie Ma, Yoshinari Fujinuma, Miguel Ballesteros, Vittorio Castelli, Dan Roth. “Comparing Biases and the Impact of Multilingual Training across Multiple Languages” (EMNLP 2023)[paper]
    • Sophie Groenwold*, Lily Ou*, Aesha Parekh*, Samhita Honnavalli*, Sharon Levy, Diba Mirza and William Yang Wang. “Investigating African-American Vernacular English in Transformer-Based Text Generation” (EMNLP 2020) [paper]
    • Samhita Honnavalli*, Aesha Parekh*, Lily Ou*, Sophie Groenwold*, Sharon Levy, Vicente Ordonez and William Yang Wang. “Towards Understanding Gender-Seniority Compound Bias in Natural Language Generation” (LREC 2022) [paper]
  2. Safety: We have already seen cases in which NLP models have generated harmful information to users. This line of research studies issues related to user safety such as covertly unsafe text, i.e. text that is implicitly harmful and requires commonsense reasoning to deem it unsafe. Relevant papers include:
    • Sharon Levy, Emily Allaway, Melanie Subbiah, Lydia Chilton, Desmond Patton, Kathleen McKeown and William Yang Wang. “SafeText: A Benchmark for Exploring Physical Safety in Language Models” (EMNLP 2022)[paper]
    • Alex Mei*, Anisha Kabir*, Sharon Levy, Melanie Subbiah, Emily Allaway, John N. Judge, Desmond Patton, Bruce Bimber, Kathleen McKeown and William Yang Wang. “Mitigating Covertly Unsafe Text within Natural Language Systems” (EMNLP Findings 2022)[paper]
    • Alex Mei*, Sharon Levy*, William Yang Wang. “Foveate, Attribute, and Rationalize: Towards Safe and Trustworthy AI” (ACL Findings 2023)[paper]
    • Alex Mei*, Sharon Levy*, William Yang Wang. “ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models” (EMNLP Findings 2023)[paper]
  3. Trustworthiness: As natural language systems are trained, they may learn information that becomes outdated over time or is incorrect in general. We focus on both harmful (e.g. conspiracy theories) and less harmful (e.g. satire) types of misinformation. The following lists related papers:
    • Sharon Levy, Michael Saxon, William Yang Wang. “Investigating Memorization of Conspiracy Theories in Text Generation” (ACL Findings 2021)[paper][blog post]
    • Sharon Levy*, Kai Nakamura*, William Yang Wang. “r/Fakeddit: A New Multimodal Benchmark Dataset for Fine-grained Fake News Detection” (LREC 2020)[paper]
  4. Question Answering: Related to trustworthiness, this focuses specifically on how natural language systems can be used as question answering applications and ensures that models can answer queries truthfully. Related papers include:
    • Matthew Ho*, Aditya Sharma*, Justin Chang*, Michael Saxon, Sharon Levy, Yujie Lu and William Yang Wang. “WikiWhy: Answering and Explaining Cause-and-Effect Questions” (ICLR 2023)[paper]
    • Sharon Levy, Kevin Mo, Wenhan Xiong, William Yang Wang. “Open-Domain Question-Answering for COVID-19 and Other Emergent Domains” (EMNLP 2021 Demo)[paper]
    • Kai Nakamura, Sharon Levy, Yi-Lin Tuan, Wenhu Chen, William Yang Wang. “HybriDialogue: An Information-Seeking Dialogue Dataset Grounded on Tabular and Textual Data” (ACL Findings 2022)[paper]
    • Alon Albalak, Sharon Levy, William Yang Wang. “Addressing Issues of Cross-Linguality in Open-Retrieval Question Answering Systems For Emergent Domains” (EACL 2023 Demo)[paper]
  5. Computational Social Science: Here, we use NLP as a tool to study interactions in the real world and online. The following lists my relevant work:
    • Sharon Levy, Robert E. Kraut, Jane A. Yu, Kristen M. Altenburger, Yi-Chia Wang. “Understanding Conflicts in Online Conversations” (WWW ‘22)[paper]

What is my advising style?

I plan to work closely with my students and hold weekly 1-on-1s with each student. In addition to these meetings, I will also hold a weekly lab group meeting in which students will provide updates on their progress in order to help form collaborations and overcome project difficulties. These meetings will also contain a reading group, where students will present interesting and relevant research papers that have been recently published to practice their presentation skills and stay up-to-date on current research.

Prospective PhD Students

If you are interested in working with me, please apply to the Rutgers CS department as a PhD student and specify my name as a potential advisor. You can find more information here. The deadline is January 1st. As I am receiving a high volume of emails, please do not email me with your CV. I will go through the applications once the deadline is over and if I see a good fit, I will contact you for an interview.

Current Rutgers Students

I will also be accepting students who currently study at Rutgers. If you are interested in working with me, please email me with the header [Current Rutgers Student] closer to the start of the Fall 2024 semester.