publications

(* indicates equal contribution)

2025

  1. reverse.png
    Are Large Language Models Reliable AI Scientists? Assessing Reverse-Engineering of Black-Box Systems
    Jiayi Geng*, Howard Chen*, Dilip Arumugam, and Thomas L Griffiths
    arXiv preprint arXiv:2505.17968, 2025
  2. cognitive.png
    Using the tools of cognitive science to understand large language models at different levels of analysis
    Alexander Ku, Declan Campbell, Xuechunzi Bai, Jiayi Geng, 8 authors, and Thomas L Griffiths
    arXiv preprint arXiv:2503.13401, 2025
  3. mind_your_step.png
    Mind your step (by Step): Chain-of-thought Can Reduce Performance on Tasks Where Thinking Makes Humans Worse
    Ryan Liu*Jiayi Geng*, Addison J Wu, Ilia Sucholutsky, Tania Lombrozo, and Thomas L Griffiths
    ICML, 2025
  4. llm_rationality.png
    Large language Models Assume People are More Rational than We Really are
    Ryan Liu*Jiayi Geng*, Joshua C Peterson, Ilia Sucholutsky, and Thomas L Griffiths
    ICLR, 2025

2024

  1. cm.png
    Continual Memorization of Factoids in Large Language Models
    Howard Chen*Jiayi Geng*, Adithya Bhaskar, Dan Friedman, and Danqi Chen
    arXiv preprint 2411.07175, 2024
  2. treebon.png
    TreeBoN: Enhancing Inference-Time Alignment with Speculative Tree-Search and Best-of-N Sampling
    Jiahao Qiu, Yifu Lu, Yifan Zeng, Jiacheng Guo, Jiayi Geng, Huazheng Wang, Kaixuan Huang, Yue Wu, and Mengdi Wang
    arXiv preprint 2410.16033, 2024
  3. dr_gpt.png
    Dr. GPT in Campus Counseling: Understanding Higher Education Students’ Opinions on LLM-assisted Mental Health Services
    Owen Xingjian Zhang, Shuyao Zhou, Jiayi Geng, Yuhan Liu, and Sunny Xun Liu
    arXiv preprint 2409.17572, 2024
  4. science_tutor.png
    Language Models as Science Tutors
    Alexis Chevalier, Jiayi Geng, Alexander Wettig, Howard Chen, 16 authors, Sanjeev Arora, and Danqi Chen
    ICML, 2024

2023

  1. corgi.png
    Corgi-pm: A Chinese Corpus for Gender Bias Probing and Mitigation
    Ge Zhang, Yizhi Li, Yaoyao Wu, Linyuan Zhang, Chenghua Lin, Jiayi Geng, Shi Wang, and Jie Fu
    arXiv preprint 2301.00395, 2023