Aurora Team, “Aurora-M: The First Open Source Multilingual Language Model Red-teamed according to the U.S. Executive Order”, Manuscript, 2024.
Aurora Team, “Aurora-M: The First Open Source Multilingual Language Model Red-teamed according to the U.S. Executive Order”, Manuscript, 2024.
Zhou Yang, Zhensu Sun, Terry Zhuo Yue, Premkumar Devanbu, David Lo, “Robustness, Security, Privacy, Explainability, Efficiency, and Usability of Large Language Models for Code”, Manuscript, 2024.
BigCode Team, “StarCoder 2 and The Stack v2: The Next Generation”, Manuscript, 2024.
Niklas Muennighoff, Qian Liu, Armel Zebaze, Qinkai Zheng, Binyuan Hui, Terry Yue Zhuo, Swayam Singh, Xiangru Tang, Leandro von Werra and Shayne Longpreo, “OctoPack: Instruction Tuning Code Large Language Models”, International Conference on Learning Representations (ICLR), 2024. [Github]
Terry Yue Zhuo, “ICE-Score: Instructing Large Language Models to Evaluate Code”, Findings of European Chapter of the Association for Computational Linguistics (EACL), 2024. [Github]
Terry Yue Zhuo, Armel Zebaze, Nitchakarn Suppattarachai, Leandro von Werra, Harm de Vries, Qian Liu, Niklas Muennighoff, “Astraios: Parameter-Efficient Instruction Tuning Code Large Language Models”, Manuscript, 2024. [MarkTechPost] [Tweet] [Github]
Guang Yang, Yu Zhou, Xiang Chen, Xiangyu Zhang, Terry Yue Zhuo, Taolue Chen, “Chain-of-Thought in Neural Code Generation: From and For Lightweight Language Models”, Manuscript, 2023.
Terry Yue Zhuo, Zhou Yang, Zhensu Sun, Yufei Wang, Li Li, Xiaoning Du, Zhenchang Xing and David Lo, “Data Augmentation Approaches for Source Code Models: A Survey”, Manuscript, 2023. [Github]
BigCode Team, “StarCoder: May The Source Be With You!”, Transactions on Machine Learning Research (TMLR), 2023. [Tweet 1] [Tweet 2] [TechCrunch] [StarCoderBase] [StarCoder] [Github]
Terry Yue Zhuo, Xiaoning Du, Zhenchang Xing, Jiamou Sun, Haowei Quan, Li Li and Liming Zhu, “Probing API Name Knowledge in Pre-trained Code Models”, Manuscript, 2023.
BigCode Team, “SantaCoder: don't reach for the stars!”, Deep Learning For Code workshop (DL4C) @ ICLR, 2023. [Best Paper Award] [Tweet 1] [Tweet 2] [SantaCoder]
Han Hu, Yujin Huang, Qiuyuan Chen, Terry Yue Zhuo and Chunyang Chen, “A first look at the robustness of on-device models on iOS apps”, ACM Transactions on Software Engineering and Methodology (TOSEM), 2023.
Terry Yue Zhuo, Zhuang Li, Yujin Huang, Yuan-Fang Li, Fatemeh Shiri, Weiqing Wang and Reza Haffari, “On Robustness of Prompt-based Semantic Parsing with Large Pre-trained Language Model: An Empirical Study on Codex”, European Chapter of the Association for Computational Linguistics (EACL), 2023.
Yujin Huang*, Terry Yue Zhuo*, Qiongkai Xu, Han Hu, Xingliang Yuan and Chunyang Chen, “Training-free Lexical Backdoor Attacks on Language Models”, The Web Conference (WWW), 2023.
Xiaoxi Kang, Lizhen Qu, Lay-Ki Soon, Adnan Trakic, Terry Yue Zhuo, Patrick Charles Emerton and Genevieve Grant, “Can ChatGPT Perform Reasoning Using the IRAC Method in Analyzing Legal Scenarios Like a Lawyer?”, Findings of Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023.
Jinyan Su, Terry Yue Zhuo, Di Wang and Preslav Nakov, “DetectLLM: Leveraging Log Rank Information for Zero-Shot Detection of Machine-Generated Text”, Findings of Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023. [Github]
Jinyan Su*, Terry Yue Zhuo*, Di Wang and Preslav Nakov, “Fake News Detectors are Biased against Texts Generated by Large Language Models”, Manuscript, 2023. [Hacker News]
Terry Yue Zhuo, Qiongkai Xu, Xuanli He and Trevor Cohn, “Rethinking Round-trip Translation for Machine Translation Evaluation”, Findings of Annual Meeting of the Association for Computational Linguistics (ACL), 2023. [Github]
Zhuang Li, Yuyang Chai*, Terry Yue Zhuo*, Lizhen Qu, Gholamreza Haffari, Fei Li, Donghong Ji and Quan Hung Tran, “FACTUAL: A Benchmark for Faithful and Consistent Multimodal Scene Graph Parsing”, Findings of Annual Meeting of the Association for Computational Linguistics (ACL), 2023.
Terry Yue Zhuo, Yujin Huang, Chunyang Chen and Zhenchang Xing, “Red teaming ChatGPT via Jailbreaking: Bias, Robustness, Reliability and Toxicity”, Manuscript, 2023. [University of West Florida Library] [Brunel University London Course Material]
Terry Yue Zhuo, Yaqing Liao, Yuecheng Lei, Yazhou Ren, Xiaojun Chang, Zenglin Xu, Lizhen Qu and Gerard de Melo, “ViLPAct: A Benchmark for Compositional Generalization on Multimodal Human Activities”, Findings of European Chapter of the Association for Computational Linguistics (EACL), 2023. [Github]
Fatemeh Shiri*, Terry Yue Zhuo*, Zhuang Li*, Van Nguyen, Shirui Pan, Teresa Wang, Reza Haffari and Yuan-Fang Li, “Paraphrasing Techniques for Maritime QA System”, International Conference on Information Fusion (FUSION), 2022.
Farhad Moghimifar, Lizhen Qu, Terry Yue Zhuo, Mahsa Baktashmotlagh and Gholamreza Haffari, “Dynamic Neural-Symbolic Reasoner on Commonsense Knowledge”, Manuscript, 2021.
Farhad Moghimifar, Lizhen Qu, Terry Yue Zhuo, Gholamreza Haffari and Mahsa Baktashmotlagh, “Neural Symbolic Commonsense Reasoner with Relation Predictors”, Annual Meeting of the Association for Computational Linguistics (ACL), 2021.
Jason Rumengan, Terry Yue Zhuo and Conrad Sanderson, “PyArmadillo: an alternative approach to linear algebra in Python”, Journal of Open Source Software (JOSS), 2021.
Farhad Moghimifar, Lizhen Qu, Yue Zhuo, Mahsa Baktashmotlagh and Gholamreza Haffari, “CosMo: Conditional SEQ2SEQ-based Mixture Model for Zero-Shot Commonsense Question Answering”, International Conference on Computational Linguistics (COLING), 2020.