BigCode Team, “BigCodeBench: Benchmarking Code Generation with Diverse Function Calls and Complex Instructions ”, Manuscript, 2024. [Github]
I am getting my PhD at CSIRO's Data61 and Monash University.
My research concerns the intelligence of program-aided language models, mainly in function calls, reasoning, and generation. The long-term goal is to build AGI with code intelligence.
Paper is cheap. Show me what you can build for the community. What you create defines who you are.
I am always open to collaboration. Feel free to reach out to me if you are interested in my work.
Recent Publications
BigCode Team, “OctoPack: Instruction Tuning Code Large Language Models”, International Conference on Learning Representations (ICLR), 2024. [Github]
Terry Yue Zhuo, “ICE-Score: Instructing Large Language Models to Evaluate Code”, Findings of European Chapter of the Association for Computational Linguistics (EACL), 2024. [Github]
BigCode Team, “Astraios: Parameter-Efficient Instruction Tuning Code Large Language Models”, Manuscript, 2024.
BigCode Team, “StarCoder: May The Source Be With You!”, Transactions on Machine Learning Research (TMLR), 2023. [Tweet 1] [Tweet 2] [TechCrunch] [StarCoderBase] [StarCoder] [Github]
BigCode Team, “SantaCoder: don't reach for the stars!”, Deep Learning For Code workshop (DL4C) @ ICLR, 2023. [Best Paper Award] [Tweet 1] [Tweet 2] [SantaCoder]