SSNLP 2024

The 2024 Singapore Symposium on Natural Language Processing

Welcome!

We are excited to announce that the Singapore Symposium on Natural Language Processing (SSNLP 2024) will take place on Wednesday, November 6, as a full-day event. SSNLP, an annual pre-conference workshop, gathers the Natural Language Processing community in Singapore, bringing together local students, practitioners, and faculty. It offers a valuable platform to connect, exchange ideas, and foster collaboration.

Since its inception in 2018, SSNLP has steadily grown in both popularity and influence, with successful editions held in 2018, 2019, 2020, 2022, and 2023. We look forward to continuing this tradition in 2024 and delivering another impactful experience for all participants.

This year's event will be held at the Mapletree Business City, Town Hall Auditorium, (10 Pasir Panjang Road, Singapore 117438, [Google Map], [Venue Roadmap], [Outdoor photo], [Indoor photo]). Please note that seating is limited. In the event of oversubscription, we may consider virtual attendance for registered participants. We encourage early bird registration to secure your spot. The on-site registration deadline is October 31st, 23:59; make sure you complete the registration before this!

On-site Registration Closed

Our in-person registration is now closed due to high demand. For those interested in virtual attendance, please complete the form below. Please note that priority for all Q&A sessions will be given to on-site attendees. The virtual registration deadline is Nov 5th, 17:00 SGT.

Virtual Registration Closed

Latest news


Nov 1, 2024 — Virtual registration deadline is Nov 5th, 17:00 SGT, please register before this!

Oct 30, 2024 — On-site registration deadline is October 31st, 23:59, please register before this!

Oct 20, 2024 — Registration is open, please register now

Oct 15, 2024 — The date is confirmed: November 6, 2024

Programme

We've planned to host Oral and Poster sessions for research presentations, and also have invited Keynote Presentations as well as Industry Talks.

Time Event
09:00 - 09:15 Welcome and Opening Remarks
09:15 - 10:15 Remote Keynote 1 (45mins + 15mins Q&A)
speaker:   Jason Wei
session chair:   Roy Lee    
10:15 - 11:45 Remote Keynote 2 - 4 (each 20mins + 10mins Q&A)
speakers:   Bing Liu & Yue Zhang & Jing Ma    
session chairs:   Yang Deng & Anh Tuan Luu    
11:45 - 12:45 Research Oral Presentation 1 (each 12mins + 3mins Q&A)
session chair:   Haonan Wang    
12:45 - 14:00 Poster Session 1
Lunch (w/ Career & Networking Session)
14:00 - 15:30 On-Site Industry Talks (each 20mins + 10mins Q&A)
speakers:   Tianyu Pang & Wenxuan Zhang & Taifeng Wang
session chair:   Jiaying Wu    
15:30 - 16:45 Poster Session 2
15:45 - 16:30 Research Oral Presentation 2 (each 12mins + 3mins Q&A)
session chair:   Moxin Li    
16:30 - 17:00 Coffee break
17:00 - 17:45 Research Oral Presentation 3 (each 12mins + 3mins Q&A)
session chair:   Zhiyuan Hu    
17:45 - 18:00 Closing

Oral & Poster Presentations

Oral presentations are for 12 minutes plus 3 minutes for immediate questions. Poster boards can accommodate (1m x 1m) sized posters, in either portrait or landscape. We arrange the posters into Poster Session 1 and Poster Session 2, divided as follows, with each containing around 20 posters.

Click to see the paper list ↓
Paper session 1   

[1] Zhiyuan Hu, Chumin Liu, Xidong Feng, Yilun Zhao, See-Kiong Ng, Anh Tuan Luu, Junxian He, Pang Wei Koh, Bryan Hooi. Uncertainty of Thoughts: Uncertainty-Aware Planning Enhances Information Seeking in Large Language Models. NeurIPS (Slot: 11:45 - 12:00)
[2] Lin Xu, Zhiyuan Hu, Daquan Zhou, Hongyu Ren, Zhen Dong, Kurt Keutzer, See-Kiong Ng, Jiashi Feng. MAgIC: Investigation of Large Language Model Powered Multi-Agent in Cognition, Adaptability, Rationality and Collaboration. EMNLP main (Slot: 12:00 - 12:15)
[3] Xuan Zhang, Chao Du, Tianyu Pang, Qian Liu, Wei Gao, Min Lin. Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs. NeurIPS (Slot: 12:15 - 12:30)
[4] Xiaobao Wu, Liangming Pan, William Yang Wang, Anh Tuan Luu. AKEW: Assessing Knowledge Editing in the Wild. EMNLP main (Slot: 12:30 - 12:45)

Paper session 2   

[1] Ming Shan Hee, Aditi Kumaresan, Roy Ka-Wei Lee. Bridging Modalities: Enhancing Cross-Modality Hate Speech Detection with Few-Shot In-Context Learning. EMNLP main (Slot: 15:45 - 16:00)
[2] Yang Deng, Yong Zhao, Moxin Li, See-Kiong Ng, Tat-Seng Chua. Don't Just Say "I Don't Know"! Self-aligning Large Language Models for Responding to Unknown Questions with Explanations. EMNLP main (Slot: 16:00 - 16:15)
[3] Jiahao Ying, Yixin Cao, Yushi Bai, Qianru Sun, Bo Wang, Wei Tang, Zhaojun Ding, Yizhe Yang, Xuanjing Huang, Shuicheng Yan. Automating Dataset Updates Towards Reliable and Timely Evaluation of Large Language Models. NeurIPS (Slot: 16:15 - 16:30)

Paper session 3   

[1] Yew Ken Chia, Guizhen Chen, Weiwen Xu, Luu Anh Tuan, Soujanya Poria, Lidong Bing. Reasoning Paths Optimization: Learning to Reason and Explore From Diverse Paths. EMNLP main (Slot: 17:00-17:15)
[2] Yiran Zhao, Wenyue Zheng, Tianle Cai, Xuan Long Do, Kenji Kawaguchi, Anirudh Goyal, Michael Shieh. Accelerating Greedy Coordinate Gradient and General Prompt Optimization via Probe Sampling. NeurIPS (Slot: 17:15-17:30)
[3] Wenyang Hu, Yao Shu, Zongmin Yu, Zhaoxuan Wu, Xiaoqiang Lin, Zhongxiang Dai, See-Kiong Ng, Bryan Kian Hsiang Low. Localized Zeroth-Order Prompt Optimization. NeurIPS spotlight (Slot: 17:30 - 17:45)

Poster session 1  

[1] Hai Ye, Hwee Tou Ng. Preference-Guided Reflective Sampling for Aligning Language Models. (Board: 1)
[2] Do Xuan Long, Duong Ngoc Yen, Anh Tuan Luu, Kenji Kawaguchi, Min-Yen Kan, Nancy F. Chen. Multi-expert Prompting Improves Reliability, Safety and Usefulness of Large Language Models (Board: 2)
[3] Naaman Tan, Josef Valvoda, Tianyu Liu, Anej Svete, Yanxia Qin, Min-Yen Kan, Ryan Cotterell. A Fundamental Trade-off in Aligned Language Models and its Relation to Sampling Adaptors (Board: 3)
[4] Esther Gan*, Yiran Zhao*, Liying Cheng, Mao Yancan, Anirudh Goyal, Kenji Kawaguchi, Min-Yen Kan, Michael Shieh. Reasoning Robustness of LLMs to Adversarial Typographical Errors (Board: 4)
[5] Hongfu Liu, Yuxi Xie, Ye Wang, Michael Shieh. Advancing Adversarial Suffix Transfer Learning on Aligned Large Language Models (Board: 5)
[6] Thong Nguyen, Zhiyuan Hu, Xiaobao Wu, Cong-Duy T Nguyen, See-Kiong Ng, Anh Tuan Luu. Encoding and Controlling Global Semantics for Long-form Video Question Answering (Board: 6)
[7] Gregory Kang Ruey Lau, Xinyuan Niu, Hieu Dao, Jiangwei Chen, Chuan Sheng Foo, Bryan Kian Hsiang Low. Waterfall: Framework for Robust and Scalable Text Watermarking (Board: 7)
[8] Yunze Xiao, Yujia Hu, Kenny Tsu Wei Choo, Roy Ka-Wei Lee. ToxiCloakCN: Evaluating Robustness of Offensive Language Detection in Chinese with Cloaking Perturbations (Board: 8)
[9] Miaoyu Li, Haoxin Li, Zilin Du, Boyang Li. Diversify, Rationalize, and Combine: Ensembling Multiple QA Strategies for Zero-shot Knowledge-based VQA (Board: 9)
[10] Xuan Zhang, Yang Deng, Zifeng Ren, See-Kiong Ng, Tat-Seng Chua. Ask-before-Plan: Proactive Language Agents for Real-World Planning (Board: 10)
[11] Muhammad Reza Qorib, Alham Fikri Aji, Hwee Tou Ng. Efficient and Interpretable Grammatical Error Correction with Mixture of Experts (Board: 11)
[12] Fengzhu Zeng, Wenqian Li, Wei Gao, Yan Pang. Multimodal Misinformation Detection by Learning from Synthetic Data with Multimodal LLMs (Board: 12)
[13] Jiahao Ying, Mingbao Lin, Yixin Cao, Wei Tang, Bo Wang, Qianru Sun, Xuanjing Huang, Shuicheng Yan. LLMs-as-Instructors: Learning from Errors Toward Automating Model Improvement (Board: 13)
[14] Xinyi Xu, Zhaoxuan Wu, Rui Qiao, Arun Verma, Yao Shu, Jingtan Wang, Xinyuan Niu, Zhenfeng He, Jiangwei Chen, Zijian Zhou, Gregory Kang Ruey Lau, Hieu Dao, Lucas Agussurja, Rachael Hwee Ling Sim, Xiaoqiang Lin, Wenyang Hu, Zhongxiang Dai, Pang Wei Koh, Bryan Kian Hsiang Low. Position Paper: Data-Centric AI in the Age of Large Language Models (Board: 14)
[15] Ming Shan Hee, Shivam Sharma, RUI CAO, Palash Nandi, Preslav Nakov, Tanmoy Chakraborty, Roy Ka-Wei Lee. Recent Advances in Online Hate Speech Moderation: Multimodality and the Role of Large Models (Board: 15)
[16] Xu Guo, Zilin Du, Boyang Li, Chunyan Miao. Generating Synthetic Datasets for Few-shot Prompt Tuning (Board: 16)
[17] Quanyu Long, Yin Wu, Wenya Wang, Sinno Jialin Pan. Does In-Context Learning Really Learn? Rethinking How Large Language Models Respond and Solve Tasks via In-Context Learning (Board: 17)
[18] Yujia Hu, Zhiqiang Hu, Chun Wei Seah, Roy Ka-Wei Lee. InstructAV: Instruction Fine-tuning Large Language Models for Authorship Verification (Board: 18)
[19] Zhengyuan Liu, Stella Xin Yin, Geyu Lin, Nancy F. Chen. Personality-aware Student Simulation for Conversational Intelligent Tutoring Systems (Board: 19)
[20] Moxin Li, Wenjie Wang, Fuli Feng, Fengbin Zhu, Qifan Wang, Tat-Seng Chua. Think Twice Before Trusting: Self-Detection for Large Language Models through Comprehensive Answer Reflection (Board: 20)

Poster session 2  

[1] Suzanna Sia, David Mueller, Kevin Duh. Where does In context learning happen in LLMs (Board: 1)
[2] Hao Fei, Shengqiong Wu, Hanwang Zhang, Tat-Seng Chua, Shuicheng Yan. VITRON: A Unified Pixel-level Vision LLM for Understanding, Generating, Segmenting, Editing (Board: 2)
[3] Zijian Zhou, Xiaoqiang Lin, Xinyi Xu, Alok Prakash, Daniela Rus, Bryan Kian Hsiang Low. DETAIL: Task DEmonsTration Attribution for Interpretable In-context Learning (Board: 3)
[4] Zhaoxuan Wu, Xiaoqiang Lin, Zhongxiang Dai, Wenyang Hu, Yao Shu, See-Kiong Ng, Patrick Jaillet, Bryan Kian Hsiang Low. Prompt Optimization with EASE? Efficient Ordering-aware Automated Selection of Exemplars (Board: 4)
[5] Zhuanghua Liu, Luo Luo, Bryan Kian Hsiang Low. Gradient-Free Methods for Nonconvex Nonsmooth Stochastic Compositional Optimization (Board: 5)
[6] Mingzhe Du, Anh Tuan Luu, Bin Ji, Qian Liu, See-Kiong Ng. Mercury: A Code Efficiency Benchmark for Code Large Language Models (Board: 6)
[7] Bhardwaj, Rishabh, Do Duc Anh, Soujanya Poria. Language Models are Homer Simpson! Safety Re-Alignment of Fine-tuned Language Models through Task Arithmetic (Board: 7)
[8] Ruichao Yang, Wei Gao, Jing Ma, Hongzhan Lin, Bo Wang. Reinforcement Tuning for Detecting Stances and Debunking Rumors Jointly with Large Language Models (Board: 8)
[9] Fengzhu Zeng, Wei Gao. JustiLM: Few-Shot Justification Generation for Explainable Fact-Checking of Real-world Claims (Board: 9)
[10] Xinze Li, Yixin Cao, Liangming Pan, Yubo Ma, Aixin Sun. Towards Verifiable Generation: A Benchmark for Knowledge-aware Language Model Attribution (Board: 10)
[11] Jiahao Ying, Yixin Cao, Kai Xiong, Yidong He, Long Cui, Yongbin Liu. Intuitive or Dependent? Investigating LLMs' Behavior Style to Conflicting Prompts (Board: 11)
[12] Do Xuan Long*, Yiran Zhao*, Hannah Brown*, Yuxi Xie, James Zhao, Nancy Chen, Kenji Kawaguchi, Michael Shieh, Junxian He. Prompt Optimization via Adversarial In-Context Learning (Board: 12)
[13] Jundong Xu, Hao Fei, Liangming Pan, Qian Liu, Mong-Li Lee, Wynne Hsu. Faithful Logical Reasoning via Symbolic Chain-of-Thought (Board: 13)
[14] Cunxiao Du et al.. GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding (Board: 14)
[15] Xiangming Gu*, Xiaosen Zheng*, Tianyu Pang*, Chao Du, Qian Liu, Ye Wang, Jing Jiang, Min Lin. Agent Smith: A Single Image Can Jailbreak One Million Multimodal LLM Agents Exponentially Fast (Board: 15)
[16] Shengqiong Wu, Hao Fei, Leigang Qu, Wei Ji, Tat-Seng Chua. NExT-GPT: Any-to-Any Multimodal LLM (Board: 16)
[17] Hao Fei, Shengqiong Wu, Wei Ji, Hanwang Zhang, Meishan Zhang, Mong-Li Lee, Wynne Hsu. Video-of-thought: Step-by-step video reasoning from perception to cognition (Board: 17)
[18] Suzanna Sia, Alexandra Delucia, Kevin Duh. Anti-Lm Decoding for zeroshot in context MT (Board: 18)
[19] Anthony Tiong, Junqi Zhao, et al.. What Are We Measuring When We Evaluate Large Vision-Language Models? An Analysis of Latent Factors and Biases (Board: 19)
[20] Yidan Sun, Qin Chao, Boyang Li. Event Causality Is Key to Computational Story Understanding (Board: 20)
[21] Brian Formento, Wenjie Feng, Chuan Sheng Foo, Luu Anh Tuan, See-Kiong Ng. SemRoDe: Macro Adversarial Training to Learn Representations That are Robust to Word-Level Attacks (Board: 21)
[22] Fu Jinlan, Ng See-Kiong, Jiang Zhengbao, Liu Pengfei. GPTScore: Evaluate as You Desire (Board: 22)
[23] Meng Luo, Hao Fei, Bobo Li, Shengqiong Wu, Qian Liu, Soujanya Poria, Erik Cambria, Mong-Li Lee, Wynne Hsu. PanoSent: A Panoptic Sextuple Extraction Benchmark for Multimodal Conversational Aspect-based Sentiment Analysis (Board: 23)
[24] Yubo Ma, Yuhang Zang, Liangyu Chen, Meiqi Chen, Yizhu Jiao, Xinze Li, Xinyuan Lu, Ziyu Liu, Yan Ma, Xiaoyi Dong, Pan Zhang, Liangming Pan, Yu-Gang Jiang, Jiaqi Wang, Yixin Cao, Aixin Sun. MMLongBench-Doc: Benchmarking Long-context Document Understanding with Visualizations (Board: 24)
[25] Yajing Yang, Qian Liu, Min-Yen Kan. DataTales: A Benchmark for Real-World Intelligent Data Narration (Board: 25)

Keynote Speakers

The following speakers from both academia and industry are invited to give keynotes at SSNLP 2024. Please click the profile image to view the detailed description of the talk.

Title: Scaling paradigms for large language models
Speaker: Jason Wei @ OpenAI

Abstract: In this talk I will tell you about the role of scaling in the past five years of artificial intelligence. In the first scaling paradigm, which started around five years ago, our field scaled large language models by training with more compute on more data. Such scaling led to the success of ChatGPT and other AI chat engines, which were surprisingly capable and general purpose. With the release of OpenAI o1, we are at the beginning of a new paradigm where we do not just scale training time compute, but we also scale test-time compute. These new models are trained via reinforcement learning on chain-of-thought reasoning, and by thinking harder for more-challenging tasks can solve even competition-level math and programming problems.


Bio: Dr. Jason Wei is an AI researcher based in San Francisco. He currently works at OpenAI, where he contributed to OpenAI o1, a frontier model trained to do chain-of-thought reasoning via reinforcement learning. From 2020 to 2023, Jason was a research scientist at Google Brain, where his work popularized chain-of-thought prompting, instruction tuning, and emergent phenomena.



Title: Lifelong Learning Dialogue Systems
Speaker: Bing Liu @ UIC

Abstract: Dialogue systems, commonly known as chatbots, have gained escalating popularity in recent times due to their wide-spread applications in carrying out chit-chat conversations with users and task-oriented dialogues to accomplish various user tasks. Existing chatbots are usually trained from pre-collected and manually labeled data. Many also use manually compiled knowledge bases (KBs). Their ability to understand natural language is still limited. Typically, they need to be constantly improved by engineers with more labeled data and more manually compiled knowledge. In this talk, I would like to introduce the new paradigm of lifelong learning dialogue systems to endow chatbots the ability to learn continually by themselves through their own self-initiated interactions with their users and working environments. As the systems chat more and more with users, they become more and more knowledgeable and better and better at conversing.


Bio: Dr. Bing Liu is a Distinguished Professor and Peter L. and Deborah K. Wexler Professor of Computing at the University of Illinois Chicago. He received his Ph.D. in Artificial Intelligence (AI) from the University of Edinburgh. His current research interests include continual/lifelong learning, lifelong learning dialogue systems, sentiment analysis, machine learning and natural language processing. He has published extensively in prestigious conferences and journals and authored five books: one about lifelong machine learning, one about lifelong learning dialogue systems, two about sentiment analysis, and one about Web mining. Three of his papers have received the Test-of-Time awards, and another one received Test-of-Time honorable mention. Some of his works have also been widely reported in popular and technology press internationally. He served as the Chair of ACM SIGKDD from 2013-2017 and as program chair of many leading data mining conferences. He is also the winner of 2018 ACM SIGKDD Innovation Award, and is a Fellow of ACM, AAAI, and IEEE.



Title: AutoSurvey: Large Language Models Can Automatically Write Surveys
Speaker: Yue Zhang @ Westlake Unv

Abstract: This talk introduces AutoSurvey, a speedy and well-organized methodology for automating the creation of comprehensive literature surveys in rapidly evolving fields like artificial intelligence. Traditional survey paper creation faces challenges due to the vast volume and complexity of information, prompting the need for efficient survey methods. While large language models (LLMs) offer promise in automating this process, challenges such as context window limitations, parametric knowledge constraints, and the lack of evaluation benchmarks remain. AutoSurvey addresses these challenges through a systematic approach that involves initial retrieval and outline generation, subsection drafting by specialized LLMs, integration and refinement, and rigorous evaluation and iteration. Our contributions include a comprehensive solution to the survey problem, a reliable evaluation method, and experimental validation demonstrating AutoSurvey's effectiveness.


Bio: Dr. Yue Zhang is a tenured Professor at Westlake University. His research interests include NLP and its underlying machine learning algorithms. His major contributions to the field include psycholinguistically motivated machine learning algorithm, learning-guided beam search for structured prediction, pioneering neural NLP models including graph LSTM, and OOD generalization for NLP. He authored the Cambridge University Press book ``Natural Language Processing -- a Machine Learning Perspective''. He is the PC co-chair for CCL 2020 and EMNLP 2022, and action editor for Transactions for ACL. He also served as associate editor for IEEE/ACM Transactions of Audio Speech and Language Processing (TASLP), ACM Transactions on Asian and Low-Resource Languages (TALLIP), IEEE Transactions on Big Data (TBD) and Computer, Speech and Language (CSL). He won the best paper awards of IALP 2017 and COLING 2018, best paper honorable mention of SemEval 2020, and best paper nomination for ACL 2018 and ACL 2023.



Title: Large Language Model for Social Safety
Speaker: Jing Ma @ HKBU

Abstract: A chaotic phenomenon characterized as the massive spread of toxic content (such as misinformation, harmful memes, etc) has become increasingly a daunting issue in human society. Recent advances in large language models (LLMs) and vision-language models (VLMs) offer transformative opportunities for enhancing social safety on digital platforms. This talk delves into innovative methods focusing on rumor detection, explainable fake news detection, harmful meme detection, and sarcasm detection. We explore LLM-based approaches for detecting textural rumors and fake news, highlighting how LLMs can flag misinformation and provide justifications behind the detection. Moving to multimodal challenges, we examine the detection of harmful memes and sarcasm tasks. VLMs can capture implicit clues by analyzing both visual and textual signals. This talk aims to provide insights for deploying advanced AI responsibly to address the growing challenges of safety issues.


Bio: Dr. Ma Jing is currently an Assistant Professor at Department of Computer Science, Hong Kong Baptist University. She received her PhD at The Chinese University of Hong Kong at 2020. She has long-term and strong research interests in Natural Language Processing, Social Network Analysis and Mining, Fact-Checking, Information Retrieval, Large Language Model and Vision Language Model. She has co-authored more than 50 publications in refereed journals and conferences, including ACL, WWW, EMNLP, IJCAI, CIKM, TKDE and TIST, with more than 5400 citations so far. She was recognised as one of “The 2022 Women in AI” by Aminer, “The World’s top 2% scientists” released by Stanford University, and her paper was selected as Top Five Outstanding TIST Articles. During 2018.12-2019.08, she was a visiting scholar at Nanyang Technological University, Singapore. During 2019.12-2020.02, she was a visiting scholar at Institute for Basic Science, South Korea. In recent years, she served as Area Chair for AACL 2023, NAACL 2024, ACL 2024, EMNLP 2024 and NLPCC 2024; Program Committee Member for WSDM 2023, WWW2021 2023, AAAI 2019-2021, ACL 2019, EMNLP 2019, etc, and was invited to review journal papers such as TIST, TKDE, TOMM, TNNLS, TPAMI, etc.



Title: Your LLM is Secretly a Fool and You Should Treat it Like One
Speaker: Tianyu Pang @ Sea AI

Abstract: In this talk, I will present our recent works on jailbreaking/cheating LLMs and multimodal LLMs (MLLMs). This involves a quick overview of adversarial attacks and shows how LLMs/MLLMs facilitate much more flexible attacking strategies. For examples, we show that a null model that always returns a constant output can achieve a 86.5% LC win rate on AlpacaEval 2.0; we could also jailbreak one million MLLM agents exponentially fast in, say, 5 minutes.


Bio: Dr. Tianyu Pang is a Senior Research Scientist at Sea AI Lab. He received Ph.D. and B.S. degrees from Tsinghua University. His research interests span the areas of machine learning, including Trustworthy AI and Generative Models. He has published over 40 papers on top-tier conferences and journals including ICML/NeurIPS/ICLR and CVPR/ICCV/ECCV/TPAMI. His published papers have received over 9,000 citations. He is a recipient of Microsoft Research Asia Fellowship (2020), Baidu Scholarship (2020), NVIDIA Pioneering Research Award (2018), Zhong Shimo Scholarship (2020), CAAI Outstanding Doctoral Dissertation Award (2023), WAIC Rising Star Award (2023), and World's Top 2% Scientists (2024).



Title: Auto-Arena: Towards Fully Automated LLM Evaluations
Speaker: Wenxuan Zhang @ DAMO

Abstract: As large language models (LLMs) rapidly evolve, the challenge of evaluating their capabilities becomes increasingly crucial. In this talk, I will discuss the paradigm shift in LLM evaluation, tracing its evolution from traditional static benchmark-based methods to the LLM-as-a-judge approach, and ultimately to the renowned Chatbot Arena platform based on human voting. Throughout this journey, we observe a trend towards automation in various components of the evaluation process. Building on this trend, I will introduce our innovative solution: the Auto-Arena for LLMs. This automated evaluation framework leverages LLM-based agents to streamline the entire assessment process, from generating questions and participating in debates to evaluating one another within a committee. Remarkably, the Auto-Arena produces results that exhibit state-of-the-art correlation with human preferences—all without human intervention. I will conclude by sharing interesting findings from this project and exploring potential future directions in the realm of automated LLM evaluation and LLM improvements.


Bio: Dr. Wenxuan Zhang is currently a research scientist at Alibaba DAMO Academy in Singapore. He received his Ph.D. degree from the Chinese University of Hong Kong and then joined Alibaba Singapore with the Ali Star award. His primary research areas are natural language processing (NLP) and trustworthy AI. His research aims to advance NLP models that are inclusive, supporting diverse languages and cultures through multilingual language models, while also trustworthy by improving the safety and robustness of the models. He has published over 40 papers in top-tier AI conferences and journals, including ICLR, NeurIPS, ACL, EMNLP, SIGIR, WWW, TOIS, and TKDE. He is the core tech lead of the SeaLLMs project (LLMs specialized for Southeast Asian languages), which has received significant community attention with over 200k downloads. He also regularly serves on the (senior) program committees of multiple leading conferences and journals.



Title: Train Large-Scale Language Model with High-Quality Data
Speaker: Taifeng Wang @ ByteDance

Abstract: In the rapidly evolving field of artificial intelligence, training large-scale language models has emerged as a crucial area of research and development. Today's talk focuses on the significance of training large scale language models with high-quality data. As language models continue to grow in size and complexity, the quality of the training data becomes paramount. High-quality data ensures more accurate and reliable language understanding and generation. It enables the model to capture nuanced language patterns, semantic relationships, and context. We will talk about the various techniques employed to train large-scale language models effectively.


Bio: Taifeng Wang is currently a principal researcher at ByteDance. He got his master's degree from the University of Science and Technology of China. He is an expert on AI Algorithms with over 20 years of R&D experience, who has served as Principal Researcher at Microsoft Research Asia, AI Director of the Intelligent Engine Department at Ant Financial, and Head of AI Algorithms at Biomap. His research spans the entire spectrum from natural language processing (NLP), graph learning, distributed machine learning, multimodal learning, and AI-driven biopharmaceuticals. He has served as a Senior Area Chair for ACL and is the brilliant mind behind LightGBM, famously known as a "Kaggle leaderboard weapon". Holding 17 Chinese patents and 20 U.S. patents, Taifeng's research has garnered over 16,000 citations on Google Scholar. His team is now working on building the Large Language foundation model for ByteDance.



Organizers

General Chair:

Jiaying Wu, National University of Singapore

Local Chairs:

Ambuj Mehrish, Singapore University of Technology and Design

Ming Shan Hee, Singapore University of Technology and Design

Gerard Christopher Yeo, National University of Singapore

Program & Invitation Chairs:

Yang Deng, Singapore Management University

Anh Tuan Luu, Nanyang Technological University

Industry Relations Chair:

Yixin Cao, Fudan University

Publicity Chairs:

Wenya Wang, Nanyang Technological University

Hao Fei, National University of Singapore

Advisory Committee:

Min-Yen Kan, National University of Singapore

Soujanya Poria, Singapore University of Technology and Design

Roy Lee, Singapore University of Technology and Design

Partners


Location

SSNLP 2024 will be held at the Mapletree Business City, Town Hall Auditorium, (10 Pasir Panjang Road, Singapore 117438.

Past SSNLP

Contact Us

Please feel free to reach out if you have any inquiries: Ambuj Mehrish, Ming Shan Hee and Gerard Christopher Yeo.