Department of Mathematics - Seminar on Data Science - AI Safety by the People, for the People

11:00am - 12:00pm
Room 1103 (near Lift 19)

Supporting the below United Nations Sustainable Development Goals:支持以下聯合國可持續發展目標:支持以下联合国可持续发展目标:

The rapid advancement in large language models has heightened the importance of AI safety, a concept that encompasses two critical dimensions: ensuring safety 'by the people' and 'for the people'. 'By the people' focuses on the responsibility of developers to use legitimate training data. 'For the people' emphasizes the alignment of AI models with core human values — helpfulness, harmlessness, and honesty

 

In the first part of tha talk, we will delve into strategies for addressing concerns regarding the legitimate use of data in AI training. A case study from March 2023 highlights OpenAI's challenge: demonstrating the legitimacy of their training data and logic while maintaining the confidentiality of ChatGPT's weights and data. We introduce zkDL (Zero-Knowledge Deep Learning), an innovative solution offering efficient zero-knowledge proofs for deep learning. This technology allows for the creation of proofs in a second for neural networks with 20M parameters, achieving a 2000× speedup on an NVIDIA's A100 GPU.

 

The second part of the talk explores how integrating self-evaluation and rewind mechanisms in unaligned large language models (LLMs) can produce outputs that resonate with human preferences through self-boosting. We present the Rewindable Auto-regressive INference (RAIN) framework, enabling pre-trained LLMs to assess their own outputs and utilize these evaluations to inform and refine their response generation. This innovative approach is notable for its ability to enhance AI safety without the need for additional alignment data, training, gradient computations, or parameter updates.

講者/ 表演者:
Prof. Hongyang ZHANG
University of Waterloo
語言
英文
適合對象
校友
教職員
研究生
本科生
主辦單位
數學系
新增活動
請各校內團體將活動發布至大學活動日曆。