天元系列报告 | Optimizing LLMs and Optimizing with LLMs
报 告 人: 袁晓明 教授
所在单位: 香港大学
报告地点: 吉林大学正新楼209
报告时间: 2025-07-28 10:00:00
报告简介:

Large Language Models (LLMs) have brought about a profound revolution across various sectors, spanning from industry and economy to education and entertainment. While the advancement of LLMs has primarily relied on hardware such as GPUs and engineering technologies, the field is now entering a new phase that necessitates deeper engagement with scientific disciplines. The integration of mathematical and quantitative insights is crucial for driving the next wave of breakthroughs in the realm of LLMs. We will reevaluate the core tasks involved in the lifecycle of LLMs through an optimization lens, with a focus on enhancing the scientific understanding of key phases including pre-training, post-training, and serving. We will discuss specific tasks such as (distributed) low-bit training, pruning, quantization, prefill-decode disaggregation architecture, and the management of training chips in distributed/centralized settings. Our primary objective is to reduce the computation and memory footprint of LLMs throughout their lifecycle. We will also initiate two new concepts: Optimization Agents and Intelligence-Collaborative Computation.

上一篇 下一篇
主讲人简介:
袁晓明,香港大学数学系教授。主要研究领域为优化、最优控制、人工智能和云计算。Clarivate Analytics高被引学者。2024年香港裘搓基金会高级研究学者。2023年带领香港大学和华为云研究团队入围Franz Edelman Award决赛。