This Monday - Tianqi Chen (OctoAI & CMU) on “Enabling LLM Deployment Across Cloud and Edge with ML Compilation”

7 views
Skip to first unread message

Nadav Timor

unread,
Jan 9, 2025, 6:23:56 AMJan 9
to Faster LLM Inference Seminar

Date & Time:
Saturday, January 13, 2025, at 1:00 PM EST

Abstract:
In this talk, we will discuss the lessons learned in building an efficient large language model deployment system for both server and edge settings. We will cover general techniques in machine learning compilation and system support for efficient structure generation. We will also discuss the future opportunities in system co-design for cloud-edge model deployments.

Bio:
Tianqi Chen is currently an Assistant Professor at the Machine Learning Department and Computer Science Department of Carnegie Mellon University. He is also a distinguished engineer at NVIDIA. He received his PhD. from the Paul G. Allen School of Computer Science & Engineering at the University of Washington. He has created many major learning systems that are widely adopted: XGBoost, Apache TVM, and MLC-LLM.

Registration:
https://faster-llms.vercel.app

We look forward to seeing you there!

Reply all
Reply to author
Forward
0 new messages