Groups
Groups
Sign in
Groups
Groups
Faster LLM Inference Seminar
Conversations
About
Send feedback
Help
Faster LLM Inference Seminar
Contact owners and managers
1–11 of 11
Mark all as read
Report group
0 selected
Nadav Timor
Jan 12
This Wednesday: Dan Alistarh (IST Austria & Red Hat) on “Massive Models in Low Precision”
Date & Time: Wednesday, Jan 14, at 3:00 PM EST (Add to calendar) Title: Massive Models in Low
unread,
This Wednesday: Dan Alistarh (IST Austria & Red Hat) on “Massive Models in Low Precision”
Date & Time: Wednesday, Jan 14, at 3:00 PM EST (Add to calendar) Title: Massive Models in Low
Jan 12
Nadav Timor
7/22/25
Tomorrow: Junchen Jiang (UChicago) on LMCache
Date & Time: Tomorrow, July 23, at 3:00 PM EST (Add to calendar) Title: Next-Gen Long-Context LLM
unread,
Tomorrow: Junchen Jiang (UChicago) on LMCache
Date & Time: Tomorrow, July 23, at 3:00 PM EST (Add to calendar) Title: Next-Gen Long-Context LLM
7/22/25
Nadav Timor
4/10/25
Starting in <45min: “Optimizing attention for modern hardware” - Tri Dao (Princeton & Together AI)
Date & Time: Today, April 10, at 12:00 PM EST (Add to calendar) Abstract: Attention, as a core
unread,
Starting in <45min: “Optimizing attention for modern hardware” - Tri Dao (Princeton & Together AI)
Date & Time: Today, April 10, at 12:00 PM EST (Add to calendar) Abstract: Attention, as a core
4/10/25
Faster LLM Inference Seminar
3/17/25
On Thursday: “Adaptive Compute LLMs with Early Exits” - Tal Schuster (Google DeepMind)
Date & Time: Thursday, March 20, at 3:00 PM EST (Add to calendar) Abstract: Scaling LLMs is a
unread,
On Thursday: “Adaptive Compute LLMs with Early Exits” - Tal Schuster (Google DeepMind)
Date & Time: Thursday, March 20, at 3:00 PM EST (Add to calendar) Abstract: Scaling LLMs is a
3/17/25
Faster LLM Inference Seminar
3/4/25
Today: “Accelerating LLM Inference with vLLM (and SGLang)” - Ion Stoica (Berkeley & Anyscale & Databricks)
Date & Time: Today at 3:30 PM EST (Add to calendar) Abstract: Inference efficiency remains a
unread,
Today: “Accelerating LLM Inference with vLLM (and SGLang)” - Ion Stoica (Berkeley & Anyscale & Databricks)
Date & Time: Today at 3:30 PM EST (Add to calendar) Abstract: Inference efficiency remains a
3/4/25
Nadav Timor
2/12/25
Today at 3pm ET - Hao Zhang (UCSD & Snowflake) on “Efficiently Serving Reasoning Programs with Certaindex”
Date & Time: Today at 3:00 PM EST (Add to calendar) Abstract: The rapid evolution of large
unread,
Today at 3pm ET - Hao Zhang (UCSD & Snowflake) on “Efficiently Serving Reasoning Programs with Certaindex”
Date & Time: Today at 3:00 PM EST (Add to calendar) Abstract: The rapid evolution of large
2/12/25
Nadav Timor
1/9/25
This Monday - Tianqi Chen (OctoAI & CMU) on “Enabling LLM Deployment Across Cloud and Edge with ML Compilation”
Date & Time: Saturday, January 13, 2025, at 1:00 PM EST Abstract: In this talk, we will discuss
unread,
This Monday - Tianqi Chen (OctoAI & CMU) on “Enabling LLM Deployment Across Cloud and Edge with ML Compilation”
Date & Time: Saturday, January 13, 2025, at 1:00 PM EST Abstract: In this talk, we will discuss
1/9/25
Faster LLM Inference Seminar
11/22/24
This Monday - Hongyang Zhang (Waterloo & Vector Institute) on "EAGLE & EAGLE-2: Lossless Inference Acceleration for LLMs"
Registration: https://faster-llms.vercel.app See you then! Nadav
unread,
This Monday - Hongyang Zhang (Waterloo & Vector Institute) on "EAGLE & EAGLE-2: Lossless Inference Acceleration for LLMs"
Registration: https://faster-llms.vercel.app See you then! Nadav
11/22/24
Faster LLM Inference Seminar
11/14/24
Starting in 3 hrs - Ce Zhang (UChicago & Together AI) on "The Token/s Game and Beyond"
Registration: https://faster-llms.vercel.app See you soon! Nadav
unread,
Starting in 3 hrs - Ce Zhang (UChicago & Together AI) on "The Token/s Game and Beyond"
Registration: https://faster-llms.vercel.app See you soon! Nadav
11/14/24
Nadav Timor
9/23/24
Starting in 10 min - Sasha Rush (Cornell & Hugging Face) on "SSMs and The Foundation Model Design Space"
Registration: https://faster-llms.vercel.app/ See you soon! Nadav
unread,
Starting in 10 min - Sasha Rush (Cornell & Hugging Face) on "SSMs and The Foundation Model Design Space"
Registration: https://faster-llms.vercel.app/ See you soon! Nadav
9/23/24
Nadav Timor
8/28/24
Today at noon ET - Xupeng Miao (Purdue University) - "Towards Fast and Affordable Serving Systems for LLMs"
Hi all, Happy to invite you to our session featuring Xupeng Miao from Purdue University. Please
unread,
Today at noon ET - Xupeng Miao (Purdue University) - "Towards Fast and Affordable Serving Systems for LLMs"
Hi all, Happy to invite you to our session featuring Xupeng Miao from Purdue University. Please
8/28/24