CS294-162 Machine Learning Systems (Fall 2024)
Course Logistics
- Lectures: Mon/Wed 10:30 am – 12:00 pm at Soda 310
- Office Hours: By appointment
- Communications: Slack (invite only)
Course Description
The recent success of AI has been at least partly driven by advances in hardware and software systems. These systems have enabled training increasingly complex models on ever larger datasets. In the process, these systems have also simplified model development, enabling the rapid growth in the machine learning community. These new hardware and software systems include a new generation of GPUs and hardware accelerators (e.g., TPU) as well as open source frameworks such as TensorFlow and PyTorch (and many others) and have shaped AI research and practice.
A fundamental hypothesis in AI-Systems research is that advances in systems will enable the continued scaling of models and data to unlock new AI capabilities. In the past few years, we have started to see significant evidence of this hypothesis with very large models unlock new AI capabilities especially in the context of text and image generation.
In this course, we will study the latest trends in systems designs to better support the next generation of AI applications. We will focus on advances in generative AI and specifically LLMs and how they have mirrored advances in computer systems for AI. We will cover the key pieces of work from the system and AI literature that have driven these advances and may reveal where research is headed next.
Course Format and Policies
TBD
Course Staff
Instructors
Seminar Coordinators