Computer Science Seminar: Tim Dettmers
Description
Tim Dettmers, a doctoral candidate at the University of Washington, will give a talk titled "Accessible Foundation Models: Systems, Algorithms, and Science" for the Department of Computer Science.
Abstract:
The ever-increasing scale of foundation models, such as ChatGPT and AlphaFold, has revolutionized artificial intelligence (AI) and science more generally. However, increasing scale also steadily raises computational barriers, blocking almost everyone from studying, adapting, or otherwise using these models for anything beyond static API queries. In this talk, Tim Dettmers will present research that significantly lowers these barriers for a wide range of use cases, including inference algorithms that are used to make predictions after training, fine-tuning approaches that adapt a trained model to new data, and finally, full training of foundation models from scratch. For inference, he will describe the LLM.int8() algorithm, which showed how to enable high-precision 8-bit matrix multiplication that is both fast and memory efficient. LLM.int8() is based on the discovery and characterization of sparse outlier sub-networks that only emerge at large model scales, but are crucial for effective Int8 quantization. For fine-tuning, he will introduce the QLoRA algorithm, which pushes such quantization much further to unlock fine-tuning of very large models on a single GPU by only updating a small set of the parameters while keeping most of the network in a new information-theoretically optimal 4-bit representation. For full training, he will present SWARM parallelism, which allows collaborative training of foundation models across continents on standard internet infrastructure while still being 80% as effective as the prohibitively expensive supercomputers that are currently used. Finally, he will close by outlining his plans to make foundation models 100x more accessible, which will be needed to maintain truly open AI-based scientific innovation as models continue to scale.
Who can attend?
- Faculty
- Staff
- Students