IEEE VIS 2024 Content: Compress and Compare: Interactively Evaluating Efficiency and Behavior Across ML Model Compression Experiments

Compress and Compare: Interactively Evaluating Efficiency and Behavior Across ML Model Compression Experiments

Angie Boggust - Massachusetts Institute of Technology, Cambridge, United States

Venkatesh Sivaraman - Carnegie Mellon University, Pittsburgh, United States

Yannick Assogba - Apple, Cambridge, United States

Donghao Ren - Apple, Seattle, United States

Dominik Moritz - Apple, Pittsburgh, United States

Fred Hohman - Apple, Seattle, United States

Screen-reader Accessible PDF

Room: Bayshore V

2024-10-17T13:18:00ZGMT-0600Change your timezone on the schedule page
2024-10-17T13:18:00Z
Exemplar figure, described by caption below
Compress and Compare helps ML practitioners analyze and compare compression experiments. The Model Map helps practitioners understand what experiments were run and find high-performing sequences of operations, while the Model Scatterplot and Selection Details views help compare accuracy and efficiency metrics quantitatively. Our paper describes the challenges that Compress and Compare addresses, how we designed the system, and a study with eight experts demonstrating its potential to support compression workflows.
Fast forward
Full Video
Keywords

Efficient machine learning, model compression, visual analytics, model comparison

Abstract

To deploy machine learning models on-device, practitioners use compression algorithms to shrink and speed up models while maintaining their high-quality output. A critical aspect of compression in practice is model comparison, including tracking many compression experiments, identifying subtle changes in model behavior, and negotiating complex accuracy-efficiency trade-offs. However, existing compression tools poorly support comparison, leading to tedious and, sometimes, incomplete analyses spread across disjoint tools. To support real-world comparative workflows, we develop an interactive visual system called Compress and Compare. Within a single interface, Compress and Compare surfaces promising compression strategies by visualizing provenance relationships between compressed models and reveals compression-induced behavior changes by comparing models’ predictions, weights, and activations. We demonstrate how Compress and Compare supports common compression analysis tasks through two case studies, debugging failed compression on generative language models and identifying compression artifacts in image classification models. We further evaluate Compress and Compare in a user study with eight compression experts, illustrating its potential to provide structure to compression workflows, help practitioners build intuition about compression, and encourage thorough analysis of compression’s effect on model behavior. Through these evaluations, we identify compression-specific challenges that future visual analytics tools should consider and Compress and Compare visualizations that may generalize to broader model comparison tasks.