Multi-GPU Inference with Accelerate bengubler.com 14 points by nebrelbug 3 years ago · 1 comment Reader PiP Save nebrelbugOP 3 years ago Quick tutorial on how to use Accelerate to run inference on LLMs in parallel