Call for testing: Vulkan-inference support (#17) · Issues · 林博仁 Buo-ren Lin / Unofficial snap packaging for whisper.cpp · GitLab

1 min read Original article ↗
Call for testing: Vulkan-inference support
Hello, since revision 5b61d5e8 we now supports Vulkan-based inference/acceleration. This provides drastically better transcribe speed(~39x) than the poorly optimized CPU inference mode on supported systems: ![Screenshot depicting transcribe duration difference between Vulkan-vs-CPU inference modes, where CPU inference takes 99 minutes and Vulkan-based inference only taking about 2.5](/uploads/bb8a2c2a2113065e4a272b6327c732c1/cpu-vs-vulkan.png "Screenshot depicting transcribe duration difference between Vulkan-vs-CPU inference mode"){width=900 height=540} As this is a rather big change we would like to hold a call for testing to see whether there's any regressions we need to look into. Please install the build in the `beta` channel, run your usual audio transcribe workload, and made a reply or file a new issue when you encountered a problem(the following is the typical commands to do so): ```bash sudo snap install --channel=beta whisper-cpp whisper-cpp.download-ggml-model medium whisper-cpp.cli -m ggml-medium.bin --threads $(nproc) --language auto --print-progress --flash-attn --output-srt _input_file_ ``` Cheers! :champagne_glass: