Run llama-bench over various GPUs using the vastai CLI
Vastai has multiple GPUs, including older architechtures like Volta and Pascal. This makes it ideal for llama.cpp development as one of the project's goals is broad hardware support.
- Copy your patch ( for eg.
git diff upstream/master > patch.diff) into the local folder - Modify
setup_script.shto run the appropriatellama-benchcommand - Run command eg.
python3 vastai_benchmark.py --instance RTX_4090
- [] Add ability to specify multiple GPUs
- [] Add cost for a particular test
- Sometimes, the automatic instance finding takes a long time. You can also supply an existing instance via
--instance-id