Help! Hope to get an inference configuration that can run on multiple GPUs.

#25
by Lokis - opened

I have an 8*A100 80Gb server, but I've been testing for a long time and can't use the multi-GPU configuration stably. When using Auto, the output is very slow. Are there any documents or open-source configuration files that I can learn from?

Sign up or log in to comment