by michaeloblak on 4/12/2023, 11:28:26 PM with 4 comments
Is it possible? Let's say Alice has 2 GPUs idle at the moment, Ben has 1 GPU, and Chris needs 3 GPUs for the next 12 hours. How to build such a system, and what problems there might occur? How to handle turning off one's machine? Does it even make sense to run training or inference on such (latency)?
Is it possible? Let's say Alice has 2 GPUs idle at the moment, Ben has 1 GPU, and Chris needs 3 GPUs for the next 12 hours. How to build such a system, and what problems there might occur? How to handle turning off one's machine? Does it even make sense to run training or inference on such (latency)?