Back to All Events

PAST: Squeeze more juice out of a single GPU in deep learning

WATCH HERE

Presenter: Weiguang Guan, SHARCNET

It’s well known that GPUs can significantly accelerate neural network training. However, not everyone knows that a single GPU is sufficient to train most neural networks except for a few large ones (like LLMs). In fact, a GPU is under-utilized in most cases. In this talk, we are addressing the under-utilization issue and proposing a way to make full use of the GPU capacity. The goal is to increase the throughput with a single GPU. We will use a small NN training as an example to illustrate how to achieve the goal by splitting a physical GPU into multiple logical GPUs and then running a particular training process per logical GPU.

Previous
Previous
November 8

PAST: Reference-Counted Multidimensional Arrays for C++ with rarray

Next
Next
November 22

PAST: Web scraping in Python