NVIDIA GPU

LayerOps supports deploying services that require NVIDIA GPU capabilities. This feature allows you to run GPU-intensive services like Ollama or other AI/ML workloads that benefit from GPU acceleration.

Setting up GPU-enabled Instances

There are two ways to set up GPU-enabled instances in LayerOps:

  1. Attach an External GPU Instance

You can attach any external machine (VM, server, or computer) running Ubuntu that has an NVIDIA GPU. When attaching the instance to LayerOps, you will need to specify that it has NVIDIA GPU capabilities.

external with gpu

Make sure NVIDIA drivers are properly installed on the machine before attaching it to LayerOps.

apt install ubuntu-drivers-common && \
add-apt-repository ppa:graphics-drivers/ppa && \
apt install $(nvidia-detector)
  1. Deploy GPU-compatible Instances via Providers

You can create an instance pool with GPU capabilities through your configured cloud providers. Simply select a GPU-compatible instance type. This allows you to scale your GPU workloads automatically based on your needs.

instance type with gpu

Deploying Services with GPU Access

To enable GPU access for your service, follow these steps:

  1. Create a new service by clicking on the "Create new Service" button
  2. Fill out your service configuration form
  3. In the "Something more?" section at the end of the form, click on the "Devices" button
  4. Check the "Use NVIDIA GPU" option to enable GPU capabilities for your service

service with gpu

Services configured to use NVIDIA GPUs will only be deployed to instances that have GPU capabilities. Make sure you have appropriate GPU-enabled instances in your pool.