Skip to content

Self-Service Fractional GPUs with Rafay GPU PaaS

Enterprises and GPU Cloud providers are rapidly evolving toward a self-service model for developers and data scientists. They want to provide instant access to high-performance compute — especially GPUs — while keeping utilization high and costs under control.

Rafay GPU PaaS enables enterprises and GPU Clouds to achieve exactly that: developers and data scientists can spin up resources such as Developer Pods or Jupyter Notebooks backed by fractional GPUs, directly from an intuitive self-service interface.

This is Part-1 in a multi-part series on end user, self service access to Fractional GPU based AI/ML resources.

Fractional GPU


Simplifying Access to Fractional GPUs

Traditionally, provisioning GPU resources for experimentation or model training meant filing a ticket and waiting for IT or platform teams to allocate hardware. That process doesn’t work in modern, agile AI environments.

When GPUs are allocated as fractions, multiple workloads can share a single physical GPU safely and efficiently. With Rafay GPU PaaS, these fractional GPU classes are abstracted into simple drop-down options such as 0.25 GPU or 0.5 GPU,” making them accessible to any developer without needing to understand underlying GPU profiles, MIG instances, or time-slicing policies. These users also do not require any form of privileged and administrative access to the underlying infrastructure.


Developer Pods: A Full Linux Workspace in Seconds

A developer simply fills in a few intuitive fields:

  1. Name and Description
  2. CPU (e.g. 500 Mi) and Memory (e.g. 512 MiB)
  3. GPU Fraction (e.g. 0.25)
  4. Base images (e.g. Ubuntu 24.04, preloaded with CUDA, Python)

Based on the user's selection, Rafay dynamically calculates and provides a real-time cost estimate. This provides users with immediate visibility and accountability — i.e. users know what their environment costs before launching it.

Developer Pod

With one click, Rafay provisions the pod, applying all necessary Kubernetes and KAI Scheduler logic behind the scenes:

  • KAI assigns a fractional GPU slice (e.g., 25% of an A100 or L40S GPU).
  • Rafay enforces quotas, scheduling policies, and isolation.

Developers connect to the remote Ubuntu instance via SSH.

Developer Pod Success

From the user’s perspective, it feels like getting a dedicated Linux server with a full GPU. In reality, it’s a fraction of a GPU, allocated dynamically and efficiently through KAI. In the example below, we have accessed the remote Ubuntu image via SSH and have run the nvidia-smi command.

Developer Pod Access


Jupyter Notebooks with Fractional GPUs

For data scientists, Rafay provides an equally streamlined workflow for self service access to Jupyter notebooks with fractional GPUs. Users provide a name and specify resources interactively. For example:

  1. CPU: e.g., 1000m ($0.25/hour)
  2. Memory: e.g., 4Gi ($0.80/hour)
  3. GPU Fraction: e.g., 0.25 ($0.50/hour)
  4. Base Images (e.g. Minimal, PyTorch, Tensorflow etc)

They can select a Notebook Profile (such as Minimal Environment) preconfigured with Jupyter, TensorFlow, PyTorch, or RAPIDS. Rafay will dynamically calculate and present a cost estimate. This clarity allows teams to track compute costs in real time while experimenting freely.

Notebook

Once the notebook is launched, Rafay does the following automatically:

  • Allocate a fractional GPU to the notebook pod.
  • Provisions ingress, TLS certificates, and authentication automatically.

Notebook URL

Once the user clicks on the URL for the Jupyter notebook and authenticates using the token, they land in a secure Jupyter environment ready to run GPU-accelerated code. This provides all the convenience of managed notebook services — without vendor lock-in and with full control over costs, policy, and compliance.

In the image below, we have launched the "Terminal" in the Jupyter notebook and typed the "nvidia-smi" command to check what kind of GPU resources are available.

Notebook Nvidia GPU


Benefits

The benefits for users are apparent. This experience equals rapid iteration — running experiments, tuning models, and testing code in minutes rather than days.

Instant Access

Users do not have to wait on approvals or infrastructure provisioning.

Right-Sized Resources

Developers choose exactly the fraction of GPU needed for their workload.

Predictable Costing

Transparent per-hour and monthly estimates displayed in the UI. Integrated usage tracking enabling platform teams and GPU Cloud providers to bill users.

Familiar Environments

Users can choose from list of prebuilt and ready to use ML frameworks making them immediately productive.


Behind the Scenes

Rafay GPU PaaS integrates with multiple frameworks/tooling behind the scenes. Users are provided with same seamless, self service experience immaterial of the underlying technology used. Some common options and key considerations are described below.

Rafay Integrations


NVIDIA MIG

Rafay integrates with MIG. Hardware-based fractioning technology like NVIDIA’s Multi-Instance GPU (MIG) delivers improved isolation through dedicated fractional GPU partitions.

MIG Tech Approach

Key Considerations

  1. MIG requires static partitioning of the GPU with pre-defined profiles which admins may find restrictive.
  2. MIG is supported only on advanced and expensive datacenter class GPU models.

NVIDIA KAI Scheduler

Rafay integrates with KAI Scheduler for scenarios where MIG is not supported on the GPUs.

Key Considerations

KAI leverages “time slicing” to implement sharing. With KAI, you can implement various types of sharing: "Fractional GPUs", "Fractional GPU Memory", "Queues" etc.

KAI Tech Approach

Although it offers flexible GPU sharing, it lacks memory isolation. For example, in the notebook example above, although we requested a fraction of the GPU for the pod, we can actually see all 23GB GPU memory.

Notebook Nvidia GPU Memory

As documented, "In order to make sure the pods share the GPU device nicely, it is important that the running processes will allocate GPU memory up to the requested amount and not beyond that".


The New Normal for Seamless GPU Access

With Rafay GPU PaaS, both enterprises and GPU Clouds can transform their GPU clouds into fully self-service, multi-tenant environments. Users no longer need to understand MIG profiles or GPU topology — they simply choose a fractional GPU size and launch. Whether running a Developer Pod for code exploration or a Jupyter Notebook for model experimentation, users get:

  • GPU-backed environments in minutes,
  • Fractional GPU allocation at scale,
  • Simplified developer access via intuitive UI
  • Visibility into costs,
  • Scalable, shareable infrastructure that maximizes ROI.

The result: a GPU cloud where developers and data scientists move faster — while every GPU dollar works harder.

In the next blog, we will update this to provide users with the ability to select fractional GPU memory via self service. So, instead of getting a fraction such as 25% of a GPU, they can select exactly how much GPU memory should be allocated for their resource.