Rafay Releases New PaaS Capabilities for GPU-based Workloads
May 15, 2024

Rafay Systems has extended the capabilities of its enterprise PaaS for modern infrastructure to support graphics processing unit- (GPU-) based workloads.

This makes compute resources for AI instantly consumable by developers and data scientists with the enterprise-grade guardrails Rafay customers leverage today. The company also launched a new AI Suite with standards-based pipelines for machine learning operations (MLOps) and large language model operations (LLMOps) to help enterprise platform teams quicken the development and deployment of AI applications for developers and data scientists.

Rafay’s enterprise customers have long leveraged Rafay’s PaaS for modern infrastructure to rapidly give developers access to central processing unit- (CPU-) based infrastructure on-premises and in all the major public clouds, with guardrails included. The same issues that needed to be addressed for CPU-based workloads — environment standardization, self-service consumption of compute, secure use of multi-tenant environments, cost optimization, zero-trust connectivity enforcement and auditability — now have to be addressed with GPU-based workloads. Aspects such as cost are even more critical to control in the new age of AI.

In addition to applying its existing capabilities to GPU-based workloads, Rafay has extended its enterprise PaaS with features and capabilities that specifically support GPU workloads and infrastructure. Rafay makes AI-focused compute resources instantly consumable by developers and data scientists, enabling customers to empower every developer and data scientist to accelerate the speed of AI-driven innovation — and do it within the guidelines and policies set forth by the enterprise.

Haseeb Budhani, co-founder and CEO of Rafay Systems, said, “Beyond the multi-cluster matchmaking capabilities and other powerful PaaS features that deliver a self-service compute consumption experience for developers and data scientists, platform teams can also make users more productive with turnkey MLOps and LLMOps capabilities available on the Rafay platform. This announcement makes Rafay a must-have partner for enterprises, as well as GPU and sovereign cloud operators, looking to speed up modern application delivery.”

To address challenges associated with building and deploying AI-based applications, Rafay’s newly added support for GPU workloads helps enterprises and managed service providers power a new GPU-as-a-Service experience for internal developers and customers, respectively. This provides developers and data scientists with:

- Developer and Data Scientist Self-service: Easy to use, self-service experience to request for GPU-enabled workspaces

- AI-optimized User Workspaces: Pre-configured workspaces for AI model development, training and servicing with necessary AI tools including Jupyter Notebooks and Virtual Studio Code (VSCode) internal developer environment (IDE) integrations

- GPU Matchmaking: Similarly for CPUs, dynamically match the user workspaces with available GPUs or pools of GPUs based on criteria such as proximity, cost efficiency, GPU type and more to improve utilization

- GPU Virtualization: Time slicing and multi-instance GPU sharing to virtualized GPUs across workloads and lower the costs of running GPU hardware with dashboards to visualize GPU usage

Rafay’s new AI Suite adds to Rafay’s existing portfolio of suites, which consists of the company's Standardization Suite, Public Cloud Suite, and Private Cloud Suite. New capabilities include:

- Pre-configured LLMOps Playgrounds: Help developers experiment with generative AI (GenAI) by rapidly training, tuning and testing GenAI apps with approved models, vector databases, inference servers and more

- Turnkey MLOps Pipeline: Deliver an enhanced developer experience with an all-in-one MLOps pipeline, complete with GPU support, a company-wide model registry, and integrations with Jupyter Notebooks and VSCode IDEs

- Central Management of LLM Providers and Prompts: Built-in prompt compliance, cost controls on public LLM use such as OpenAI and Anthropic to ensure developers consistently comply with internal policies

- AI Data Source Integrations and Governance: Leverage pre-configured integrations with enterprise data sources such as Databricks and Snowflake while controlling usage for AI application development and deployments

Rafay’s newly added support for GPU workloads also expands and enhances the solutions the company jointly brings to market with global partners such as NTT DATA.

NTT DATA helps grow and transform its clients’ businesses, and it is now able to do so faster than ever before, thanks to Rafay’s new AI capabilities. This collaboration underscores NTT DATA's commitment to staying at the forefront of technological advancement, ultimately enabling NTT DATA to drive greater value for clients and achieve its mission objectives with agility and precision.

"Building on our successful partnership with Rafay, NTT DATA is proud to deepen its collaboration with the team to help deliver its new GPU support and AI Suite to the global market. This expanded alliance underscores our shared commitment to provide cutting-edge AI solutions that empower businesses worldwide. Rafay’s approach satisfies users responsible for application development and management, making it easy to cross-collaborate within enterprises’ security and budget boundaries,” said Mike Jones, vice president of partners and alliances at NTT DATA UK&I.

The new GPU-based capabilities in Rafay’s PaaS, along with the AI Suite are now generally available for customers.

Share this

Industry News

November 21, 2024

Red Hat announced the general availability of Red Hat Enterprise Linux 9.5, the latest version of the enterprise Linux platform.

November 21, 2024

Securiti announced a new solution - Security for AI Copilots in SaaS apps.

November 20, 2024

Spectro Cloud completed a $75 million Series C funding round led by Growth Equity at Goldman Sachs Alternatives with participation from existing Spectro Cloud investors.

November 20, 2024

The Cloud Native Computing Foundation® (CNCF®), which builds sustainable ecosystems for cloud native software, has announced significant momentum around cloud native training and certifications with the addition of three new project-centric certifications and a series of new Platform Engineering-specific certifications:

November 20, 2024

Red Hat announced the latest version of Red Hat OpenShift AI, its artificial intelligence (AI) and machine learning (ML) platform built on Red Hat OpenShift that enables enterprises to create and deliver AI-enabled applications at scale across the hybrid cloud.

November 20, 2024

Salesforce announced agentic lifecycle management tools to automate Agentforce testing, prototype agents in secure Sandbox environments, and transparently manage usage at scale.

November 19, 2024

OpenText™ unveiled Cloud Editions (CE) 24.4, presenting a suite of transformative advancements in Business Cloud, AI, and Technology to empower the future of AI-driven knowledge work.

November 19, 2024

Red Hat announced new capabilities and enhancements for Red Hat Developer Hub, Red Hat’s enterprise-grade developer portal based on the Backstage project.

November 19, 2024

Pegasystems announced the availability of new AI-driven legacy discovery capabilities in Pega GenAI Blueprint™ to accelerate the daunting task of modernizing legacy systems that hold organizations back.

November 19, 2024

Tricentis launched enhanced cloud capabilities for its flagship solution, Tricentis Tosca, bringing enterprise-ready end-to-end test automation to the cloud.

November 19, 2024

Rafay Systems announced new platform advancements that help enterprises and GPU cloud providers deliver developer-friendly consumption workflows for GPU infrastructure.

November 19, 2024

Apiiro introduced Code-to-Runtime, a new capability using Apiiro’s deep code analysis (DCA) technology to map software architecture and trace all types of software components including APIs, open source software (OSS), and containers to code owners while enriching it with business impact.

November 19, 2024

Zesty announced the launch of Kompass, its automated Kubernetes optimization platform.

November 18, 2024

MacStadium announced the launch of Orka Engine, the latest addition to its Orka product line.