Instant GPUs, Infinite AI: GMI Cloud Launches On-Demand GPU Cloud Product

May 17, 2024

Why managing AI risk presents new challenges

Aliquet morbi justo auctor cursus auctor aliquam. Neque elit blandit et quis tortor vel ut lectus morbi. Amet mus nunc rhoncus sit sagittis pellentesque eleifend lobortis commodo vestibulum hendrerit proin varius lorem ultrices quam velit sed consequat duis. Lectus condimentum maecenas adipiscing massa neque erat porttitor in adipiscing aliquam auctor aliquam eu phasellus egestas lectus hendrerit sit malesuada tincidunt quisque volutpat aliquet vitae lorem odio feugiat lectus sem purus.

  • Lorem ipsum dolor sit amet consectetur lobortis pellentesque sit ullamcorpe.
  • Mauris aliquet faucibus iaculis vitae ullamco consectetur praesent luctus.
  • Posuere enim mi pharetra neque proin condimentum maecenas adipiscing.
  • Posuere enim mi pharetra neque proin nibh dolor amet vitae feugiat.

The difficult of using AI to improve risk management

Viverra mi ut nulla eu mattis in purus. Habitant donec mauris id consectetur. Tempus consequat ornare dui tortor feugiat cursus. Pellentesque massa molestie phasellus enim lobortis pellentesque sit ullamcorper purus. Elementum ante nunc quam pulvinar. Volutpat nibh dolor amet vitae feugiat varius augue justo elit. Vitae amet curabitur in sagittis arcu montes tortor. In enim pulvinar pharetra sagittis fermentum. Ultricies non eu faucibus praesent tristique dolor tellus bibendum. Cursus bibendum nunc enim.

Id suspendisse massa mauris amet volutpat adipiscing odio eu pellentesque tristique nisi.

How to bring AI into managing risk

Mattis quisque amet pharetra nisl congue nulla orci. Nibh commodo maecenas adipiscing adipiscing. Blandit ut odio urna arcu quam eleifend donec neque. Augue nisl arcu malesuada interdum risus lectus sed. Pulvinar aliquam morbi arcu commodo. Accumsan elementum elit vitae pellentesque sit. Nibh elementum morbi feugiat amet aliquet. Ultrices duis lobortis mauris nibh pellentesque mattis est maecenas. Tellus pellentesque vivamus massa purus arcu sagittis. Viverra consectetur praesent luctus faucibus phasellus integer fermentum mattis donec.

Pros and cons of using AI to manage risks

Commodo velit viverra neque aliquet tincidunt feugiat. Amet proin cras pharetra mauris leo. In vitae mattis sit fermentum. Maecenas nullam egestas lorem tincidunt eleifend est felis tincidunt. Etiam dictum consectetur blandit tortor vitae. Eget integer tortor in mattis velit ante purus ante.

  1. Vestibulum faucibus semper vitae imperdiet at eget sed diam ullamcorper vulputate.
  2. Quam mi proin libero morbi viverra ultrices odio sem felis mattis etiam faucibus morbi.
  3. Tincidunt ac eu aliquet turpis amet morbi at hendrerit donec pharetra tellus vel nec.
  4. Sollicitudin egestas sit bibendum malesuada pulvinar sit aliquet turpis lacus ultricies.
“Lacus donec arcu amet diam vestibulum nunc nulla malesuada velit curabitur mauris tempus nunc curabitur dignig pharetra metus consequat.”
Benefits and opportunities for risk managers applying AI

Commodo velit viverra neque aliquet tincidunt feugiat. Amet proin cras pharetra mauris leo. In vitae mattis sit fermentum. Maecenas nullam egestas lorem tincidunt eleifend est felis tincidunt. Etiam dictum consectetur blandit tortor vitae. Eget integer tortor in mattis velit ante purus ante.

As AI adoption accelerates across industries, companies are encountering unprecedented barriers in accessing the GPU resources necessary for innovation. High down payments, long contracts, and multi-month lead times have placed AI innovation just out of reach for many. But today, GMI Cloud is changing that landscape with the launch of its On-Demand GPU Cloud Product, providing instant, scalable, and affordable access to top-tier NVIDIA GPUs.

Versatile Optionality to Meet Global Demand for Compute:

The current surge in global demand for AI compute power requires companies to be strategic in their approach for accessing GPUs. In a fast-evolving landscape, organizations are being asked to pay a 25–50% down payment and sign up for a 3-year contract with the promise of gaining access to reserved GPU infrastructure in 6–12 months.

While certainly valuable for large-scale AI initiatives and projects such as foundation model training or ongoing inferencing, reserved bare-metal/private cloud solutions are not fit for all use cases. Certain businesses, especially startups, don’t always have the budget or long-term forecasting capabilities to commit to large GPU installations. They need flexibility to scale up or down based on application requirements. Similarly, enterprise data science teams often require agility to experiment, prototype, and evaluate AI applications quickly.

GMI Cloud On-Demand GPUs

GMI Cloud is dedicated to driving innovation by providing increased accessibility to top-tier GPU compute. Today we are launching an On-Demand GPU Cloud Product that offers a needed solution, allowing organizations to bypass long lead-times and access GPU resources without the need for long-term contracts. We’ve seen the frustration that companies feel from not being able to access GPUs in an effective manner. Accessibility is currently the primary roadblock to innovation for many companies – we built GMI Cloud On-Demand to eliminate this problem. The on-demand model is perfect for users who need instant short-term access to one or two instances to take on projects that demand high computational power like rapid prototyping or model fine-tuning. GMI Cloud On-Demand offers almost instantaneous access to NVIDIA H100 computing resources and gives additional optionality next to our reserved private cloud GPUs.

Benefits of GMI Cloud’s On-Demand Model

  • Added Flexibility: Scale GPU resources up or down almost instantaneously without long-term commitments or down payments.
  • Hassle-Free Deployment: Deploy AI models effortlessly with one-click container launches using our expertly pre-built docker image library. We reduce time and complexity in setting up environments, allowing your teams to focus on innovation rather than infrastructure.
  • Cloud-Native Orchestration: Manage and scale AI workloads seamlessly with NVIDIA software and Kubernetes integration, from control plane to management APIs. We provide scalability and flexibility, enabling your business to adapt quickly to changing demands without compromising on performance.

Technical Features and Benefits:

NVIDIA Software Stack Integration:

GMI Cloud’s On-Demand GPU Cloud Product includes a comprehensive NVIDIA software stack for seamless deployment and inference:

  • TensorRT: High-performance deep learning inference library optimized for NVIDIA GPUs. TensorRT accelerates the inference of models across different frameworks, significantly reducing latency for real-time applications.
  • NVIDIA Triton Inference Server: An open-source inference serving software that supports multiple frameworks, including TensorFlow, PyTorch, ONNX, and OpenVINO. Triton allows deployment of ensembles, dynamic batching, and model optimization for efficient inferencing.
  • NVIDIA NGC Containers: Access prebuilt NVIDIA GPU-optimized containers from the NGC catalog. Includes models and containers for vision, NLP, speech, and recommendation systems.

Kubernetes Orchestration:

GMI Cloud’s Kubernetes-managed platform offers scalable orchestration for ML workloads

  • Multi-Tenancy and Isolation: Kubernetes namespaces and resource quotas ensure secure isolation and efficient resource allocation.
  • Automatic Scaling: Horizontal Pod Autoscaling (HPA) dynamically adjusts the number of pod replicas based on workload demands.
  • GPU Resource Scheduling: Native support for NVIDIA GPUs via Kubernetes Device Plugins, ensuring optimal GPU utilization and scheduling.

Inference Model Deployment:

GMI Cloud’s On-Demand GPU Cloud Product simplifies the deployment and inferencing of various models:

  • LLaMA 3: Fine-tune and infer across different LLaMA 3 model sizes, ranging from 8B to 70B parameters.
  • Mixtral 8x7B: Deploy Mixtral, a multi-LLM ensemble designed for parallel inferencing.
  • Stable Diffusion: Efficiently generate high-quality images using Stable Diffusion’s state-of-the-art diffusion models.
  • Gemma 8x16B: Inference support for Google’s large-scale Gemma models, optimized for parallel inference serving.

On-Demand GPU Use Cases

Startups and Researchers:

  • Early-Stage Startups: Quickly prototype AI projects and scale GPU resources based on traction without the need for long-term contracts or large capital investments.
  • ML Researchers: Experiment with new models, algorithms, and techniques using flexible pay-as-you-go pricing, perfect for short-term or unpredictable workloads.
  • Fine-Tuning Specialists: Optimize and fine-tune models like LLaMA 3, Mixtral, and Gemma without the overhead of setting up private infrastructure.

Enterprise Data Science Teams:

  • Data Scientists and Analysts: Prototype, evaluate, and scale AI applications with almost instantaneous GPU access, enabling agile experimentation and testing.
  • AI Teams with Tight Deadlines: Accelerate model training and inference while avoiding delays from multi-month lead times and long-term commitments.
  • Private Cloud Complement: Use On-Demand instances to supplement existing private cloud infrastructure, offering overflow capacity for burst workloads.

ML Practitioners and DevOps Engineers:

  • ML Engineers: Efficiently deploy and infer models like Stable Diffusion, Mixtral, and Triton with preconfigured NVIDIA software stack environments.
  • DevOps Teams: Leverage Kubernetes orchestration with GPU scheduling, namespace isolation, and automatic scaling to streamline ML workflows.
  • Model Deployment Specialists: Seamless integration with NVIDIA Triton, TensorRT, and NGC containers ensures hassle-free inferencing across various AI models.

Getting Started:

GMI Cloud offers competitive pricing at $4.39/hour for on-demand access to NVIDIA H100 GPUs for 14-days. Visit gmicloud.ai to access our On-Demand GPU Cloud and unlock unlimited AI potential.

Visit GMI Cloud’s booth at Computex in Taiwan in June for hands-on demonstrations of our On-Demand GPU Cloud Product and other innovative AI solutions.

Get started today

Give GMI Cloud a try and see for yourself if it's a good fit for AI needs.

Get started
14-day trial
No long-term commits
No setup needed
On-demand GPUs

Starting at

$4.39/GPU-hour

$4.39/GPU-hour
Private Cloud

As low as

$2.50/GPU-hour

$2.50/GPU-hour