How to Build and Own Your AI Stack: Insights from GMI Cloud’s Virtual Event with Jordan Nanos from HPE

June 5, 2024

Why managing AI risk presents new challenges

Aliquet morbi justo auctor cursus auctor aliquam. Neque elit blandit et quis tortor vel ut lectus morbi. Amet mus nunc rhoncus sit sagittis pellentesque eleifend lobortis commodo vestibulum hendrerit proin varius lorem ultrices quam velit sed consequat duis. Lectus condimentum maecenas adipiscing massa neque erat porttitor in adipiscing aliquam auctor aliquam eu phasellus egestas lectus hendrerit sit malesuada tincidunt quisque volutpat aliquet vitae lorem odio feugiat lectus sem purus.

  • Lorem ipsum dolor sit amet consectetur lobortis pellentesque sit ullamcorpe.
  • Mauris aliquet faucibus iaculis vitae ullamco consectetur praesent luctus.
  • Posuere enim mi pharetra neque proin condimentum maecenas adipiscing.
  • Posuere enim mi pharetra neque proin nibh dolor amet vitae feugiat.

The difficult of using AI to improve risk management

Viverra mi ut nulla eu mattis in purus. Habitant donec mauris id consectetur. Tempus consequat ornare dui tortor feugiat cursus. Pellentesque massa molestie phasellus enim lobortis pellentesque sit ullamcorper purus. Elementum ante nunc quam pulvinar. Volutpat nibh dolor amet vitae feugiat varius augue justo elit. Vitae amet curabitur in sagittis arcu montes tortor. In enim pulvinar pharetra sagittis fermentum. Ultricies non eu faucibus praesent tristique dolor tellus bibendum. Cursus bibendum nunc enim.

Id suspendisse massa mauris amet volutpat adipiscing odio eu pellentesque tristique nisi.

How to bring AI into managing risk

Mattis quisque amet pharetra nisl congue nulla orci. Nibh commodo maecenas adipiscing adipiscing. Blandit ut odio urna arcu quam eleifend donec neque. Augue nisl arcu malesuada interdum risus lectus sed. Pulvinar aliquam morbi arcu commodo. Accumsan elementum elit vitae pellentesque sit. Nibh elementum morbi feugiat amet aliquet. Ultrices duis lobortis mauris nibh pellentesque mattis est maecenas. Tellus pellentesque vivamus massa purus arcu sagittis. Viverra consectetur praesent luctus faucibus phasellus integer fermentum mattis donec.

Pros and cons of using AI to manage risks

Commodo velit viverra neque aliquet tincidunt feugiat. Amet proin cras pharetra mauris leo. In vitae mattis sit fermentum. Maecenas nullam egestas lorem tincidunt eleifend est felis tincidunt. Etiam dictum consectetur blandit tortor vitae. Eget integer tortor in mattis velit ante purus ante.

  1. Vestibulum faucibus semper vitae imperdiet at eget sed diam ullamcorper vulputate.
  2. Quam mi proin libero morbi viverra ultrices odio sem felis mattis etiam faucibus morbi.
  3. Tincidunt ac eu aliquet turpis amet morbi at hendrerit donec pharetra tellus vel nec.
  4. Sollicitudin egestas sit bibendum malesuada pulvinar sit aliquet turpis lacus ultricies.
“Lacus donec arcu amet diam vestibulum nunc nulla malesuada velit curabitur mauris tempus nunc curabitur dignig pharetra metus consequat.”
Benefits and opportunities for risk managers applying AI

Commodo velit viverra neque aliquet tincidunt feugiat. Amet proin cras pharetra mauris leo. In vitae mattis sit fermentum. Maecenas nullam egestas lorem tincidunt eleifend est felis tincidunt. Etiam dictum consectetur blandit tortor vitae. Eget integer tortor in mattis velit ante purus ante.

In a recent GMI Cloud virtual event, industry experts shared pivotal insights about building and managing AI infrastructure. The event featured presentations and discussions from Alex Yeh, CEO of GMI Cloud; Jordan Nanos, Machine Learning Architect at Hewlett Packard Enterprise; and Yujing Qian, Software Engineer Manager at GMI Cloud. This article provides a detailed overview and summary of the key points discussed in the webinar.

The entire webinar can be watched on YouTube here:

Webinar Highlights

Vision and Strategy

Alex Yeh began by highlighting GMI Cloud’s vision to enable users to build AI applications effortlessly, similar to how Shopify democratizes e-commerce. “We want to empower anyone to build AI applications with one click,” Alex emphasized. He discussed the company’s goal to be the “AI TSMC,” supporting the entire stack of building AI applications on their cloud platform. This involves providing not only raw compute power but also all necessary tools and resources.

Infrastructure Focus

A key aspect of GMI Cloud’s strategy is controlling their hardware to ensure high availability, reliability, and optimized performance. Alex noted, “We control our nodes, which ensures that our customers always have the highest availability and reliability.”

Data Centers and Global Reach

Currently, GMI Cloud operates multiple data centers across the Asia-Pacific region, with plans to expand to 30 locations globally by the end of next year. These data centers are strategically placed in densely populated regions to minimize latency and optimize performance. Alex highlighted, “We have three data centers up and running and two more coming online by mid-August.”

Technical Capabilities

GMI Cloud provides access to top-tier GPU hardware, including the NVIDIA H100, and an in-house designed software stack that simplifies large-scale deployment. The company also offers a cluster engine layer, this includes multi-tenant Kubernetes for expert-level control and customization for container orchestration, essential for both training and inference workloads.

Service Models

To cater to different customer needs, GMI Cloud offers two major service models: On-Demand and Reserved. The On-Demand model is suitable for startups and researchers with unpredictable workloads, while the Reserved model is ideal for large enterprises with stable, long-term requirements. This flexible and predictable pricing structure ensures that various operational scenarios are efficiently managed.

Opening Presentation

Alex Yeh (CEO of GMI Cloud)

Alex emphasized GMI Cloud’s goal to support the entire stack of building AI applications, drawing on the company’s strong roots in the industry and the extensive experience of its core team from Google and OEM backgrounds in Taiwan. He stressed the importance of strategic data center locations, particularly in the Asia-Pacific region, to minimize latency and optimize performance. “Our goal is to have 30 data centers by the end of next year, providing the broadest serving GPU fleets across Asia and eventually expanding globally,” Alex explained.

Fireside Chat: Key Takeaways

During the fireside chat, Alex Yeh and Jordan Nanos delved into scalability and efficiency challenges, with Alex explaining the importance of infrastructure management. “We aim to provide a robust infrastructure that simplifies the complexity of managing AI systems,” he said. Jordan added, “It’s about ensuring reliability and performance through strategic hardware control.”

The discussion also covered data privacy and security. Jordan elaborated on the importance of securing data at multiple layers and leveraging the open-source community for continuous innovation while maintaining compliance. “The open-source ecosystem is vibrant and essential for AI advancement, but we must ensure data integrity and security,” he emphasized.

Jordan also discussed the challenges of managing AI infrastructure, emphasizing the complexity and costs involved. He highlighted the need for robust operations to ensure high uptime and reliability, saying, “Managing hardware is incredibly expensive and complex. Our goal is to simplify these steps for our customers.” Jordan also addressed security concerns, detailing the three layers of security: data privacy, model security, and application compliance. “Ensuring data privacy at multiple layers, from data ingestion to model deployment, is crucial,” he noted.

Looking to the future, Alex and Jordan discussed the short-term disruptions and long-term innovations expected in the AI industry. Alex mentioned, “The advertisement and commerce sectors will see significant changes through tailored AI solutions.” He also highlighted the potential for AI to revolutionize biotech, material science, and other fields, saying, “AI agents will support various enterprise functions, accelerating innovation in multiple industries.”

Alex emphasized GMI Cloud’s goal to support the entire stack of building AI applications, drawing on the company’s strong roots in the industry and the extensive experience of its core team from Google and OEM backgrounds in Taiwan. He stressed the importance of strategic data center locations, particularly in the Asia-Pacific region, to minimize latency and optimize performance. “Our goal is to have 30 data centers by the end of next year, providing the broadest serving GPU fleets across Asia and eventually expanding globally,” Alex explained.

Demo Session: Deploying Llama 3 on GMI Cloud

Yujing Qian’s demo session provided a practical demonstration of GMI Cloud’s platform capabilities. He showcased how the platform allows for the seamless deployment of Llama 3, highlighting its user-friendly interface and flexible container solutions. “Our platform’s flexibility enables quick setup and efficient AI model deployment,” Yujing demonstrated, emphasizing the robust performance of GMI Cloud’s GPUs in handling extensive AI workloads.

Conclusion

The GMI Cloud virtual event underscored the company’s commitment to empowering AI innovation through robust infrastructure, strategic hardware control, and a user-centric platform. By addressing key challenges in scalability, efficiency, and security, GMI Cloud positions itself as a leader in the AI infrastructure space, ready to support enterprises and developers in building the future of AI. A big thanks to Jordan Nanos and our partners at HPE for joining the event and providing valuable insights into the industry.

Stay tuned for more insights and updates from GMI Cloud. Follow us on LinkedIn and YouTube for the latest developments in the AI industry. Feel free to reach out to our sales team (sales@gmicloud.ai) with any questions or comments.

Get started today

Give GMI Cloud a try and see for yourself if it's a good fit for AI needs.

Get started
14-day trial
No long-term commits
No setup needed
On-demand GPUs

Starting at

$4.39/GPU-hour

$4.39/GPU-hour
Private Cloud

As low as

$2.50/GPU-hour

$2.50/GPU-hour