06 Jul

NVIDIA LaunchPad + NVIDIA AI Enterprise

What is NVIDIA LaunchPad

NVIDIA LaunchPad is a free program that provides users short-term access to a large catalog of hands-on labs. Enterprises and organizations can immediately tap into the necessary hardware and software stacks to experience end-to-end solution workflows in the areas of AI, data science, 3D design collaboration and simulation, and more.

What to Expect on NVIDIA LaunchPad

What is NVIDIA AI Enterprise

NVIDIA AI Enterprise Catalog gives you the entire access to an end-to-end, cloud-native, suite of AI and data analytics software, optimized and certified by NVIDIA. It’s certified to deploy anywhere—from the enterprise data center to the public cloud—and includes global enterprise support and training. It includes key enabling technologies and software from NVIDIA for rapid deployment, management, and scaling of AI workloads in the modern hybrid cloud.

NVIDIA AI Enterprise enables the following:

  1. Leverage fully integrated, optimized, certified, and supported software from NVIDIA for AI workloads.
  2. Run NVIDIA AI frameworks and tools optimized for GPU acceleration, reducing deployment time and ensuring reliable performance.
  3. Deploy anywhere – including on popular data center platforms from VMware and Red Hat, mainstream NVIDIA-Certified Systems configured with or without GPUs, and on GPU-accelerated instances in the public cloud.
  4. Leverage the jointly certified NVIDIA and Red Hat solution to deploy and manage AI workloads in containers or VMs with optimized software.
  5. Scale out to multiple nodes, enabling even the largest deep learning training models to run on the VMware vSphere. Previously, scaling with bare metal performance in a fully virtualized environment was limited to a single node, limiting the complexity and size of AI workloads that could be supported.
  6. Run AI workloads at near bare-metal performance with new optimizations for GPU acceleration on vSphere, including support for the latest Ampere architecture including the NVIDIA A100. Additionally, technologies like GPUDirect Communications can now be supported on vSphere. This provides communication between GPU memory and storage across a cluster for improved performance.

The NVIDIA AI Enterprise includes:

  • TensorFlow and Pytorch for machine learning
  • NVIDIA TAO Toolkit for a faster, easier way to accelerate training and quickly create highly accurate and performant, domain-specific vision, and conversational AI models
  • NVIDIA Tensor RT, for GPU optimized deep learning inference and Triton Inference Server to deploy trained AI models at scale
  • Triton Inference Server supports all major frameworks, such as TensorFlow, TensorRT, PyTorch, MXNet, Python and more. Triton Inference Server also includes the RAPIDS FIL backend for the best inference performance for tree-based models on GPUs.
  • NVIDIA RAPIDS, for end-to-end data science, machine learning and analytics pipeline
  • NVIDIA GPU and Network Operators, to deploy and manage NVIDIA GPU and Networking resources in Kubernetes
  • NVIDIA vGPU Software, to deploy vGPU on common data center platforms, including VMware and Red Hat

Deploying NVIDIA AI Enterprise on NVIDIA LaunchPad

Through the LaunchPad portal, NVIDIA provides a series of steps that you can follow to install the NVIDIA AI Enterprise software stack. For example, below demonstration is to “Train and Deploy an AI Support Chatbot” using TensorFlow, TensorRT and Triton Inference Server from the NVIDIA AI Enterprise Catalog.

  1. Connect to the VMware vCenter

2. Create Your First NVIDIA AI Enterprise VM and install the Operating System. We also added an NVIDIA Virtual GPU device to the VM to accelerate the workloads later on.

3. Installing Docker and Docker Utility Engine for NVIDIA GPUs

4. Installing Tensorflow to train the BERT model for conversational AI (NLP)

5. Installing Triton Inference Server to serve the trained BERT model

For more details and enquiries about NVIDIA Products, including NVIDIA Launchpad and NVIDIA AI Enterprise, please contact our sales team at:

[email protected]

20 Feb

OUR BREAKTHROUGH WORKING WITH NVIDIA CLARA ON BLOOD VESSELS

After much experimentation, VesselNet’s Keras model has been successfully converted into a Tensorflow-TensorRT optimized model via the following transformation:

  • A custom function called freeze_session() that freezes the graph session of a trained TF model and saved to a ProtoBuf (.pb) format:
  • Utilizing the Python API of TF-TRT, trt.TrtGraphConverter() method to convert the previous ProtoBuf format file into a TRT-Optimized frozen graph:

The coming week we’ll move towards uploading to AIAA server and run several tests on it with training data.

Commands to upload to AIAA Server:

Here are the samples of the testing datasets provided by the VesselNet’s author:

Summary

The VesselNet’s Keras model was successfully converted into a TRT-Optimized frozen graph that now can be uploaded to AIAA server to integrate with existing pipeline. Next, we’ll upload and run through the workflow with testing datasets.

10 Jun

JUNE 2019: NOVAGLOBAL VIRTUAL GPU TRAINING & DEVELOPMENT (vGTD) PLATFORM

Figure 1: AI + RV / Visualizing Deep Learning Results on vGTD Platform with Laptop

Novaglobal is delivering a vGPU Training & Development Platform solution based on NVIDIA Tesla T4, Virtual GPU (vGPU) and NVDOCKER supported with NVIDIA GPU CLOUD (NGC). The platform will focus on providing GPU readied solutions for:

  1. AI / Deep Learning
  2. HPC Applications
  3. 3D Remote Visualization
  4. Digital Rendering / VR

Key benefits of system:

  1. Open Source solution except for NVIDIA vGPU licences
  2. Support for up to 32 vGPU instances concurrently
  3. With vGPU template, VM can be built up to support DL teaching environment very quickly
  4. Support all major DL frameworks
  5. Support Windows & various flavours of Linux.
  6. Does not require VDI. GUI/RV is built-in with VM.
  7. Supported with DOCKER/NVDOCKER for fast deployment
  8. Integrated to NVIDIA GPU CLOUD (NGC)

Hardware Requirements:

  • ASUS 4U ESC8000 G4 GPU Server
  • Dual Intel CPU
  • 384GB System RAM
  • 4 x SSD
  • 8 x NVIDIA Tesla T4