Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Deep Dive into Training and Inferencing Large Language Models on Azure (Presented by Microsoft)

      , Principal Software Engineer, Microsoft
      , Technical Program Manager 2, Microsoft
      In the coming years, the foundational models in artificial intelligence will continue to rapidly evolve towards multi-modality and targeted small language models in addition to improving the capabilities of existing large language models. As developers of AI, you need an infrastructure that can scale seamlessly and be flexible enough to cater to the demands of the AI model that you want to develop. In this session, we will explore the advancements we have made in accelerating training and inferencing of Large Language Models (LLMs) using Azure's cutting-edge infrastructure powered by NVIDIA H100 SXM Tensor Core GPUs and NVIDIA H100 NVL 94GB PCIe Tensor Core GPUs. We will dive into the performance, scaling, and readiness of our AI infrastructure and will extend our vision in deploying inferencing in a cost-effective way on Azure's NC H100 v5-series.
      活动: GTC 24
      日期: March 2024
      话题: AI Inference
      行业: All Industries
      级别: Intermediate Technical
      NVIDIA technology: NVLink / NVSwitch
      语言: English
      所在地: