Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Optimizing Inference Performance and Incorporating New LLM Features in Desktops and Workstations

      , Deep Learning Solution Architect, NVIDIA
      , Product Manager, NVIDIA
      TensorRT has become the preferred choice for independent software vendor applications used in desktop and workstation environments, including those developed by Topaz, BlackMagic, and others. As these applications adapt to embrace the emerging generative AI trend, they seek to incorporate more features driven by large language models (LLMs) and stable diffusion techniques. We'll describe the journey as developer on how to apply TensorRT optimizations to achieve the speed-of-light inference performance, and share best practices. We'll also tell stories of how NVIDIA and partners worked together to come up with new features and improvements to support the release of TensorRT.
      活动: GTC 24
      日期: March 2024
      话题: AI Inference
      行业: All Industries
      级别: Intermediate Technical
      NVIDIA technology: TensorRT
      语言: English
      所在地: