Skip to content
Menu
PyCon Hong Kong
  • Schedule
  • Sponsors
    • Supporting Organizations
    • Organizers
  • Volunteers
  • Code of Conduct
  • Ticket
  • About
    • PyCon HK
    • 2022
    • 2021
      • Privacy Policy Statement 2021
    • 2020 Fall
    • 2020 Spring
    • 2018
    • 2017
      • Photos
      • Videos
    • 2016
      • Photos
      • Videos
    • 2015
      • Photos
PyCon Hong Kong

[PyCon HK x COSCUP 2022 Online Special Track] Triton As NLP Model Inference Back-end

Posted on October 30, 2022November 4, 2022

If there is a Chatbot with massive users and uses your own NLP model, it might face a bottleneck that back-end service is not easy to handle such NLP concurrency. To solve this kind problem, Nvidia has introduced an open-source inference server, Triton, which allowed to deploy AI model on GPU or CPU. It will maximize utilization of GPU/CPU for faster inference.
In this session, I will introduce Triton Inference Server and deploy NLP model Triton with a practical sample.

Date and Time : November 5, 2022 / 14:15-14:45 ( UTC+8 ) 
Language : English
Speaker : Mr. Ko Ko / Chatbot Developers Taiwan / Taiwan

Speaker Introduction

Mr. Ko Ko

Ko Ko is a Microsoft AI MVP. He is dedicated to sharing AI and Chatbot related technology. He is a famous technical lecturer and got invited by many large conferences, such as COSCUP, .NET CONF, PyCon APAC and so on. And Ko Ko is also a core member in Chatbot Developers Taiwan.

Gold Sponsors

Silver Sponsors

Categories

  • 2022
  • 2021
  • 2020 Fall
  • 2020 Spring
  • 2018
  • Conference Highlights

Archives

©2023 PyCon Hong Kong | Powered by WordPress and Superb Themes!
← Ticket – PyCon Hong Kong 2022 ← [PyCon HK x COSCUP 2022 Online Special Track] Deep Learning with PyTorch: From Beginner to Research