Skip to content
Menu
PyCon HK
  • Schedule
    • 2023
    • 2022
    • 2021
    • 2020 Fall
    • 2020 Spring
    • 2018
    • 2017
    • 2016
    • 2015
  • Sponsors
    • 2023
    • 2022
    • 2021
    • 2020 Fall
    • 2018
    • 2017
    • 2016
    • 2015
  • Organizers
    • Organizers
    • Volunteers
    • Booths
  • Supporting Organizations
  • Code of Conduct
    • Procedures for Reporting Incidents
    • Enforcement Procedures
  • About
    • PyCon HK
    • Conference Highlights
    • 2023
    • 2022
    • 2021
    • 2020 Fall
    • 2020 Spring
    • 2018
    • 2017
      • Photos
      • Videos
    • 2016
      • Photos
      • Videos
    • 2015
      • Photos
PyCon HK

[PyCon HK x COSCUP 2022 Online Special Track] Triton As NLP Model Inference Back-end

Posted on October 30, 2022November 4, 2022

If there is a Chatbot with massive users and uses your own NLP model, it might face a bottleneck that back-end service is not easy to handle such NLP concurrency. To solve this kind problem, Nvidia has introduced an open-source inference server, Triton, which allowed to deploy AI model on GPU or CPU. It will maximize utilization of GPU/CPU for faster inference.
In this session, I will introduce Triton Inference Server and deploy NLP model Triton with a practical sample.

Date and Time : November 5, 2022 / 14:15-14:45 ( UTC+8 ) 
Language : English
Speaker : Mr. Ko Ko / Chatbot Developers Taiwan / Taiwan

Speaker Introduction

Mr. Ko Ko

Ko Ko is a Microsoft AI MVP. He is dedicated to sharing AI and Chatbot related technology. He is a famous technical lecturer and got invited by many large conferences, such as COSCUP, .NET CONF, PyCon APAC and so on. And Ko Ko is also a core member in Chatbot Developers Taiwan.

  • Instagram
  • LinkedIn
  • Facebook
  • Twitter
  • YouTube

Archives

©2025 PyCon HK | Powered by SuperbThemes!
← Ticket – PyCon Hong Kong 2022 ← [PyCon HK x COSCUP 2022 Online Special Track] Deep Learning with PyTorch: From Beginner to Research