AI TRACK: Running OpenAI's Whisper Automatic Speech Recognition on a Live Video Transcoding Server


Event Time

Originally Aired - Monday, April 15   |   10:15 AM - 10:45 AM PT

Event Location

Pass Required: Streaming Summit Pass

Don't have this pass? Register Now!

Info Alert

Create or Log in to myNAB Show to see Videos and Resources.

Videos

Resources

{{video.title}}

Log in to your myNAB Show to join the zoom meeting!

Resources

Info Alert

This Session Has Not Started Yet

Be sure to come back after the session starts to have access to session resources.

As video streaming workflows increasingly move from file-based to live, functions that previously could be done offline can pose technical challenges or be exceedingly expensive to implement. One such function is the requirement for subtitle creation and transcription for live video streams. In this demonstration, you will see how ten NETINT T1U VPUs installed in a 1RU server with a 96-core Ampere Altra Max processor running OpenAI’s Whisper ASR model can produce dozens of simultaneous ABR ladder transcodes using AV1, HEVC, or H.264.

Keywords: Streaming Media, Streaming Summit, Streaming Video, Dan Rayburn, OTT



Speakers

Alex Liu
Co-founder and COO
NETINT Technologies
Sean Varley
Chief Evangelist, VP, Business Development
Ampere