URGENT - Stream 'audio.pcm' and 'video.yuv' in real time via websocket

I am working on the bot (a headless meeting bot using the Zoom Meeting Linux SDK within a Docker container) which joins the meeting, ask permission to record after which I am getting two files and ‘audio.pcm’ and ‘video.yuv’ after starting the recording. However, I want to process the audio and video in real time via websocket to wss://audio.info & wss://video.info respectively. Audio and Video should be extracted and sent in real-time for AI audio and video analysis. I want to process it using a neural network. How to do this effectively at production level? Detailed explanation along with some examples or repositories or code snippets etc would really help.

(Video websockets are optimized for those doing real-time AI video analysis, providing 720p PNG image frames at 2fps.)

@ chunsiong.zoom
@ jake-recall-ai

@pramodkoujalagi this is beyond the scope of this forum.