Machine Learning Engineer at Quantiphi·

We have to process the video stream to identify emotions. For which we need to use Amazon Rekognition/custom model on Amazon SageMaker. With Kinesis WebRTC Javascript SDK, currently, video can be streamed only into the kinesis signaling channel. Signaling channel data is available for streaming only and not processing (ML). So, how can we get real-time data for processing into Kinesis Streams from the frontend?

For streaming the video from frontend to backend into the Amazon Kinesis Video Streams for processing, we tested with Kinesis webRTC JavaScript SDK, and we are facing issues while implementing as mentioned above, so would Chime SDK serve as an alternative to this?

In Rekognition, "create-stream-processor" has a settings parameter. This currently only supports FaceSearch. We are looking to Detect and analyze faces. Is that possible with "create-stream-processor" in the Python SDK? Or do we have to use the Java SDK?

READ LESS
4 upvotes·23.1K views
Avatar of Harsh Solanki

Harsh Solanki

Machine Learning Engineer at Quantiphi