AI/ML Workshop
Using Amazon Transcribe Websocket in a static web application, we can capture voice, transcribe the voice stream into text, and then perform sentiment analysis on the text to display the POSITIVE, NEGATIVE, NUTURL sentiments.
Components built and deployed:
- A static site with Amplify
- real-time audio transcription via Amazon Transcribe over a WebSocket.
- real-time comprehend API call to extract sentiments
This demo app uses browser microphone input and client-side JavaScript to demonstrate the real-time streaming audio transcription capability of Amazon Transcribe using WebSockets.
Check out the Amazon Transcribe WebSocket docs.
You have two options
Even though this is a static site consisting only of HTML, CSS, and client-side JavaScript, there is a build step required. Some of the modules used were originally made for server-side code and do not work natively in the browser.
We use browserify to enable browser support for the JavaScript modules we require().
- Clone the repo
- run
npm install - run
npm run-script buildto generatedist/main.js.
Once you've bundled the JavaScript, all you need is a webserver. For example, from your project directory:
npm install --global local-web-server
ws
This project is based on code written by Karan Grover from the Amazon Transcribe team, who did the hard work (audio encoding, event stream marshalling). Sentiment fetch based on the transcription is added by Jianjun Xu
This library is licensed under the Apache 2.0 License.