Transcribing Audio Files With Amazon Transcribe, Lambda & S3 Part 2

Amazon Transcribe is one of AWS’s numerous machine learning services that is used to convert speech to text. Transcribe combines a deep learning process called Automatic Speech Recognition(ASR) and Natural Language Processing (NLP) to transcribe audio files. Across the globe, several organizations are leveraging this technology to automate media closed captioning & subtitling. Also, Amazon Transcribe supports transcription in over 30 languages including Hebrew, Japanese, Arabic, German, and others In this tutorial, we will be working with Amazon Transcribe to perform automatic speech recognition. Architecture A user or an application uploads an audio file to an S3 bucket. This upload triggers a Lambda function which will instruct Transcribe to begin the speech-to-text process. Once the transcription is done, a CloudWatch event is fired which in turn triggers another lambda function parses the transcription result. Create an S3 Bucket: First, we need to create an S3 Bucket which will serve as a repository for our audio and transcribed files. Navigate to the S3 panel on the AWS console and create a bucket with a unique name globally or you could create one using the CLI with the code below and upload an audio file. Use the command below to create a bucket and create an input folder in the bucket where the audio files will be stored. [...x]