.Make sure compatibility along with numerous structures, including.NET 6.0,. NET Structure 4.6.2, and.NET Standard 2.0 and also above.Minimize reliances to prevent variation problems as well as the need for binding redirects.Translating Audio Data.One of the key performances of the SDK is audio transcription. Creators can easily transcribe audio documents asynchronously or even in real-time. Below is actually an instance of just how to transcribe an audio documents:.utilizing AssemblyAI.making use of AssemblyAI.Transcripts.var client = new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby documents, similar code may be used to achieve transcription.await making use of var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var transcript = await client.Transcripts.TranscribeAsync(.flow,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK likewise holds real-time sound transcription making use of Streaming Speech-to-Text. This feature is actually particularly beneficial for uses calling for urgent processing of audio information.making use of AssemblyAI.Realtime.await utilizing var transcriber = new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for receiving audio coming from a microphone for instance.GetAudio( async (part) => wait for transcriber.SendAudioAsync( piece)).await transcriber.CloseAsync().Making Use Of LeMUR for LLM Functions.The SDK includes with LeMUR to allow designers to construct large language version (LLM) applications on vocal data. Below is actually an example:.var lemurTaskParams = brand new LemurTaskParams.Cause="Offer a short rundown of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var reaction = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intelligence Models.Also, the SDK possesses integrated support for audio intellect designs, making it possible for feeling analysis and various other sophisticated features.var records = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more details, see the formal AssemblyAI blog.Image resource: Shutterstock.