.Guarantee being compatible along with multiple frameworks, including.NET 6.0,. Web Framework 4.6.2, and.NET Requirement 2.0 and also above.Reduce addictions to avoid version disputes as well as the need for binding redirects.Recording Audio Info.Some of the major capabilities of the SDK is audio transcription. Designers can easily record audio reports asynchronously or in real-time. Below is an instance of exactly how to transcribe an audio documents:.making use of AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local files, identical code could be made use of to achieve transcription.await using var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = wait for client.Transcripts.TranscribeAsync(.flow,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK also reinforces real-time audio transcription using Streaming Speech-to-Text. This component is especially useful for applications needing instant handling of audio data.making use of AssemblyAI.Realtime.wait for making use of var scribe = brand-new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Final: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for getting sound from a microphone as an example.GetAudio( async (chunk) => await transcriber.SendAudioAsync( portion)).wait for transcriber.CloseAsync().Taking Advantage Of LeMUR for LLM Functions.The SDK combines along with LeMUR to enable programmers to build big foreign language design (LLM) functions on vocal information. Listed here is actually an example:.var lemurTaskParams = brand new LemurTaskParams.Urge="Deliver a brief summary of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intelligence Models.In addition, the SDK features built-in support for audio knowledge models, allowing feeling study as well as various other sophisticated features.var transcript = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more information, go to the official AssemblyAI blog.Image source: Shutterstock.