.Ensure being compatible with numerous frameworks, including.NET 6.0,. NET Structure 4.6.2, and.NET Requirement 2.0 and also above.Minimize dependences to avoid model problems and the need for tiing redirects.Transcribing Sound Record.One of the primary functions of the SDK is audio transcription. Creators can translate audio files asynchronously or in real-time. Below is actually an example of just how to translate an audio report:.utilizing AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = brand-new AssemblyAIClient(" YOUR_API_KEY").var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local files, similar code can be used to accomplish transcription.await using var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var transcript = wait for client.Transcripts.TranscribeAsync(.flow,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK likewise holds real-time audio transcription using Streaming Speech-to-Text. This attribute is actually especially beneficial for applications calling for quick processing of audio data.making use of AssemblyAI.Realtime.wait for making use of var transcriber = new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for obtaining sound coming from a mic for instance.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( part)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Functions.The SDK combines with LeMUR to enable designers to create large language version (LLM) functions on vocal records. Listed below is an example:.var lemurTaskParams = brand new LemurTaskParams.Cue="Deliver a quick review of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var feedback = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intellect Versions.Additionally, the SDK includes integrated assistance for audio intelligence styles, enabling view evaluation and other advanced components.var transcript = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// GOOD, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, see the main AssemblyAI blog.Image resource: Shutterstock.