.Ensure being compatible with several platforms, including.NET 6.0,. NET Structure 4.6.2, and.NET Requirement 2.0 as well as above.Reduce dependences to stop model problems and the necessity for binding redirects.Translating Audio Files.Some of the major capabilities of the SDK is actually audio transcription. Programmers can translate audio data asynchronously or even in real-time. Below is actually an instance of how to translate an audio file:.using AssemblyAI.using AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var transcript = await client.Transcripts.TranscribeAsync( brand-new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For regional files, comparable code can be made use of to achieve transcription.wait for using var stream = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK likewise reinforces real-time audio transcription making use of Streaming Speech-to-Text. This feature is actually especially helpful for treatments needing prompt handling of audio information.making use of AssemblyAI.Realtime.await using var transcriber = brand-new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Last: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for receiving sound from a mic as an example.GetAudio( async (part) => await transcriber.SendAudioAsync( piece)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Apps.The SDK combines with LeMUR to enable creators to build large language version (LLM) applications on voice records. Below is actually an instance:.var lemurTaskParams = brand-new LemurTaskParams.Urge="Supply a quick rundown of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Knowledge Models.In addition, the SDK comes with built-in assistance for audio intellect designs, making it possible for belief review as well as various other innovative functions.var records = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or even NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To read more, visit the formal AssemblyAI blog.Image source: Shutterstock.