.Make sure compatibility with several structures, including.NET 6.0,. NET Structure 4.6.2, and.NET Specification 2.0 and above.Decrease dependences to avoid model conflicts and the need for tiing redirects.Recording Sound Info.One of the key functions of the SDK is audio transcription. Creators can translate audio files asynchronously or in real-time. Below is actually an instance of how to transcribe an audio file:.making use of AssemblyAI.using AssemblyAI.Transcripts.var client = new AssemblyAIClient(" YOUR_API_KEY").var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For regional documents, similar code can be utilized to attain transcription.wait for making use of var flow = brand new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also reinforces real-time audio transcription making use of Streaming Speech-to-Text. This attribute is specifically helpful for applications needing urgent handling of audio information.using AssemblyAI.Realtime.await making use of var transcriber = brand-new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for acquiring audio coming from a mic as an example.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( portion)).wait for transcriber.CloseAsync().Utilizing LeMUR for LLM Functions.The SDK combines along with LeMUR to enable programmers to build huge language version (LLM) functions on voice data. Below is actually an example:.var lemurTaskParams = brand-new LemurTaskParams.Cue="Supply a quick rundown of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var reaction = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intelligence Versions.Additionally, the SDK features integrated help for audio cleverness designs, allowing belief analysis as well as other sophisticated components.var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, check out the formal AssemblyAI blog.Image resource: Shutterstock.