.Ensure being compatible with several structures, including.NET 6.0,. NET Structure 4.6.2, and.NET Requirement 2.0 as well as above.Decrease dependencies to prevent version disputes as well as the need for tiing redirects.Recording Sound Data.Among the key functions of the SDK is audio transcription. Creators may transcribe audio data asynchronously or even in real-time. Below is an example of how to transcribe an audio documents:.utilizing AssemblyAI.making use of AssemblyAI.Transcripts.var customer = new AssemblyAIClient(" YOUR_API_KEY").var records = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local documents, identical code can be made use of to accomplish transcription.wait for using var flow = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = wait for client.Transcripts.TranscribeAsync(.stream,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also reinforces real-time sound transcription making use of Streaming Speech-to-Text. This feature is actually especially beneficial for uses requiring prompt processing of audio records.using AssemblyAI.Realtime.await utilizing var transcriber = brand-new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Final: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for obtaining audio coming from a microphone for example.GetAudio( async (chunk) => wait for transcriber.SendAudioAsync( portion)).await transcriber.CloseAsync().Taking Advantage Of LeMUR for LLM Applications.The SDK incorporates with LeMUR to make it possible for developers to develop big foreign language design (LLM) functions on vocal records. Listed below is an example:.var lemurTaskParams = brand-new LemurTaskParams.Trigger="Offer a short conclusion of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var reaction = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intellect Styles.In addition, the SDK features built-in help for audio intellect designs, permitting belief study and various other state-of-the-art components.var records = await client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// FAVORABLE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For more information, check out the official AssemblyAI blog.Image resource: Shutterstock.