.Make sure being compatible along with multiple structures, including.NET 6.0,. Web Structure 4.6.2, and.NET Standard 2.0 and also above.Decrease reliances to avoid version disagreements and the necessity for binding redirects.Transcribing Sound Information.One of the main functions of the SDK is audio transcription. Creators can easily translate audio reports asynchronously or in real-time. Below is actually an instance of just how to transcribe an audio report:.utilizing AssemblyAI.making use of AssemblyAI.Transcripts.var customer = new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local documents, similar code could be made use of to accomplish transcription.wait for using var stream = brand-new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK additionally holds real-time sound transcription using Streaming Speech-to-Text. This attribute is actually particularly helpful for applications requiring immediate processing of audio records.making use of AssemblyAI.Realtime.wait for using var transcriber = brand new RealtimeTranscriber( new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Partial: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for receiving sound from a mic for example.GetAudio( async (part) => await transcriber.SendAudioAsync( chunk)).wait for transcriber.CloseAsync().Utilizing LeMUR for LLM Apps.The SDK integrates with LeMUR to allow programmers to create big language design (LLM) applications on voice data. Right here is actually an instance:.var lemurTaskParams = brand new LemurTaskParams.Motivate="Provide a short summary of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var feedback = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intellect Models.In addition, the SDK comes with integrated help for audio intelligence versions, making it possible for conviction review as well as various other innovative features.var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// GOOD, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, visit the formal AssemblyAI blog.Image source: Shutterstock.