.Guarantee being compatible along with numerous structures, including.NET 6.0,. Web Structure 4.6.2, and.NET Criterion 2.0 and above.Lessen dependences to avoid model problems and also the necessity for binding redirects.Transcribing Sound Information.Some of the key functions of the SDK is audio transcription. Developers can record audio data asynchronously or in real-time. Below is actually an example of how to transcribe an audio documents:.utilizing AssemblyAI.utilizing AssemblyAI.Transcripts.var client = brand-new AssemblyAIClient(" YOUR_API_KEY").var transcript = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For local data, comparable code can be used to obtain transcription.await making use of var flow = new FileStream("./ nbc.mp3", FileMode.Open).var transcript = wait for client.Transcripts.TranscribeAsync(.stream,.brand-new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK also holds real-time audio transcription making use of Streaming Speech-to-Text. This component is specifically practical for applications requiring quick processing of audio data.utilizing AssemblyAI.Realtime.wait for making use of var scribe = brand-new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Final: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for getting audio from a microphone as an example.GetAudio( async (portion) => wait for transcriber.SendAudioAsync( chunk)).await transcriber.CloseAsync().Utilizing LeMUR for LLM Apps.The SDK integrates along with LeMUR to allow developers to construct sizable foreign language version (LLM) apps on vocal data. Listed below is actually an instance:.var lemurTaskParams = brand new LemurTaskParams.Cause="Provide a brief review of the records.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var action = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Cleverness Versions.Also, the SDK features integrated support for audio knowledge models, allowing view study and other enhanced components.var records = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var lead to transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// GOOD, NEUTRAL, or even downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For additional information, check out the official AssemblyAI blog.Image source: Shutterstock.