.Guarantee compatibility with several structures, including.NET 6.0,. NET Framework 4.6.2, and.NET Standard 2.0 and above.Minimize dependences to avoid model conflicts as well as the requirement for binding redirects.Recording Sound Record.One of the major functionalities of the SDK is audio transcription. Developers can easily transcribe audio documents asynchronously or in real-time. Below is an example of exactly how to record an audio documents:.making use of AssemblyAI.utilizing AssemblyAI.Transcripts.var client = brand new AssemblyAIClient(" YOUR_API_KEY").var records = wait for client.Transcripts.TranscribeAsync( brand new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For nearby data, identical code can be made use of to obtain transcription.await making use of var stream = brand new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Sound Transcription.The SDK additionally sustains real-time sound transcription utilizing Streaming Speech-to-Text. This attribute is especially useful for treatments requiring urgent handling of audio records.making use of AssemblyAI.Realtime.wait for using var scribe = new RealtimeTranscriber( brand new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( records =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Ultimate: transcript.Text "). ).await transcriber.ConnectAsync().// Pseudocode for obtaining audio coming from a mic for instance.GetAudio( async (piece) => await transcriber.SendAudioAsync( part)).wait for transcriber.CloseAsync().Making Use Of LeMUR for LLM Applications.The SDK combines with LeMUR to permit programmers to build big foreign language design (LLM) functions on vocal records. Below is actually an instance:.var lemurTaskParams = new LemurTaskParams.Prompt="Give a quick review of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var reaction = await client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Audio Intellect Designs.Also, the SDK possesses built-in support for audio knowledge versions, making it possible for belief analysis and various other innovative features.var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = true. ).foreach (var result in transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// BENEFICIAL, NEUTRAL, or downside.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").To learn more, visit the official AssemblyAI blog.Image resource: Shutterstock.