.Ensure being compatible along with various structures, including.NET 6.0,. Internet Platform 4.6.2, and.NET Standard 2.0 as well as above.Minimize reliances to stop variation disagreements as well as the necessity for binding redirects.Recording Audio Data.Among the primary capabilities of the SDK is audio transcription. Developers can transcribe audio files asynchronously or in real-time. Below is an example of just how to record an audio report:.using AssemblyAI.utilizing AssemblyAI.Transcripts.var customer = new AssemblyAIClient(" YOUR_API_KEY").var records = await client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3". ).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).For regional data, comparable code may be used to achieve transcription.await using var stream = new FileStream("./ nbc.mp3", FileMode.Open).var records = await client.Transcripts.TranscribeAsync(.stream,.brand new TranscriptOptionalParams.LanguageCode = TranscriptLanguageCode.EnUs.).transcript.EnsureStatusCompleted().Console.WriteLine( transcript.Text).Real-Time Audio Transcription.The SDK likewise holds real-time audio transcription using Streaming Speech-to-Text. This attribute is actually specifically helpful for treatments needing quick handling of audio records.utilizing AssemblyAI.Realtime.wait for utilizing var transcriber = brand new RealtimeTranscriber( brand-new RealtimeTranscriberOptions.ApiKey="YOUR_API_KEY",.SampleRate = 16_000. ).transcriber.PartialTranscriptReceived.Subscribe( transcript =>Console.WriteLine($" Limited: transcript.Text "). ).transcriber.FinalTranscriptReceived.Subscribe( records =>Console.WriteLine($" Ultimate: transcript.Text "). ).wait for transcriber.ConnectAsync().// Pseudocode for acquiring audio from a mic for example.GetAudio( async (piece) => wait for transcriber.SendAudioAsync( piece)).wait for transcriber.CloseAsync().Taking Advantage Of LeMUR for LLM Apps.The SDK incorporates with LeMUR to enable developers to develop huge language model (LLM) functions on vocal data. Here is actually an instance:.var lemurTaskParams = brand new LemurTaskParams.Urge="Supply a short rundown of the transcript.",.TranscriptIds = [transcript.Id],.FinalModel = LemurModel.AnthropicClaude3 _ 5_Sonnet..var response = wait for client.Lemur.TaskAsync( lemurTaskParams).Console.WriteLine( response.Response).Sound Intellect Versions.In addition, the SDK includes built-in help for audio cleverness models, permitting view evaluation and other sophisticated features.var transcript = wait for client.Transcripts.TranscribeAsync( new TranscriptParams.AudioUrl="https://storage.googleapis.com/aai-docs-samples/nbc.mp3",.SentimentAnalysis = real. ).foreach (var cause transcript.SentimentAnalysisResults!).Console.WriteLine( result.Text).Console.WriteLine( result.Sentiment)// POSITIVE, NEUTRAL, or NEGATIVE.Console.WriteLine( result.Confidence).Console.WriteLine($" Timestamp: result.Start - result.End ").For additional information, see the official AssemblyAI blog.Image resource: Shutterstock.