代码
using Azure;
using Azure.AI.Inference;
// Azure endpoint configuration
var endpoint = new Uri("https://###.services.ai.azure.com/models");
var credential = new AzureKeyCredential("###");
var model = "Phi-4-multimodal-instruct";
var client = new ChatCompletionsClient(
endpoint,
credential,
new AzureAIInferenceClientOptions());
ChatMessageContentItem[] userContent =
{
new ChatMessageAudioContentItem(audioFilePath: "sample.mp3", AudioContentFormat.Mp3)
};
var requestOptions = new ChatCompletionsOptions()
{
Messages =
{
new ChatRequestSystemMessage("Based on the attached audio, generate a comprehensive text transcription of the spoken content."),
new ChatRequestUserMessage(userContent),
},
Model = model,
Temperature = 1,
MaxTokens = 1000,
};
Response<ChatCompletions> response = client.Complete(requestOptions);
System.Console.WriteLine(response.Value.Content);
与GitHub和Azure终点同时进行的,结果相同 veried mp3是有效的
phi-4-Multimodal-Instruct应该通过C#客户端支持音频输入?有一种不同的方法来格式化此特定模型的音频输入?
用基于URI的构造函数将文件路径构造函数重新放置:
此更改可确保模型以预期格式接收音频。