feat: 流式处理统一返回用户/系统内容并完善消息存储

引入 StreamProcessResult 统一封装流式处理结果,补充各 API 类型下用户输入与系统输出内容的提取与累计,用于会话消息持久化与用量统计;同时增强 Gemini 请求与响应内容解析能力,确保流式场景下消息与 token 使用数据完整一致。
This commit is contained in:
ccnetcore
2026-01-11 13:48:20 +08:00
parent 6f316d3e51
commit 5e37859157
2 changed files with 205 additions and 22 deletions

View File

@@ -6,6 +6,83 @@ namespace Yi.Framework.AiHub.Domain.Shared.Dtos.Gemini;
public static class GeminiGenerateContentAcquirer
{
/// <summary>
/// 从请求体中提取用户最后一条消息内容
/// 路径: contents[last].parts[last].text
/// </summary>
public static string GetLastUserContent(JsonElement request)
{
var contents = request.GetPath("contents");
if (!contents.HasValue || contents.Value.ValueKind != JsonValueKind.Array)
{
return string.Empty;
}
var contentsArray = contents.Value.EnumerateArray().ToList();
if (contentsArray.Count == 0)
{
return string.Empty;
}
var lastContent = contentsArray[^1];
var parts = lastContent.GetPath("parts");
if (!parts.HasValue || parts.Value.ValueKind != JsonValueKind.Array)
{
return string.Empty;
}
var partsArray = parts.Value.EnumerateArray().ToList();
if (partsArray.Count == 0)
{
return string.Empty;
}
// 获取最后一个 part 的 text
var lastPart = partsArray[^1];
return lastPart.GetPath("text").GetString() ?? string.Empty;
}
/// <summary>
/// 从响应中提取文本内容(非 thought 类型)
/// 路径: candidates[0].content.parts[].text (where thought != true)
/// </summary>
public static string GetTextContent(JsonElement response)
{
var candidates = response.GetPath("candidates");
if (!candidates.HasValue || candidates.Value.ValueKind != JsonValueKind.Array)
{
return string.Empty;
}
var candidatesArray = candidates.Value.EnumerateArray().ToList();
if (candidatesArray.Count == 0)
{
return string.Empty;
}
var parts = candidatesArray[0].GetPath("content", "parts");
if (!parts.HasValue || parts.Value.ValueKind != JsonValueKind.Array)
{
return string.Empty;
}
// 遍历所有 parts只取非 thought 的 text
foreach (var part in parts.Value.EnumerateArray())
{
var isThought = part.GetPath("thought").GetBool();
if (!isThought)
{
var text = part.GetPath("text").GetString();
if (!string.IsNullOrEmpty(text))
{
return text;
}
}
}
return string.Empty;
}
public static ThorUsageResponse? GetUsage(JsonElement response)
{
var usage = response.GetPath("usageMetadata");

View File

@@ -1177,27 +1177,27 @@ public class AiGateWayManager : DomainService
}
}
}, cancellationToken);
ThorUsageResponse? tokenUsage = null;
StreamProcessResult? processResult = null;
switch (apiType)
{
case ModelApiTypeEnum.Completions:
tokenUsage = await ProcessCompletionsStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
processResult = await ProcessCompletionsStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
case ModelApiTypeEnum.Messages:
tokenUsage = await ProcessAnthropicStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
processResult = await ProcessAnthropicStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
case ModelApiTypeEnum.Responses:
tokenUsage = await ProcessOpenAiResponsesStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
processResult = await ProcessOpenAiResponsesStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
case ModelApiTypeEnum.GenerateContent:
tokenUsage = await ProcessGeminiStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
processResult = await ProcessGeminiStreamAsync(messageQueue, requestBody, modelDescribe, cancellationToken);
break;
default:
throw new UserFriendlyException($"不支持的API类型: {apiType}");
}
// 标记完成并等待消费任务结束
isComplete = true;
await outputTask;
@@ -1206,25 +1206,25 @@ public class AiGateWayManager : DomainService
await _aiMessageManager.CreateUserMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = "不予存储",
Content = sessionId is null ? "不予存储" : processResult?.UserContent ?? string.Empty,
ModelId = sourceModelId,
TokenUsage = tokenUsage,
TokenUsage = processResult?.TokenUsage,
}, tokenId);
await _aiMessageManager.CreateSystemMessageAsync(userId, sessionId,
new MessageInputDto
{
Content = "不予存储",
Content = sessionId is null ? "不予存储" : processResult?.SystemContent ?? string.Empty,
ModelId = sourceModelId,
TokenUsage = tokenUsage
TokenUsage = processResult?.TokenUsage
}, tokenId);
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, tokenUsage, tokenId);
await _usageStatisticsManager.SetUsageAsync(userId, sourceModelId, processResult?.TokenUsage, tokenId);
// 扣减尊享token包用量
if (userId.HasValue && tokenUsage is not null && modelDescribe.IsPremium)
if (userId.HasValue && processResult?.TokenUsage is not null && modelDescribe.IsPremium)
{
var totalTokens = tokenUsage.TotalTokens ?? 0;
var totalTokens = processResult?.TokenUsage.TotalTokens ?? 0;
if (totalTokens > 0)
{
await PremiumPackageManager.TryConsumeTokensAsync(userId.Value, totalTokens);
@@ -1234,10 +1234,20 @@ public class AiGateWayManager : DomainService
#region - API类型的具体实现
/// <summary>
/// 流式处理结果,包含用户输入、系统输出和 token 使用情况
/// </summary>
private class StreamProcessResult
{
public string UserContent { get; set; } = string.Empty;
public string SystemContent { get; set; } = string.Empty;
public ThorUsageResponse TokenUsage { get; set; } = new();
}
/// <summary>
/// 处理 OpenAI Completions 格式流式响应
/// </summary>
private async Task<ThorUsageResponse> ProcessCompletionsStreamAsync(
private async Task<StreamProcessResult> ProcessCompletionsStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
@@ -1246,6 +1256,9 @@ public class AiGateWayManager : DomainService
var request = requestBody.Deserialize<ThorChatCompletionsRequest>(ThorJsonSerializer.DefaultOptions)!;
_specialCompatible.Compatible(request);
// 提取用户最后一条消息
var userContent = request.Messages?.LastOrDefault()?.MessagesStore ?? string.Empty;
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
@@ -1256,6 +1269,7 @@ public class AiGateWayManager : DomainService
var chatService = LazyServiceProvider.GetRequiredKeyedService<IChatCompletionService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.CompleteChatStreamAsync(modelDescribe, request, cancellationToken);
var tokenUsage = new ThorUsageResponse();
var systemContentBuilder = new StringBuilder();
try
{
@@ -1267,6 +1281,13 @@ public class AiGateWayManager : DomainService
tokenUsage = data.Usage;
}
// 累加系统输出内容 (choices[].delta.content)
var deltaContent = data.Choices?.FirstOrDefault()?.Delta?.Content;
if (!string.IsNullOrEmpty(deltaContent))
{
systemContentBuilder.Append(deltaContent);
}
var message = JsonSerializer.Serialize(data, ThorJsonSerializer.DefaultOptions);
messageQueue.Enqueue($"data: {message}\n\n");
}
@@ -1275,6 +1296,7 @@ public class AiGateWayManager : DomainService
{
_logger.LogError(e, "Ai对话异常");
var errorContent = $"对话Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
var model = new ThorChatCompletionsResponse()
{
Choices = new List<ThorChatChoiceResponse>()
@@ -1296,13 +1318,18 @@ public class AiGateWayManager : DomainService
}
messageQueue.Enqueue("data: [DONE]\n\n");
return tokenUsage;
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage
};
}
/// <summary>
/// 处理 Anthropic Messages 格式流式响应
/// </summary>
private async Task<ThorUsageResponse> ProcessAnthropicStreamAsync(
private async Task<StreamProcessResult> ProcessAnthropicStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
@@ -1311,6 +1338,16 @@ public class AiGateWayManager : DomainService
var request = requestBody.Deserialize<AnthropicInput>(ThorJsonSerializer.DefaultOptions)!;
_specialCompatible.AnthropicCompatible(request);
// 提取用户最后一条消息
var lastMessage = request.Messages?.LastOrDefault();
var userContent = lastMessage?.Content ?? string.Empty;
if (string.IsNullOrEmpty(userContent) && lastMessage?.Contents != null && lastMessage.Contents.Any())
{
// 如果是 Contents 数组,提取第一个 text 类型的内容
var textContent = lastMessage.Contents.FirstOrDefault(c => c.Type == "text");
userContent = textContent?.Text ?? System.Text.Json.JsonSerializer.Serialize(lastMessage.Contents);
}
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
@@ -1321,6 +1358,7 @@ public class AiGateWayManager : DomainService
var chatService = LazyServiceProvider.GetRequiredKeyedService<IAnthropicChatCompletionService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.StreamChatCompletionsAsync(modelDescribe, request, cancellationToken);
var tokenUsage = new ThorUsageResponse();
var systemContentBuilder = new StringBuilder();
try
{
@@ -1364,6 +1402,13 @@ public class AiGateWayManager : DomainService
}
}
// 累加系统输出内容 (delta.text)
var deltaText = responseResult.Item2?.Delta?.Text;
if (!string.IsNullOrEmpty(deltaText))
{
systemContentBuilder.Append(deltaText);
}
// 序列化为SSE格式字符串
var data = JsonSerializer.Serialize(responseResult.Item2, ThorJsonSerializer.DefaultOptions);
messageQueue.Enqueue($"{responseResult.Item1.Trim()}\ndata: {data}\n\n");
@@ -1373,19 +1418,25 @@ public class AiGateWayManager : DomainService
{
_logger.LogError(e, "Ai对话异常");
var errorContent = $"对话Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
throw new UserFriendlyException(errorContent);
}
tokenUsage.TotalTokens = (tokenUsage.InputTokens ?? 0) + (tokenUsage.OutputTokens ?? 0);
tokenUsage.SetSupplementalMultiplier(modelDescribe.Multiplier);
return tokenUsage;
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage
};
}
/// <summary>
/// 处理 OpenAI Responses 格式流式响应
/// </summary>
private async Task<ThorUsageResponse> ProcessOpenAiResponsesStreamAsync(
private async Task<StreamProcessResult> ProcessOpenAiResponsesStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
@@ -1393,6 +1444,27 @@ public class AiGateWayManager : DomainService
{
var request = requestBody.Deserialize<OpenAiResponsesInput>(ThorJsonSerializer.DefaultOptions)!;
// 提取用户输入内容 (input 字段可能是字符串或数组)
var userContent = string.Empty;
if (request.Input.ValueKind == JsonValueKind.String)
{
userContent = request.Input.GetString() ?? string.Empty;
}
else if (request.Input.ValueKind == JsonValueKind.Array)
{
// 获取最后一个 user 角色的消息
var inputArray = request.Input.EnumerateArray().ToList();
var lastUserMessage = inputArray.LastOrDefault(x =>
x.TryGetProperty("role", out var role) && role.GetString() == "user");
if (lastUserMessage.ValueKind != JsonValueKind.Undefined)
{
if (lastUserMessage.TryGetProperty("content", out var content))
{
userContent = content.GetString() ?? string.Empty;
}
}
}
// 处理 yi- 前缀
if (!string.IsNullOrEmpty(request.Model) &&
request.Model.StartsWith("yi-", StringComparison.OrdinalIgnoreCase))
@@ -1403,11 +1475,22 @@ public class AiGateWayManager : DomainService
var chatService = LazyServiceProvider.GetRequiredKeyedService<IOpenAiResponseService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.ResponsesStreamAsync(modelDescribe, request, cancellationToken);
ThorUsageResponse? tokenUsage = null;
var systemContentBuilder = new StringBuilder();
try
{
await foreach (var responseResult in completeChatResponse)
{
// 提取输出文本内容 (response.output_text.delta 事件)
if (responseResult.Item1.Contains("response.output_text.delta"))
{
var delta = responseResult.Item2?.GetPath("delta").GetString();
if (!string.IsNullOrEmpty(delta))
{
systemContentBuilder.Append(delta);
}
}
if (responseResult.Item1.Contains("response.completed"))
{
var obj = responseResult.Item2!.Value;
@@ -1434,29 +1517,46 @@ public class AiGateWayManager : DomainService
{
_logger.LogError(e, "Ai响应异常");
var errorContent = $"响应Ai异常异常信息\n当前Ai模型{request.Model}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
throw new UserFriendlyException(errorContent);
}
return tokenUsage ?? new ThorUsageResponse();
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage ?? new ThorUsageResponse()
};
}
/// <summary>
/// 处理 Gemini GenerateContent 格式流式响应
/// </summary>
private async Task<ThorUsageResponse> ProcessGeminiStreamAsync(
private async Task<StreamProcessResult> ProcessGeminiStreamAsync(
ConcurrentQueue<string> messageQueue,
JsonElement requestBody,
AiModelDescribe modelDescribe,
CancellationToken cancellationToken)
{
// 提取用户最后一条消息 (contents[last].parts[last].text)
var userContent = GeminiGenerateContentAcquirer.GetLastUserContent(requestBody);
var chatService = LazyServiceProvider.GetRequiredKeyedService<IGeminiGenerateContentService>(modelDescribe.HandlerName);
var completeChatResponse = chatService.GenerateContentStreamAsync(modelDescribe, requestBody, cancellationToken);
ThorUsageResponse? tokenUsage = null;
var systemContentBuilder = new StringBuilder();
try
{
await foreach (var responseResult in completeChatResponse)
{
// 累加系统输出内容 (candidates[0].content.parts[].text排除 thought)
var textContent = GeminiGenerateContentAcquirer.GetTextContent(responseResult!.Value);
if (!string.IsNullOrEmpty(textContent))
{
systemContentBuilder.Append(textContent);
}
if (responseResult!.Value.GetPath("candidates", 0, "finishReason").GetString() == "STOP")
{
tokenUsage = GeminiGenerateContentAcquirer.GetUsage(responseResult!.Value);
@@ -1483,10 +1583,16 @@ public class AiGateWayManager : DomainService
{
_logger.LogError(e, "Ai生成异常");
var errorContent = $"生成Ai异常异常信息\n当前Ai模型{modelDescribe.ModelId}\n异常信息{e.Message}\n异常堆栈{e}";
systemContentBuilder.Append(errorContent);
throw new UserFriendlyException(errorContent);
}
return tokenUsage ?? new ThorUsageResponse();
return new StreamProcessResult
{
UserContent = userContent,
SystemContent = systemContentBuilder.ToString(),
TokenUsage = tokenUsage ?? new ThorUsageResponse()
};
}
#endregion