630 lines
18 KiB
TypeScript
630 lines
18 KiB
TypeScript
import { v, ConvexError } from "convex/values";
|
|
import {
|
|
action,
|
|
internalAction,
|
|
internalMutation,
|
|
} from "./_generated/server";
|
|
import { api, internal } from "./_generated/api";
|
|
import type { FunctionReference } from "convex/server";
|
|
import {
|
|
generateImageViaOpenRouter,
|
|
DEFAULT_IMAGE_MODEL,
|
|
IMAGE_MODELS,
|
|
} from "./openrouter";
|
|
import type { Id } from "./_generated/dataModel";
|
|
import { assertNodeBelongsToCanvasOrThrow } from "./ai_utils";
|
|
|
|
const MAX_IMAGE_RETRIES = 2;
|
|
|
|
type ErrorCategory =
|
|
| "credits"
|
|
| "policy"
|
|
| "timeout"
|
|
| "transient"
|
|
| "provider"
|
|
| "unknown";
|
|
|
|
interface ErrorData {
|
|
code?: string;
|
|
[key: string]: unknown;
|
|
}
|
|
|
|
function getErrorCode(error: unknown): string | undefined {
|
|
if (error instanceof ConvexError) {
|
|
const data = error.data as ErrorData;
|
|
return data?.code;
|
|
}
|
|
return undefined;
|
|
}
|
|
|
|
function errorMessage(error: unknown): string {
|
|
if (error instanceof Error) return error.message;
|
|
return String(error ?? "Generation failed");
|
|
}
|
|
|
|
function parseOpenRouterStatus(message: string): number | null {
|
|
const match = message.match(/OpenRouter API error\s+(\d+)/i);
|
|
if (!match) return null;
|
|
const parsed = Number(match[1]);
|
|
return Number.isFinite(parsed) ? parsed : null;
|
|
}
|
|
|
|
function categorizeError(error: unknown): {
|
|
category: ErrorCategory;
|
|
retryable: boolean;
|
|
} {
|
|
const code = getErrorCode(error);
|
|
const message = errorMessage(error);
|
|
const lower = message.toLowerCase();
|
|
const status = parseOpenRouterStatus(message);
|
|
|
|
if (
|
|
code === "CREDITS_TEST_DISABLED" ||
|
|
code === "CREDITS_INVALID_AMOUNT" ||
|
|
code === "CREDITS_BALANCE_NOT_FOUND" ||
|
|
code === "CREDITS_DAILY_CAP_REACHED" ||
|
|
code === "CREDITS_CONCURRENCY_LIMIT"
|
|
) {
|
|
return { category: "credits", retryable: false };
|
|
}
|
|
|
|
if (
|
|
code === "OPENROUTER_MODEL_REFUSAL" ||
|
|
lower.includes("content policy") ||
|
|
lower.includes("policy") ||
|
|
lower.includes("moderation") ||
|
|
lower.includes("safety") ||
|
|
lower.includes("refusal") ||
|
|
lower.includes("policy_violation")
|
|
) {
|
|
return { category: "policy", retryable: false };
|
|
}
|
|
|
|
if (status !== null) {
|
|
if (status >= 500 || status === 408 || status === 429 || status === 499) {
|
|
return { category: "provider", retryable: true };
|
|
}
|
|
if (status >= 400 && status < 500) {
|
|
return { category: "provider", retryable: false };
|
|
}
|
|
}
|
|
|
|
if (
|
|
lower.includes("timeout") ||
|
|
lower.includes("timed out") ||
|
|
lower.includes("deadline") ||
|
|
lower.includes("abort") ||
|
|
lower.includes("etimedout")
|
|
) {
|
|
return { category: "timeout", retryable: true };
|
|
}
|
|
|
|
if (
|
|
lower.includes("fetch failed") ||
|
|
lower.includes("network") ||
|
|
lower.includes("connection") ||
|
|
lower.includes("econnreset") ||
|
|
lower.includes("temporarily unavailable") ||
|
|
lower.includes("service unavailable") ||
|
|
lower.includes("rate limit") ||
|
|
lower.includes("overloaded")
|
|
) {
|
|
return { category: "transient", retryable: true };
|
|
}
|
|
|
|
return { category: "unknown", retryable: false };
|
|
}
|
|
|
|
function formatTerminalStatusMessage(error: unknown): string {
|
|
const code = getErrorCode(error);
|
|
if (code) {
|
|
return code;
|
|
}
|
|
|
|
const message = errorMessage(error).trim() || "Generation failed";
|
|
const { category } = categorizeError(error);
|
|
|
|
const prefixByCategory: Record<Exclude<ErrorCategory, "unknown">, string> = {
|
|
credits: "Credits",
|
|
policy: "Policy",
|
|
timeout: "Timeout",
|
|
transient: "Netzwerk",
|
|
provider: "Provider",
|
|
};
|
|
|
|
if (category === "unknown") {
|
|
return message;
|
|
}
|
|
|
|
const prefix = prefixByCategory[category];
|
|
if (message.toLowerCase().startsWith(prefix.toLowerCase())) {
|
|
return message;
|
|
}
|
|
|
|
return `${prefix}: ${message}`;
|
|
}
|
|
|
|
function wait(ms: number) {
|
|
return new Promise<void>((resolve) => {
|
|
setTimeout(resolve, ms);
|
|
});
|
|
}
|
|
|
|
async function generateImageWithAutoRetry(
|
|
operation: () => Promise<Awaited<ReturnType<typeof generateImageViaOpenRouter>>>,
|
|
onRetry: (
|
|
retryCount: number,
|
|
maxRetries: number,
|
|
failure: { message: string; category: ErrorCategory }
|
|
) => Promise<void>
|
|
) {
|
|
let lastError: unknown = null;
|
|
const startedAt = Date.now();
|
|
|
|
for (let attempt = 0; attempt <= MAX_IMAGE_RETRIES; attempt++) {
|
|
const attemptStartedAt = Date.now();
|
|
try {
|
|
const result = await operation();
|
|
console.info("[generateImageWithAutoRetry] success", {
|
|
attempts: attempt + 1,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
lastAttemptDurationMs: Date.now() - attemptStartedAt,
|
|
});
|
|
return result;
|
|
} catch (error) {
|
|
lastError = error;
|
|
const { retryable, category } = categorizeError(error);
|
|
const retryCount = attempt + 1;
|
|
const hasRemainingRetry = retryCount <= MAX_IMAGE_RETRIES;
|
|
|
|
console.warn("[generateImageWithAutoRetry] attempt failed", {
|
|
attempt: retryCount,
|
|
maxAttempts: MAX_IMAGE_RETRIES + 1,
|
|
retryable,
|
|
hasRemainingRetry,
|
|
category,
|
|
attemptDurationMs: Date.now() - attemptStartedAt,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
message: errorMessage(error),
|
|
});
|
|
|
|
if (!retryable || !hasRemainingRetry) {
|
|
throw error;
|
|
}
|
|
|
|
await onRetry(retryCount, MAX_IMAGE_RETRIES, {
|
|
message: errorMessage(error),
|
|
category,
|
|
});
|
|
await wait(Math.min(1500, 400 * retryCount));
|
|
}
|
|
}
|
|
|
|
throw lastError ?? new Error("Generation failed");
|
|
}
|
|
|
|
export const markNodeExecuting = internalMutation({
|
|
args: {
|
|
nodeId: v.id("nodes"),
|
|
},
|
|
handler: async (ctx, { nodeId }) => {
|
|
await ctx.db.patch(nodeId, {
|
|
status: "executing",
|
|
retryCount: 0,
|
|
statusMessage: undefined,
|
|
});
|
|
},
|
|
});
|
|
|
|
export const markNodeRetry = internalMutation({
|
|
args: {
|
|
nodeId: v.id("nodes"),
|
|
retryCount: v.number(),
|
|
maxRetries: v.number(),
|
|
failureMessage: v.string(),
|
|
},
|
|
handler: async (ctx, { nodeId, retryCount, maxRetries, failureMessage }) => {
|
|
const reason =
|
|
typeof failureMessage === "string" && failureMessage.trim().length > 0
|
|
? failureMessage
|
|
: "temporärer Fehler";
|
|
await ctx.db.patch(nodeId, {
|
|
status: "executing",
|
|
retryCount,
|
|
statusMessage: `Retry ${retryCount}/${maxRetries} — ${reason}`,
|
|
});
|
|
},
|
|
});
|
|
|
|
export const finalizeImageSuccess = internalMutation({
|
|
args: {
|
|
nodeId: v.id("nodes"),
|
|
prompt: v.string(),
|
|
modelId: v.string(),
|
|
storageId: v.id("_storage"),
|
|
aspectRatio: v.optional(v.string()),
|
|
retryCount: v.number(),
|
|
},
|
|
handler: async (
|
|
ctx,
|
|
{ nodeId, prompt, modelId, storageId, aspectRatio, retryCount }
|
|
) => {
|
|
const modelConfig = IMAGE_MODELS[modelId];
|
|
if (!modelConfig) {
|
|
throw new Error(`Unknown model: ${modelId}`);
|
|
}
|
|
|
|
const existing = await ctx.db.get(nodeId);
|
|
if (!existing) {
|
|
throw new Error("Node not found");
|
|
}
|
|
|
|
const prev =
|
|
existing.data && typeof existing.data === "object"
|
|
? (existing.data as Record<string, unknown>)
|
|
: {};
|
|
const creditCost = modelConfig.creditCost;
|
|
const resolvedAspectRatio =
|
|
aspectRatio?.trim() ||
|
|
(typeof prev.aspectRatio === "string" ? prev.aspectRatio : undefined);
|
|
|
|
await ctx.db.patch(nodeId, {
|
|
status: "done",
|
|
retryCount,
|
|
statusMessage: undefined,
|
|
data: {
|
|
...prev,
|
|
storageId,
|
|
prompt,
|
|
model: modelId,
|
|
modelLabel: modelConfig.name,
|
|
modelTier: modelConfig.tier,
|
|
generatedAt: Date.now(),
|
|
creditCost,
|
|
...(resolvedAspectRatio ? { aspectRatio: resolvedAspectRatio } : {}),
|
|
},
|
|
});
|
|
|
|
return { creditCost };
|
|
},
|
|
});
|
|
|
|
export const finalizeImageFailure = internalMutation({
|
|
args: {
|
|
nodeId: v.id("nodes"),
|
|
retryCount: v.number(),
|
|
statusMessage: v.string(),
|
|
},
|
|
handler: async (ctx, { nodeId, retryCount, statusMessage }) => {
|
|
await ctx.db.patch(nodeId, {
|
|
status: "error",
|
|
retryCount,
|
|
statusMessage,
|
|
});
|
|
},
|
|
});
|
|
|
|
export const generateAndStoreImage = internalAction({
|
|
args: {
|
|
nodeId: v.id("nodes"),
|
|
prompt: v.string(),
|
|
referenceStorageId: v.optional(v.id("_storage")),
|
|
referenceImageUrl: v.optional(v.string()),
|
|
model: v.string(),
|
|
aspectRatio: v.optional(v.string()),
|
|
},
|
|
handler: async (ctx, args) => {
|
|
const startedAt = Date.now();
|
|
const apiKey = process.env.OPENROUTER_API_KEY;
|
|
if (!apiKey) {
|
|
throw new Error("OPENROUTER_API_KEY is not set");
|
|
}
|
|
|
|
console.info("[generateAndStoreImage] start", {
|
|
nodeId: args.nodeId,
|
|
model: args.model,
|
|
hasReferenceStorageId: Boolean(args.referenceStorageId),
|
|
hasReferenceImageUrl: Boolean(args.referenceImageUrl?.trim()),
|
|
aspectRatio: args.aspectRatio?.trim() || null,
|
|
promptLength: args.prompt.length,
|
|
});
|
|
|
|
let retryCount = 0;
|
|
let referenceImageUrl = args.referenceImageUrl?.trim() || undefined;
|
|
if (args.referenceStorageId) {
|
|
referenceImageUrl =
|
|
(await ctx.storage.getUrl(args.referenceStorageId)) ?? undefined;
|
|
}
|
|
|
|
try {
|
|
const result = await generateImageWithAutoRetry(
|
|
() =>
|
|
generateImageViaOpenRouter(apiKey, {
|
|
prompt: args.prompt,
|
|
referenceImageUrl,
|
|
model: args.model,
|
|
aspectRatio: args.aspectRatio,
|
|
}),
|
|
async (nextRetryCount, maxRetries, failure) => {
|
|
retryCount = nextRetryCount;
|
|
await ctx.runMutation(internal.ai.markNodeRetry, {
|
|
nodeId: args.nodeId,
|
|
retryCount: nextRetryCount,
|
|
maxRetries,
|
|
failureMessage: failure.message,
|
|
});
|
|
}
|
|
);
|
|
|
|
const decodeStartedAt = Date.now();
|
|
const binaryString = atob(result.imageBase64);
|
|
const bytes = new Uint8Array(binaryString.length);
|
|
for (let i = 0; i < binaryString.length; i++) {
|
|
bytes[i] = binaryString.charCodeAt(i);
|
|
}
|
|
console.info("[generateAndStoreImage] image decoded", {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
decodeDurationMs: Date.now() - decodeStartedAt,
|
|
bytes: bytes.length,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
});
|
|
|
|
const storageStartedAt = Date.now();
|
|
const blob = new Blob([bytes], { type: result.mimeType });
|
|
const storageId = await ctx.storage.store(blob);
|
|
console.info("[generateAndStoreImage] image stored", {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
storageDurationMs: Date.now() - storageStartedAt,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
});
|
|
|
|
return {
|
|
storageId: storageId as Id<"_storage">,
|
|
retryCount,
|
|
};
|
|
} catch (error) {
|
|
console.error("[generateAndStoreImage] failed", {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
message: errorMessage(error),
|
|
category: categorizeError(error).category,
|
|
});
|
|
throw error;
|
|
}
|
|
},
|
|
});
|
|
|
|
export const processImageGeneration = internalAction({
|
|
args: {
|
|
nodeId: v.id("nodes"),
|
|
prompt: v.string(),
|
|
modelId: v.string(),
|
|
referenceStorageId: v.optional(v.id("_storage")),
|
|
referenceImageUrl: v.optional(v.string()),
|
|
aspectRatio: v.optional(v.string()),
|
|
reservationId: v.optional(v.id("creditTransactions")),
|
|
shouldDecrementConcurrency: v.boolean(),
|
|
userId: v.string(),
|
|
},
|
|
handler: async (ctx, args) => {
|
|
const startedAt = Date.now();
|
|
console.info("[processImageGeneration] start", {
|
|
nodeId: args.nodeId,
|
|
reservationId: args.reservationId ?? null,
|
|
shouldDecrementConcurrency: args.shouldDecrementConcurrency,
|
|
userId: args.userId,
|
|
});
|
|
|
|
let retryCount = 0;
|
|
|
|
try {
|
|
const result = await ctx.runAction(internal.ai.generateAndStoreImage, {
|
|
nodeId: args.nodeId,
|
|
prompt: args.prompt,
|
|
referenceStorageId: args.referenceStorageId,
|
|
referenceImageUrl: args.referenceImageUrl,
|
|
model: args.modelId,
|
|
aspectRatio: args.aspectRatio,
|
|
});
|
|
retryCount = result.retryCount;
|
|
|
|
const { creditCost } = await ctx.runMutation(internal.ai.finalizeImageSuccess, {
|
|
nodeId: args.nodeId,
|
|
prompt: args.prompt,
|
|
modelId: args.modelId,
|
|
storageId: result.storageId,
|
|
aspectRatio: args.aspectRatio,
|
|
retryCount,
|
|
});
|
|
|
|
if (args.reservationId) {
|
|
await ctx.runMutation(internal.credits.commitInternal, {
|
|
transactionId: args.reservationId,
|
|
actualCost: creditCost,
|
|
});
|
|
}
|
|
|
|
console.info("[processImageGeneration] success", {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
reservationId: args.reservationId ?? null,
|
|
});
|
|
} catch (error) {
|
|
console.error("[processImageGeneration] failed", {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
reservationId: args.reservationId ?? null,
|
|
category: categorizeError(error).category,
|
|
message: errorMessage(error),
|
|
});
|
|
|
|
if (args.reservationId) {
|
|
try {
|
|
await ctx.runMutation(internal.credits.releaseInternal, {
|
|
transactionId: args.reservationId,
|
|
});
|
|
} catch {
|
|
// Keep node status updates best-effort even if credit release fails.
|
|
}
|
|
}
|
|
|
|
await ctx.runMutation(internal.ai.finalizeImageFailure, {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
statusMessage: formatTerminalStatusMessage(error),
|
|
});
|
|
} finally {
|
|
if (args.shouldDecrementConcurrency) {
|
|
await ctx.runMutation(internal.credits.decrementConcurrency, {
|
|
userId: args.userId,
|
|
});
|
|
}
|
|
|
|
console.info("[processImageGeneration] finished", {
|
|
nodeId: args.nodeId,
|
|
retryCount,
|
|
totalDurationMs: Date.now() - startedAt,
|
|
shouldDecrementConcurrency: args.shouldDecrementConcurrency,
|
|
});
|
|
}
|
|
},
|
|
});
|
|
|
|
export const generateImage = action({
|
|
args: {
|
|
canvasId: v.id("canvases"),
|
|
nodeId: v.id("nodes"),
|
|
prompt: v.string(),
|
|
referenceStorageId: v.optional(v.id("_storage")),
|
|
referenceImageUrl: v.optional(v.string()),
|
|
model: v.optional(v.string()),
|
|
aspectRatio: v.optional(v.string()),
|
|
},
|
|
handler: async (
|
|
ctx,
|
|
args
|
|
): Promise<{ queued: true; nodeId: Id<"nodes"> }> => {
|
|
const startedAt = Date.now();
|
|
const canvas = await ctx.runQuery(api.canvases.get, {
|
|
canvasId: args.canvasId,
|
|
});
|
|
if (!canvas) {
|
|
throw new Error("Canvas not found");
|
|
}
|
|
|
|
const node = await ctx.runQuery(
|
|
api.nodes.get as FunctionReference<"query", "public">,
|
|
{
|
|
nodeId: args.nodeId,
|
|
includeStorageUrl: false,
|
|
}
|
|
);
|
|
if (!node) {
|
|
throw new Error("Node not found");
|
|
}
|
|
assertNodeBelongsToCanvasOrThrow(node, args.canvasId);
|
|
|
|
const userId = canvas.ownerId;
|
|
const verifiedCanvasId = canvas._id;
|
|
const verifiedNodeId = node._id;
|
|
|
|
const internalCreditsEnabled =
|
|
process.env.INTERNAL_CREDITS_ENABLED === "true";
|
|
|
|
const modelId = args.model ?? DEFAULT_IMAGE_MODEL;
|
|
const modelConfig = IMAGE_MODELS[modelId];
|
|
if (!modelConfig) {
|
|
throw new Error(`Unknown model: ${modelId}`);
|
|
}
|
|
|
|
await ctx.runMutation(internal.credits.checkAbuseLimits, {});
|
|
|
|
let usageIncremented = false;
|
|
const reservationId: Id<"creditTransactions"> | null = internalCreditsEnabled
|
|
? await ctx.runMutation(api.credits.reserve, {
|
|
estimatedCost: modelConfig.creditCost,
|
|
description: `Bildgenerierung — ${modelConfig.name}`,
|
|
model: modelId,
|
|
nodeId: verifiedNodeId,
|
|
canvasId: verifiedCanvasId,
|
|
})
|
|
: null;
|
|
|
|
if (!internalCreditsEnabled) {
|
|
await ctx.runMutation(internal.credits.incrementUsage, {});
|
|
usageIncremented = true;
|
|
}
|
|
|
|
const retryCount = 0;
|
|
let backgroundJobScheduled = false;
|
|
|
|
try {
|
|
await ctx.runMutation(internal.ai.markNodeExecuting, {
|
|
nodeId: verifiedNodeId,
|
|
});
|
|
|
|
await ctx.scheduler.runAfter(0, internal.ai.processImageGeneration, {
|
|
nodeId: verifiedNodeId,
|
|
prompt: args.prompt,
|
|
modelId,
|
|
referenceStorageId: args.referenceStorageId,
|
|
referenceImageUrl: args.referenceImageUrl,
|
|
aspectRatio: args.aspectRatio,
|
|
reservationId: reservationId ?? undefined,
|
|
shouldDecrementConcurrency: usageIncremented,
|
|
userId,
|
|
});
|
|
backgroundJobScheduled = true;
|
|
console.info("[generateImage] background job scheduled", {
|
|
nodeId: verifiedNodeId,
|
|
canvasId: verifiedCanvasId,
|
|
modelId,
|
|
reservationId: reservationId ?? null,
|
|
usageIncremented,
|
|
durationMs: Date.now() - startedAt,
|
|
});
|
|
return { queued: true as const, nodeId: verifiedNodeId };
|
|
} catch (error) {
|
|
console.error("[generateImage] scheduling failed", {
|
|
nodeId: verifiedNodeId,
|
|
canvasId: verifiedCanvasId,
|
|
modelId,
|
|
reservationId: reservationId ?? null,
|
|
usageIncremented,
|
|
durationMs: Date.now() - startedAt,
|
|
category: categorizeError(error).category,
|
|
message: errorMessage(error),
|
|
});
|
|
|
|
if (reservationId) {
|
|
try {
|
|
await ctx.runMutation(api.credits.release, {
|
|
transactionId: reservationId,
|
|
});
|
|
} catch {
|
|
// Prefer returning a clear node error over masking with cleanup failures.
|
|
}
|
|
}
|
|
|
|
await ctx.runMutation(internal.ai.finalizeImageFailure, {
|
|
nodeId: verifiedNodeId,
|
|
retryCount,
|
|
statusMessage: formatTerminalStatusMessage(error),
|
|
});
|
|
|
|
throw error;
|
|
} finally {
|
|
if (usageIncremented && !backgroundJobScheduled) {
|
|
await ctx.runMutation(internal.credits.decrementConcurrency, {
|
|
userId,
|
|
});
|
|
}
|
|
}
|
|
},
|
|
});
|