Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
import { createLlamaPrompt } from "@/lib/createLlamaPrompt" | |
import { predict } from "./predict" | |
import { Preset } from "../engine/presets" | |
export const getStory = async ({ | |
preset, | |
prompt = "", | |
}: { | |
preset: Preset; | |
prompt: string; | |
}) => { | |
const query = createLlamaPrompt([ | |
{ | |
role: "system", | |
content: [ | |
`You are a comic book author specialized in ${preset.llmPrompt}`, | |
`You are going to be asked to write a comic book page, your mission is to answer a JSON array containing 4 items, to describe the page (one item per panel).`, | |
`Each array item should be a comic book panel caption the describe the environment, era, characters, objects, textures, lighting.`, | |
`Be brief in your caption don't add your own comments. Be straight to the point, and never reply things like "Sure, I can.." etc.` | |
].filter(item => item).join("\n") | |
}, | |
{ | |
role: "user", | |
content: `The story is: ${prompt}`, | |
} | |
]) | |
let result = "" | |
try { | |
result = `${await predict(query) || ""}`.trim() | |
if (!result.length) { | |
throw new Error("empty result!") | |
} | |
} catch (err) { | |
console.log(`prediction of the story failed, trying again..`) | |
try { | |
result = `${await predict(query+".") || ""}`.trim() | |
if (!result.length) { | |
throw new Error("empty result!") | |
} | |
} catch (err) { | |
console.error(`prediction of the story failed again!`) | |
throw new Error(`failed to generate the story ${err}`) | |
} | |
} | |
const tmp = result // result.split("Caption:").pop() || result | |
return tmp.replaceAll("\n", ", ") | |
} |