diff --git a/src/cookbook/tracing_with_fetched_deployed_prompt.ts b/src/cookbook/tracing_with_fetched_deployed_prompt.ts index 5cc4ac2..a5ac83b 100644 --- a/src/cookbook/tracing_with_fetched_deployed_prompt.ts +++ b/src/cookbook/tracing_with_fetched_deployed_prompt.ts @@ -75,6 +75,19 @@ async function llmJudgeEval(log: Log): Promise { } const _ragTemplate = async (deployment_id: string, llm_inputs: Record): Promise => { + // The Deployed Prompt is: + // Use the following pieces of context to answer the question. Do not make up an answer if no context is provided to help answer it. + // + // Context: + // --------- + // {{context}} + // + // --------- + // Question: {{question}} + // --------- + // + // Answer: + const deployedPrompt: UseDeployedPromptResponse = await getPrompt(deployment_id, llm_inputs); const response = await callOpenAI( deployedPrompt.prompt?.messages,