From 6e3f90cd2fd2448e2d4a6aac3d25ebfa3c2a5243 Mon Sep 17 00:00:00 2001 From: David Corbitt Date: Wed, 19 Jul 2023 18:10:23 -0700 Subject: [PATCH] Add more info to refinement --- src/server/utils/deriveNewContructFn.ts | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/src/server/utils/deriveNewContructFn.ts b/src/server/utils/deriveNewContructFn.ts index 5614744..34c023d 100644 --- a/src/server/utils/deriveNewContructFn.ts +++ b/src/server/utils/deriveNewContructFn.ts @@ -40,9 +40,9 @@ const requestUpdatedPromptFunction = async ( ) => { const originalModel = originalVariant.model as SupportedModel; let newContructionFn = ""; - const usefulTips = `Chain of thought means asking the model to think about its answer before it gives it to you. This is useful for getting more accurate answers. + const usefulTips = `Adding chain of thought means asking the model to think about its answer before it gives it to you. This is useful for getting more accurate answers. Do not add an assistant message. - This is what a prompt looks like without using function calls: + Function calls are a specific way for an LLM to return output. This is what a prompt looks like without using function calls: prompt = { model: "gpt-4", @@ -78,8 +78,8 @@ const requestUpdatedPromptFunction = async ( { name: "extract_sentiment", parameters: { - type: "object", - properties: { + type: "object", // parameters must always be an object with a properties key + properties: { // properties key is required sentiment: { type: "string", description: "one of positive/negative/neutral",