diff --git a/pages/api/chat.ts b/pages/api/chat.ts index dc65c9a..945d37b 100644 --- a/pages/api/chat.ts +++ b/pages/api/chat.ts @@ -35,7 +35,7 @@ const handler = async (req: Request): Promise => { const message = messages[i]; const tokens = encoding.encode(message.content); - if (tokenCount + tokens.length + 2000 > model.tokenLimit) { + if (tokenCount + tokens.length + 1000 > model.tokenLimit) { break; } tokenCount += tokens.length; diff --git a/utils/server/index.ts b/utils/server/index.ts index 1042a78..711ae9b 100644 --- a/utils/server/index.ts +++ b/utils/server/index.ts @@ -45,7 +45,7 @@ export const OpenAIStream = async ( }, ...messages, ], - max_tokens: 2000, + max_tokens: 1000, temperature: 1, stream: true, }),