From 1dc4f86df58edd991a822af2c38107702ae0afc2 Mon Sep 17 00:00:00 2001 From: Mckay Wrigley Date: Sun, 2 Apr 2023 06:59:47 -0600 Subject: [PATCH] change output limit --- pages/api/chat.ts | 4 ++-- utils/server/index.ts | 4 ++-- 2 files changed, 4 insertions(+), 4 deletions(-) diff --git a/pages/api/chat.ts b/pages/api/chat.ts index 4083fd9..dc65c9a 100644 --- a/pages/api/chat.ts +++ b/pages/api/chat.ts @@ -2,7 +2,7 @@ import { ChatBody, Message } from '@/types/chat'; import { DEFAULT_SYSTEM_PROMPT } from '@/utils/app/const'; import { OpenAIError, OpenAIStream } from '@/utils/server'; import tiktokenModel from '@dqbd/tiktoken/encoders/cl100k_base.json'; -import { init, Tiktoken } from '@dqbd/tiktoken/lite/init'; +import { Tiktoken, init } from '@dqbd/tiktoken/lite/init'; // @ts-expect-error import wasm from '../../node_modules/@dqbd/tiktoken/lite/tiktoken_bg.wasm?module'; @@ -35,7 +35,7 @@ const handler = async (req: Request): Promise => { const message = messages[i]; const tokens = encoding.encode(message.content); - if (tokenCount + tokens.length + 1000 > model.tokenLimit) { + if (tokenCount + tokens.length + 2000 > model.tokenLimit) { break; } tokenCount += tokens.length; diff --git a/utils/server/index.ts b/utils/server/index.ts index 8ab473d..1042a78 100644 --- a/utils/server/index.ts +++ b/utils/server/index.ts @@ -33,7 +33,7 @@ export const OpenAIStream = async ( Authorization: `Bearer ${key ? key : process.env.OPENAI_API_KEY}`, ...(process.env.OPENAI_ORGANIZATION && { 'OpenAI-Organization': process.env.OPENAI_ORGANIZATION, - }) + }), }, method: 'POST', body: JSON.stringify({ @@ -45,7 +45,7 @@ export const OpenAIStream = async ( }, ...messages, ], - max_tokens: 1000, + max_tokens: 2000, temperature: 1, stream: true, }),