199 lines
6.1 KiB
TypeScript
199 lines
6.1 KiB
TypeScript
import {
|
|
Conversation,
|
|
ErrorMessage,
|
|
KeyValuePair,
|
|
Message,
|
|
OpenAIModel,
|
|
} from '@/types';
|
|
import {
|
|
FC,
|
|
MutableRefObject,
|
|
useCallback,
|
|
useEffect,
|
|
useRef,
|
|
useState,
|
|
} from 'react';
|
|
import { useTranslation } from 'next-i18next';
|
|
import { ChatInput } from './ChatInput';
|
|
import { ChatLoader } from './ChatLoader';
|
|
import { ChatMessage } from './ChatMessage';
|
|
import { ErrorMessageDiv } from './ErrorMessageDiv';
|
|
import { ModelSelect } from './ModelSelect';
|
|
import { SystemPrompt } from './SystemPrompt';
|
|
|
|
interface Props {
|
|
conversation: Conversation;
|
|
models: OpenAIModel[];
|
|
apiKey: string;
|
|
serverSideApiKeyIsSet: boolean;
|
|
messageIsStreaming: boolean;
|
|
modelError: ErrorMessage | null;
|
|
messageError: boolean;
|
|
loading: boolean;
|
|
onSend: (message: Message, deleteCount?: number) => void;
|
|
onUpdateConversation: (
|
|
conversation: Conversation,
|
|
data: KeyValuePair,
|
|
) => void;
|
|
onEditMessage: (message: Message, messageIndex: number) => void;
|
|
stopConversationRef: MutableRefObject<boolean>;
|
|
}
|
|
|
|
export const Chat: FC<Props> = ({
|
|
conversation,
|
|
models,
|
|
apiKey,
|
|
serverSideApiKeyIsSet,
|
|
messageIsStreaming,
|
|
modelError,
|
|
messageError,
|
|
loading,
|
|
onSend,
|
|
onUpdateConversation,
|
|
onEditMessage,
|
|
stopConversationRef,
|
|
}) => {
|
|
const { t } = useTranslation('chat');
|
|
const [currentMessage, setCurrentMessage] = useState<Message>();
|
|
const [autoScrollEnabled, setAutoScrollEnabled] = useState(true);
|
|
|
|
const messagesEndRef = useRef<HTMLDivElement>(null);
|
|
const chatContainerRef = useRef<HTMLDivElement>(null);
|
|
const textareaRef = useRef<HTMLTextAreaElement>(null);
|
|
|
|
const scrollToBottom = useCallback(() => {
|
|
if (autoScrollEnabled) {
|
|
messagesEndRef.current?.scrollIntoView({ behavior: 'smooth' });
|
|
textareaRef.current?.focus();
|
|
}
|
|
}, [autoScrollEnabled]);
|
|
|
|
const handleScroll = () => {
|
|
if (chatContainerRef.current) {
|
|
const { scrollTop, scrollHeight, clientHeight } =
|
|
chatContainerRef.current;
|
|
const bottomTolerance = 5;
|
|
|
|
if (scrollTop + clientHeight < scrollHeight - bottomTolerance) {
|
|
setAutoScrollEnabled(false);
|
|
} else {
|
|
setAutoScrollEnabled(true);
|
|
}
|
|
}
|
|
};
|
|
|
|
useEffect(() => {
|
|
scrollToBottom();
|
|
setCurrentMessage(conversation.messages[conversation.messages.length - 2]);
|
|
}, [conversation.messages, scrollToBottom]);
|
|
|
|
useEffect(() => {
|
|
const chatContainer = chatContainerRef.current;
|
|
|
|
if (chatContainer) {
|
|
chatContainer.addEventListener('scroll', handleScroll);
|
|
|
|
return () => {
|
|
chatContainer.removeEventListener('scroll', handleScroll);
|
|
};
|
|
}
|
|
}, []);
|
|
|
|
return (
|
|
<div className="overflow-none relative flex-1 bg-white dark:bg-[#343541]">
|
|
{!(apiKey || serverSideApiKeyIsSet) ? (
|
|
<div className="mx-auto flex h-full w-[300px] flex-col justify-center space-y-6 sm:w-[500px]">
|
|
<div className="text-center text-2xl font-semibold text-gray-800 dark:text-gray-100">
|
|
{t('OpenAI API Key Required')}
|
|
</div>
|
|
<div className="text-center text-gray-500 dark:text-gray-400">
|
|
{t(
|
|
'Please set your OpenAI API key in the bottom left of the sidebar.',
|
|
)}
|
|
</div>
|
|
</div>
|
|
) : modelError ? (
|
|
<ErrorMessageDiv error={modelError} />
|
|
) : (
|
|
<>
|
|
<div className="max-h-full overflow-scroll" ref={chatContainerRef}>
|
|
{conversation.messages.length === 0 ? (
|
|
<>
|
|
<div className="mx-auto flex w-[350px] flex-col space-y-10 pt-12 sm:w-[600px]">
|
|
<div className="text-center text-4xl font-semibold text-gray-800 dark:text-gray-100">
|
|
{models.length === 0 ? t('Loading...') : 'Chatbot UI'}
|
|
</div>
|
|
|
|
{models.length > 0 && (
|
|
<div className="flex h-full flex-col space-y-4 rounded border border-neutral-500 p-4">
|
|
<ModelSelect
|
|
model={conversation.model}
|
|
models={models}
|
|
onModelChange={(model) =>
|
|
onUpdateConversation(conversation, {
|
|
key: 'model',
|
|
value: model,
|
|
})
|
|
}
|
|
/>
|
|
|
|
<SystemPrompt
|
|
conversation={conversation}
|
|
onChangePrompt={(prompt) =>
|
|
onUpdateConversation(conversation, {
|
|
key: 'prompt',
|
|
value: prompt,
|
|
})
|
|
}
|
|
/>
|
|
</div>
|
|
)}
|
|
</div>
|
|
</>
|
|
) : (
|
|
<>
|
|
<div className="flex justify-center border border-b-neutral-300 bg-neutral-100 py-2 text-sm text-neutral-500 dark:border-none dark:bg-[#444654] dark:text-neutral-200">
|
|
{t('Model')}: {conversation.model.name}
|
|
</div>
|
|
|
|
{conversation.messages.map((message, index) => (
|
|
<ChatMessage
|
|
key={index}
|
|
message={message}
|
|
messageIndex={index}
|
|
onEditMessage={onEditMessage}
|
|
/>
|
|
))}
|
|
|
|
{loading && <ChatLoader />}
|
|
|
|
<div
|
|
className="h-[162px] bg-white dark:bg-[#343541]"
|
|
ref={messagesEndRef}
|
|
/>
|
|
</>
|
|
)}
|
|
</div>
|
|
|
|
<ChatInput
|
|
stopConversationRef={stopConversationRef}
|
|
textareaRef={textareaRef}
|
|
messageIsStreaming={messageIsStreaming}
|
|
messages={conversation.messages}
|
|
model={conversation.model}
|
|
onSend={(message) => {
|
|
setCurrentMessage(message);
|
|
onSend(message);
|
|
}}
|
|
onRegenerate={() => {
|
|
if (currentMessage) {
|
|
onSend(currentMessage, 2);
|
|
}
|
|
}}
|
|
/>
|
|
</>
|
|
)}
|
|
</div>
|
|
);
|
|
};
|