⚡ (openai) Truncate messages sequence automatically if reaching token limit
This commit is contained in:
@@ -77,6 +77,8 @@ export const TemplatesModal = ({ isOpen, onClose, onTypebotChoose }: Props) => {
|
||||
borderRightWidth={1}
|
||||
justify="space-between"
|
||||
flexShrink={0}
|
||||
overflowY="scroll"
|
||||
className="hide-scrollbar"
|
||||
>
|
||||
<Stack spacing={5}>
|
||||
<Stack spacing={2}>
|
||||
|
||||
@@ -17,7 +17,7 @@ export const useToast = () => {
|
||||
}: Omit<ToastProps, 'onClose'>) => {
|
||||
toast({
|
||||
position: 'top-right',
|
||||
duration: details ? null : undefined,
|
||||
duration: details && status === 'error' ? null : undefined,
|
||||
render: ({ onClose }) => (
|
||||
<Toast
|
||||
title={title}
|
||||
|
||||
Reference in New Issue
Block a user