🐛 Fix openai total tokens variable set when streaming
This commit is contained in:
@ -275,7 +275,6 @@ export const ConversationContainer = (props: Props) => {
|
||||
const processClientSideActions = async (
|
||||
actions: NonNullable<ContinueChatResponse['clientSideActions']>
|
||||
) => {
|
||||
console.log('YES')
|
||||
if (isRecovered()) return
|
||||
for (const action of actions) {
|
||||
if (
|
||||
|
Reference in New Issue
Block a user