Quellcode durchsuchen

fix: infinite new chat

Timothy J. Baek vor 1 Jahr
Ursprung
Commit
fc9d037af7
2 geänderte Dateien mit 55 neuen und 40 gelöschten Zeilen
  1. 26 20
      src/routes/(app)/+page.svelte
  2. 29 20
      src/routes/(app)/c/[id]/+page.svelte

+ 26 - 20
src/routes/(app)/+page.svelte

@@ -46,6 +46,8 @@
 				currentMessage.parentId !== null ? history.messages[currentMessage.parentId] : null;
 		}
 		messages = _messages;
+	} else {
+		messages = [];
 	}
 
 	onMount(async () => {
@@ -80,13 +82,13 @@
 	// Ollama functions
 	//////////////////////////
 
-	const sendPrompt = async (userPrompt, parentId) => {
+	const sendPrompt = async (userPrompt, parentId, _chatId) => {
 		await Promise.all(
 			selectedModels.map(async (model) => {
 				if (model.includes('gpt-')) {
-					await sendPromptOpenAI(model, userPrompt, parentId);
+					await sendPromptOpenAI(model, userPrompt, parentId, _chatId);
 				} else {
-					await sendPromptOllama(model, userPrompt, parentId);
+					await sendPromptOllama(model, userPrompt, parentId, _chatId);
 				}
 			})
 		);
@@ -94,7 +96,7 @@
 		await chats.set(await $db.getChats());
 	};
 
-	const sendPromptOllama = async (model, userPrompt, parentId) => {
+	const sendPromptOllama = async (model, userPrompt, parentId, _chatId) => {
 		console.log('sendPromptOllama');
 		let responseMessageId = uuidv4();
 
@@ -231,7 +233,7 @@
 				window.scrollTo({ top: document.body.scrollHeight });
 			}
 
-			await $db.updateChatById($chatId, {
+			await $db.updateChatById(_chatId, {
 				title: title === '' ? 'New Chat' : title,
 				models: selectedModels,
 				system: $settings.system ?? undefined,
@@ -256,12 +258,12 @@
 		}
 
 		if (messages.length == 2 && messages.at(1).content !== '') {
-			window.history.replaceState(history.state, '', `/c/${$chatId}`);
-			await generateChatTitle($chatId, userPrompt);
+			window.history.replaceState(history.state, '', `/c/${_chatId}`);
+			await generateChatTitle(_chatId, userPrompt);
 		}
 	};
 
-	const sendPromptOpenAI = async (model, userPrompt, parentId) => {
+	const sendPromptOpenAI = async (model, userPrompt, parentId, _chatId) => {
 		if ($settings.OPENAI_API_KEY) {
 			if (models) {
 				let responseMessageId = uuidv4();
@@ -361,7 +363,7 @@
 						window.scrollTo({ top: document.body.scrollHeight });
 					}
 
-					await $db.updateChatById($chatId, {
+					await $db.updateChatById(_chatId, {
 						title: title === '' ? 'New Chat' : title,
 						models: selectedModels,
 						system: $settings.system ?? undefined,
@@ -387,15 +389,16 @@
 				}
 
 				if (messages.length == 2) {
-					window.history.replaceState(history.state, '', `/c/${$chatId}`);
-					await setChatTitle($chatId, userPrompt);
+					window.history.replaceState(history.state, '', `/c/${_chatId}`);
+					await setChatTitle(_chatId, userPrompt);
 				}
 			}
 		}
 	};
 
 	const submitPrompt = async (userPrompt) => {
-		console.log('submitPrompt');
+		const _chatId = JSON.parse(JSON.stringify($chatId));
+		console.log('submitPrompt', _chatId);
 
 		if (selectedModels.includes('')) {
 			toast.error('Model not selected');
@@ -421,12 +424,10 @@
 			history.messages[userMessageId] = userMessage;
 			history.currentId = userMessageId;
 
-			prompt = '';
-			files = [];
-
-			if (messages.length == 0) {
+			await tick();
+			if (messages.length == 1) {
 				await $db.createNewChat({
-					id: $chatId,
+					id: _chatId,
 					title: 'New Chat',
 					models: selectedModels,
 					system: $settings.system ?? undefined,
@@ -444,11 +445,14 @@
 				});
 			}
 
+			prompt = '';
+			files = [];
+
 			setTimeout(() => {
 				window.scrollTo({ top: document.body.scrollHeight, behavior: 'smooth' });
 			}, 50);
 
-			await sendPrompt(userPrompt, userMessageId);
+			await sendPrompt(userPrompt, userMessageId, _chatId);
 		}
 	};
 
@@ -458,7 +462,9 @@
 	};
 
 	const regenerateResponse = async () => {
-		console.log('regenerateResponse');
+		const _chatId = JSON.parse(JSON.stringify($chatId));
+		console.log('regenerateResponse', _chatId);
+
 		if (messages.length != 0 && messages.at(-1).done == true) {
 			messages.splice(messages.length - 1, 1);
 			messages = messages;
@@ -466,7 +472,7 @@
 			let userMessage = messages.at(-1);
 			let userPrompt = userMessage.content;
 
-			await sendPrompt(userPrompt, userMessage.id);
+			await sendPrompt(userPrompt, userMessage.id, _chatId);
 		}
 	};
 

+ 29 - 20
src/routes/(app)/c/[id]/+page.svelte

@@ -46,6 +46,8 @@
 				currentMessage.parentId !== null ? history.messages[currentMessage.parentId] : null;
 		}
 		messages = _messages;
+	} else {
+		messages = [];
 	}
 
 	// onMount(async () => {
@@ -113,13 +115,13 @@
 	// Ollama functions
 	//////////////////////////
 
-	const sendPrompt = async (userPrompt, parentId) => {
+	const sendPrompt = async (userPrompt, parentId, _chatId) => {
 		await Promise.all(
 			selectedModels.map(async (model) => {
 				if (model.includes('gpt-')) {
-					await sendPromptOpenAI(model, userPrompt, parentId);
+					await sendPromptOpenAI(model, userPrompt, parentId, _chatId);
 				} else {
-					await sendPromptOllama(model, userPrompt, parentId);
+					await sendPromptOllama(model, userPrompt, parentId, _chatId);
 				}
 			})
 		);
@@ -127,7 +129,8 @@
 		await chats.set(await $db.getChats());
 	};
 
-	const sendPromptOllama = async (model, userPrompt, parentId) => {
+	const sendPromptOllama = async (model, userPrompt, parentId, _chatId) => {
+		console.log('sendPromptOllama');
 		let responseMessageId = uuidv4();
 
 		let responseMessage = {
@@ -263,7 +266,7 @@
 				window.scrollTo({ top: document.body.scrollHeight });
 			}
 
-			await $db.updateChatById($chatId, {
+			await $db.updateChatById(_chatId, {
 				title: title === '' ? 'New Chat' : title,
 				models: selectedModels,
 				system: $settings.system ?? undefined,
@@ -288,12 +291,12 @@
 		}
 
 		if (messages.length == 2 && messages.at(1).content !== '') {
-			window.history.replaceState(history.state, '', `/c/${$chatId}`);
-			await generateChatTitle($chatId, userPrompt);
+			window.history.replaceState(history.state, '', `/c/${_chatId}`);
+			await generateChatTitle(_chatId, userPrompt);
 		}
 	};
 
-	const sendPromptOpenAI = async (model, userPrompt, parentId) => {
+	const sendPromptOpenAI = async (model, userPrompt, parentId, _chatId) => {
 		if ($settings.OPENAI_API_KEY) {
 			if (models) {
 				let responseMessageId = uuidv4();
@@ -393,7 +396,7 @@
 						window.scrollTo({ top: document.body.scrollHeight });
 					}
 
-					await $db.updateChatById($chatId, {
+					await $db.updateChatById(_chatId, {
 						title: title === '' ? 'New Chat' : title,
 						models: selectedModels,
 						system: $settings.system ?? undefined,
@@ -419,15 +422,16 @@
 				}
 
 				if (messages.length == 2) {
-					window.history.replaceState(history.state, '', `/c/${$chatId}`);
-					await setChatTitle($chatId, userPrompt);
+					window.history.replaceState(history.state, '', `/c/${_chatId}`);
+					await setChatTitle(_chatId, userPrompt);
 				}
 			}
 		}
 	};
 
 	const submitPrompt = async (userPrompt) => {
-		console.log('submitPrompt');
+		const _chatId = JSON.parse(JSON.stringify($chatId));
+		console.log('submitPrompt', _chatId);
 
 		if (selectedModels.includes('')) {
 			toast.error('Model not selected');
@@ -442,7 +446,8 @@
 				parentId: messages.length !== 0 ? messages.at(-1).id : null,
 				childrenIds: [],
 				role: 'user',
-				content: userPrompt
+				content: userPrompt,
+				files: files.length > 0 ? files : undefined
 			};
 
 			if (messages.length !== 0) {
@@ -452,11 +457,10 @@
 			history.messages[userMessageId] = userMessage;
 			history.currentId = userMessageId;
 
-			prompt = '';
-
-			if (messages.length == 0) {
+			await tick();
+			if (messages.length == 1) {
 				await $db.createNewChat({
-					id: $chatId,
+					id: _chatId,
 					title: 'New Chat',
 					models: selectedModels,
 					system: $settings.system ?? undefined,
@@ -474,11 +478,14 @@
 				});
 			}
 
+			prompt = '';
+			files = [];
+
 			setTimeout(() => {
 				window.scrollTo({ top: document.body.scrollHeight, behavior: 'smooth' });
 			}, 50);
 
-			await sendPrompt(userPrompt, userMessageId);
+			await sendPrompt(userPrompt, userMessageId, _chatId);
 		}
 	};
 
@@ -488,7 +495,9 @@
 	};
 
 	const regenerateResponse = async () => {
-		console.log('regenerateResponse');
+		const _chatId = JSON.parse(JSON.stringify($chatId));
+		console.log('regenerateResponse', _chatId);
+
 		if (messages.length != 0 && messages.at(-1).done == true) {
 			messages.splice(messages.length - 1, 1);
 			messages = messages;
@@ -496,7 +505,7 @@
 			let userMessage = messages.at(-1);
 			let userPrompt = userMessage.content;
 
-			await sendPrompt(userPrompt, userMessage.id);
+			await sendPrompt(userPrompt, userMessage.id, _chatId);
 		}
 	};