|
@@ -67,7 +67,7 @@
|
|
const textCompletionHandler = async () => {
|
|
const textCompletionHandler = async () => {
|
|
const model = $models.find((model) => model.id === selectedModelId);
|
|
const model = $models.find((model) => model.id === selectedModelId);
|
|
|
|
|
|
- const res = await generateOpenAIChatCompletion(
|
|
|
|
|
|
+ const [res, controller] = await generateOpenAIChatCompletion(
|
|
localStorage.token,
|
|
localStorage.token,
|
|
{
|
|
{
|
|
model: model.id,
|
|
model: model.id,
|
|
@@ -96,7 +96,7 @@
|
|
const { value, done } = await reader.read();
|
|
const { value, done } = await reader.read();
|
|
if (done || stopResponseFlag) {
|
|
if (done || stopResponseFlag) {
|
|
if (stopResponseFlag) {
|
|
if (stopResponseFlag) {
|
|
- await cancelOllamaRequest(localStorage.token, currentRequestId);
|
|
|
|
|
|
+ controller.abort('User: Stop Response');
|
|
}
|
|
}
|
|
|
|
|
|
currentRequestId = null;
|
|
currentRequestId = null;
|
|
@@ -135,7 +135,7 @@
|
|
const chatCompletionHandler = async () => {
|
|
const chatCompletionHandler = async () => {
|
|
const model = $models.find((model) => model.id === selectedModelId);
|
|
const model = $models.find((model) => model.id === selectedModelId);
|
|
|
|
|
|
- const res = await generateOpenAIChatCompletion(
|
|
|
|
|
|
+ const [res, controller] = await generateOpenAIChatCompletion(
|
|
localStorage.token,
|
|
localStorage.token,
|
|
{
|
|
{
|
|
model: model.id,
|
|
model: model.id,
|
|
@@ -182,7 +182,7 @@
|
|
const { value, done } = await reader.read();
|
|
const { value, done } = await reader.read();
|
|
if (done || stopResponseFlag) {
|
|
if (done || stopResponseFlag) {
|
|
if (stopResponseFlag) {
|
|
if (stopResponseFlag) {
|
|
- await cancelOllamaRequest(localStorage.token, currentRequestId);
|
|
|
|
|
|
+ controller.abort('User: Stop Response');
|
|
}
|
|
}
|
|
|
|
|
|
currentRequestId = null;
|
|
currentRequestId = null;
|