mirror of
https://gitee.com/wanwujie/deer-flow
synced 2026-04-19 12:24:46 +08:00
feat: prose completion api
This commit is contained in:
@@ -15,4 +15,5 @@ AGENT_LLM_MAP: dict[str, LLMType] = {
|
|||||||
"reporter": "basic",
|
"reporter": "basic",
|
||||||
"podcast_script_writer": "basic",
|
"podcast_script_writer": "basic",
|
||||||
"ppt_composer": "basic",
|
"ppt_composer": "basic",
|
||||||
|
"prose_writer": "basic",
|
||||||
}
|
}
|
||||||
|
|||||||
67
src/prose/graph/builder.py
Normal file
67
src/prose/graph/builder.py
Normal file
@@ -0,0 +1,67 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import asyncio
|
||||||
|
import logging
|
||||||
|
from langgraph.graph import END, START, StateGraph
|
||||||
|
|
||||||
|
from src.prose.graph.prose_continue_node import prose_continue_node
|
||||||
|
from src.prose.graph.prose_fix_node import prose_fix_node
|
||||||
|
from src.prose.graph.prose_improve_node import prose_improve_node
|
||||||
|
from src.prose.graph.prose_longer_node import prose_longer_node
|
||||||
|
from src.prose.graph.prose_shorter_node import prose_shorter_node
|
||||||
|
from src.prose.graph.prose_zap_node import prose_zap_node
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
|
||||||
|
def optional_node(state: ProseState):
|
||||||
|
return state["option"]
|
||||||
|
|
||||||
|
|
||||||
|
def build_graph():
|
||||||
|
"""Build and return the ppt workflow graph."""
|
||||||
|
# build state graph
|
||||||
|
builder = StateGraph(ProseState)
|
||||||
|
builder.add_node("prose_continue", prose_continue_node)
|
||||||
|
builder.add_node("prose_improve", prose_improve_node)
|
||||||
|
builder.add_node("prose_shorter", prose_shorter_node)
|
||||||
|
builder.add_node("prose_longer", prose_longer_node)
|
||||||
|
builder.add_node("prose_fix", prose_fix_node)
|
||||||
|
builder.add_node("prose_zap", prose_zap_node)
|
||||||
|
builder.add_conditional_edges(
|
||||||
|
START,
|
||||||
|
optional_node,
|
||||||
|
{
|
||||||
|
"continue": "prose_continue",
|
||||||
|
"improve": "prose_improve",
|
||||||
|
"shorter": "prose_shorter",
|
||||||
|
"longer": "prose_longer",
|
||||||
|
"fix": "prose_fix",
|
||||||
|
"zap": "prose_zap",
|
||||||
|
},
|
||||||
|
END,
|
||||||
|
)
|
||||||
|
return builder.compile()
|
||||||
|
|
||||||
|
|
||||||
|
async def _test_workflow():
|
||||||
|
workflow = build_graph()
|
||||||
|
events = workflow.astream(
|
||||||
|
{
|
||||||
|
"content": "The weather in Beijing is sunny",
|
||||||
|
"option": "continue",
|
||||||
|
},
|
||||||
|
stream_mode="messages",
|
||||||
|
subgraphs=True,
|
||||||
|
)
|
||||||
|
async for node, event in events:
|
||||||
|
e = event[0]
|
||||||
|
print({"id": e.id, "object": "chat.completion.chunk", "content": e.content})
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
from dotenv import load_dotenv
|
||||||
|
|
||||||
|
load_dotenv()
|
||||||
|
logging.basicConfig(level=logging.INFO)
|
||||||
|
asyncio.run(_test_workflow())
|
||||||
31
src/prose/graph/prose_continue_node.py
Normal file
31
src/prose/graph/prose_continue_node.py
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from langchain.schema import HumanMessage, SystemMessage
|
||||||
|
|
||||||
|
from src.config.agents import AGENT_LLM_MAP
|
||||||
|
from src.llms.llm import get_llm_by_type
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def prose_continue_node(state: ProseState):
|
||||||
|
logger.info("Generating prose continue content...")
|
||||||
|
model = get_llm_by_type(AGENT_LLM_MAP["prose_writer"])
|
||||||
|
prose_content = model.invoke(
|
||||||
|
[
|
||||||
|
SystemMessage(
|
||||||
|
content="""
|
||||||
|
You are an AI writing assistant that continues existing text based on context from prior text.
|
||||||
|
- Give more weight/priority to the later characters than the beginning ones.
|
||||||
|
- Limit your response to no more than 200 characters, but make sure to construct complete sentences.
|
||||||
|
- Use Markdown formatting when appropriate
|
||||||
|
"""
|
||||||
|
),
|
||||||
|
HumanMessage(content=state["content"]),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
return {"output": prose_content.content}
|
||||||
32
src/prose/graph/prose_fix_node.py
Normal file
32
src/prose/graph/prose_fix_node.py
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from langchain.schema import HumanMessage, SystemMessage
|
||||||
|
|
||||||
|
from src.config.agents import AGENT_LLM_MAP
|
||||||
|
from src.llms.llm import get_llm_by_type
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def prose_fix_node(state: ProseState):
|
||||||
|
logger.info("Generating prose fix content...")
|
||||||
|
model = get_llm_by_type(AGENT_LLM_MAP["prose_writer"])
|
||||||
|
prose_content = model.invoke(
|
||||||
|
[
|
||||||
|
SystemMessage(
|
||||||
|
content="""
|
||||||
|
You are an AI writing assistant that fixes grammar and spelling errors in existing text.
|
||||||
|
- Limit your response to no more than 200 characters, but make sure to construct complete sentences.
|
||||||
|
- Use Markdown formatting when appropriate.
|
||||||
|
- If the text is already correct, just return the original text.
|
||||||
|
"""
|
||||||
|
),
|
||||||
|
HumanMessage(content=f"The existing text is: {state['content']}"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
logger.info(f"prose_content: {prose_content}")
|
||||||
|
return {"output": prose_content.content}
|
||||||
31
src/prose/graph/prose_improve_node.py
Normal file
31
src/prose/graph/prose_improve_node.py
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from langchain.schema import HumanMessage, SystemMessage
|
||||||
|
|
||||||
|
from src.config.agents import AGENT_LLM_MAP
|
||||||
|
from src.llms.llm import get_llm_by_type
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def prose_improve_node(state: ProseState):
|
||||||
|
logger.info("Generating prose improve content...")
|
||||||
|
model = get_llm_by_type(AGENT_LLM_MAP["prose_writer"])
|
||||||
|
prose_content = model.invoke(
|
||||||
|
[
|
||||||
|
SystemMessage(
|
||||||
|
content="""
|
||||||
|
You are an AI writing assistant that improves existing text.
|
||||||
|
- Limit your response to no more than 200 characters, but make sure to construct complete sentences.
|
||||||
|
- Use Markdown formatting when appropriate.
|
||||||
|
"""
|
||||||
|
),
|
||||||
|
HumanMessage(content=f"The existing text is: {state['content']}"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
logger.info(f"prose_content: {prose_content}")
|
||||||
|
return {"output": prose_content.content}
|
||||||
30
src/prose/graph/prose_longer_node.py
Normal file
30
src/prose/graph/prose_longer_node.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from langchain.schema import HumanMessage, SystemMessage
|
||||||
|
|
||||||
|
from src.config.agents import AGENT_LLM_MAP
|
||||||
|
from src.llms.llm import get_llm_by_type
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def prose_longer_node(state: ProseState):
|
||||||
|
logger.info("Generating prose longer content...")
|
||||||
|
model = get_llm_by_type(AGENT_LLM_MAP["prose_writer"])
|
||||||
|
prose_content = model.invoke(
|
||||||
|
[
|
||||||
|
SystemMessage(
|
||||||
|
content="""
|
||||||
|
You are an AI writing assistant that lengthens existing text.
|
||||||
|
- Use Markdown formatting when appropriate.
|
||||||
|
"""
|
||||||
|
),
|
||||||
|
HumanMessage(content=f"The existing text is: {state['content']}"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
logger.info(f"prose_content: {prose_content}")
|
||||||
|
return {"output": prose_content.content}
|
||||||
30
src/prose/graph/prose_shorter_node.py
Normal file
30
src/prose/graph/prose_shorter_node.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from langchain.schema import HumanMessage, SystemMessage
|
||||||
|
|
||||||
|
from src.config.agents import AGENT_LLM_MAP
|
||||||
|
from src.llms.llm import get_llm_by_type
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def prose_shorter_node(state: ProseState):
|
||||||
|
logger.info("Generating prose shorter content...")
|
||||||
|
model = get_llm_by_type(AGENT_LLM_MAP["prose_writer"])
|
||||||
|
prose_content = model.invoke(
|
||||||
|
[
|
||||||
|
SystemMessage(
|
||||||
|
content="""
|
||||||
|
You are an AI writing assistant that shortens existing text.
|
||||||
|
- Use Markdown formatting when appropriate.
|
||||||
|
"""
|
||||||
|
),
|
||||||
|
HumanMessage(content=f"The existing text is: {state['content']}"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
logger.info(f"prose_content: {prose_content}")
|
||||||
|
return {"output": prose_content.content}
|
||||||
33
src/prose/graph/prose_zap_node.py
Normal file
33
src/prose/graph/prose_zap_node.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
import logging
|
||||||
|
|
||||||
|
from langchain.schema import HumanMessage, SystemMessage
|
||||||
|
|
||||||
|
from src.config.agents import AGENT_LLM_MAP
|
||||||
|
from src.llms.llm import get_llm_by_type
|
||||||
|
from src.prose.graph.state import ProseState
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
def prose_zap_node(state: ProseState):
|
||||||
|
logger.info("Generating prose zap content...")
|
||||||
|
model = get_llm_by_type(AGENT_LLM_MAP["prose_writer"])
|
||||||
|
prose_content = model.invoke(
|
||||||
|
[
|
||||||
|
SystemMessage(
|
||||||
|
content="""
|
||||||
|
You area an AI writing assistant that generates text based on a prompt.
|
||||||
|
- You take an input from the user and a command for manipulating the text."
|
||||||
|
- Use Markdown formatting when appropriate.
|
||||||
|
"""
|
||||||
|
),
|
||||||
|
HumanMessage(
|
||||||
|
content=f"For this text: {state['content']}.\nYou have to respect the command: {state['command']}"
|
||||||
|
),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
logger.info(f"prose_content: {prose_content}")
|
||||||
|
return {"output": prose_content.content}
|
||||||
20
src/prose/graph/state.py
Normal file
20
src/prose/graph/state.py
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
# Copyright (c) 2025 Bytedance Ltd. and/or its affiliates
|
||||||
|
# SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
|
from langgraph.graph import MessagesState
|
||||||
|
|
||||||
|
|
||||||
|
class ProseState(MessagesState):
|
||||||
|
"""State for the prose generation."""
|
||||||
|
|
||||||
|
# The content of the prose
|
||||||
|
content: str = ""
|
||||||
|
|
||||||
|
# Prose writer option: continue, improve, shorter, longer, fix, zap
|
||||||
|
option: str = ""
|
||||||
|
|
||||||
|
# The user custom command for the prose writer
|
||||||
|
command: str = ""
|
||||||
|
|
||||||
|
# Output
|
||||||
|
output: str = ""
|
||||||
@@ -17,11 +17,13 @@ from langgraph.types import Command
|
|||||||
from src.graph.builder import build_graph_with_memory
|
from src.graph.builder import build_graph_with_memory
|
||||||
from src.podcast.graph.builder import build_graph as build_podcast_graph
|
from src.podcast.graph.builder import build_graph as build_podcast_graph
|
||||||
from src.ppt.graph.builder import build_graph as build_ppt_graph
|
from src.ppt.graph.builder import build_graph as build_ppt_graph
|
||||||
|
from src.prose.graph.builder import build_graph as build_prose_graph
|
||||||
from src.server.chat_request import (
|
from src.server.chat_request import (
|
||||||
ChatMessage,
|
ChatMessage,
|
||||||
ChatRequest,
|
ChatRequest,
|
||||||
GeneratePodcastRequest,
|
GeneratePodcastRequest,
|
||||||
GeneratePPTRequest,
|
GeneratePPTRequest,
|
||||||
|
GenerateProseRequest,
|
||||||
TTSRequest,
|
TTSRequest,
|
||||||
)
|
)
|
||||||
from src.server.mcp_request import MCPServerMetadataRequest, MCPServerMetadataResponse
|
from src.server.mcp_request import MCPServerMetadataRequest, MCPServerMetadataResponse
|
||||||
@@ -254,6 +256,29 @@ async def generate_ppt(request: GeneratePPTRequest):
|
|||||||
raise HTTPException(status_code=500, detail=str(e))
|
raise HTTPException(status_code=500, detail=str(e))
|
||||||
|
|
||||||
|
|
||||||
|
@app.post("/api/prose/generate")
|
||||||
|
async def generate_prose(request: GenerateProseRequest):
|
||||||
|
try:
|
||||||
|
logger.info(f"Generating prose for prompt: {request.prompt}")
|
||||||
|
workflow = build_prose_graph()
|
||||||
|
events = workflow.astream(
|
||||||
|
{
|
||||||
|
"content": request.prompt,
|
||||||
|
"option": request.option,
|
||||||
|
"command": request.command,
|
||||||
|
},
|
||||||
|
stream_mode="messages",
|
||||||
|
subgraphs=True,
|
||||||
|
)
|
||||||
|
return StreamingResponse(
|
||||||
|
(f"data: {event[0].content}\n\n" async for _, event in events),
|
||||||
|
media_type="text/event-stream",
|
||||||
|
)
|
||||||
|
except Exception as e:
|
||||||
|
logger.exception(f"Error occurred during prose generation: {str(e)}")
|
||||||
|
raise HTTPException(status_code=500, detail=str(e))
|
||||||
|
|
||||||
|
|
||||||
@app.post("/api/mcp/server/metadata", response_model=MCPServerMetadataResponse)
|
@app.post("/api/mcp/server/metadata", response_model=MCPServerMetadataResponse)
|
||||||
async def mcp_server_metadata(request: MCPServerMetadataRequest):
|
async def mcp_server_metadata(request: MCPServerMetadataRequest):
|
||||||
"""Get information about an MCP server."""
|
"""Get information about an MCP server."""
|
||||||
|
|||||||
@@ -74,3 +74,11 @@ class GeneratePodcastRequest(BaseModel):
|
|||||||
|
|
||||||
class GeneratePPTRequest(BaseModel):
|
class GeneratePPTRequest(BaseModel):
|
||||||
content: str = Field(..., description="The content of the ppt")
|
content: str = Field(..., description="The content of the ppt")
|
||||||
|
|
||||||
|
|
||||||
|
class GenerateProseRequest(BaseModel):
|
||||||
|
prompt: str = Field(..., description="The content of the prose")
|
||||||
|
option: str = Field(..., description="The option of the prose writer")
|
||||||
|
command: Optional[str] = Field(
|
||||||
|
"", description="The user custom command of the prose writer"
|
||||||
|
)
|
||||||
|
|||||||
@@ -17,7 +17,6 @@
|
|||||||
"typecheck": "tsc --noEmit"
|
"typecheck": "tsc --noEmit"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@ai-sdk/react": "^1.2.9",
|
|
||||||
"@ant-design/icons": "^6.0.0",
|
"@ant-design/icons": "^6.0.0",
|
||||||
"@hookform/resolvers": "^5.0.1",
|
"@hookform/resolvers": "^5.0.1",
|
||||||
"@nanostores/react": "github:ai/react",
|
"@nanostores/react": "github:ai/react",
|
||||||
|
|||||||
@@ -2,11 +2,10 @@
|
|||||||
|
|
||||||
import { Command, CommandInput } from "../../ui/command";
|
import { Command, CommandInput } from "../../ui/command";
|
||||||
|
|
||||||
import { useCompletion } from "@ai-sdk/react";
|
|
||||||
import { ArrowUp } from "lucide-react";
|
import { ArrowUp } from "lucide-react";
|
||||||
import { useEditor } from "novel";
|
import { useEditor } from "novel";
|
||||||
import { addAIHighlight } from "novel";
|
import { addAIHighlight } from "novel";
|
||||||
import { useState } from "react";
|
import { useCallback, useState } from "react";
|
||||||
import Markdown from "react-markdown";
|
import Markdown from "react-markdown";
|
||||||
import { toast } from "sonner";
|
import { toast } from "sonner";
|
||||||
import { Button } from "../../ui/button";
|
import { Button } from "../../ui/button";
|
||||||
@@ -15,6 +14,8 @@ import { ScrollArea } from "../../ui/scroll-area";
|
|||||||
import AICompletionCommands from "./ai-completion-command";
|
import AICompletionCommands from "./ai-completion-command";
|
||||||
import AISelectorCommands from "./ai-selector-commands";
|
import AISelectorCommands from "./ai-selector-commands";
|
||||||
import { LoadingOutlined } from "@ant-design/icons";
|
import { LoadingOutlined } from "@ant-design/icons";
|
||||||
|
import { resolveServiceURL } from "~/core/api/resolve-service-url";
|
||||||
|
import { fetchStream } from "~/core/sse";
|
||||||
//TODO: I think it makes more sense to create a custom Tiptap extension for this functionality https://tiptap.dev/docs/editor/ai/introduction
|
//TODO: I think it makes more sense to create a custom Tiptap extension for this functionality https://tiptap.dev/docs/editor/ai/introduction
|
||||||
|
|
||||||
interface AISelectorProps {
|
interface AISelectorProps {
|
||||||
@@ -22,23 +23,72 @@ interface AISelectorProps {
|
|||||||
onOpenChange: (open: boolean) => void;
|
onOpenChange: (open: boolean) => void;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
function useProseCompletion() {
|
||||||
|
const [completion, setCompletion] = useState("");
|
||||||
|
const [isLoading, setIsLoading] = useState(false);
|
||||||
|
const [error, setError] = useState<Error | null>(null);
|
||||||
|
|
||||||
|
const complete = useCallback(
|
||||||
|
async (prompt: string, options?: { body?: Record<string, any> }) => {
|
||||||
|
setIsLoading(true);
|
||||||
|
setError(null);
|
||||||
|
|
||||||
|
try {
|
||||||
|
const response = await fetchStream(
|
||||||
|
resolveServiceURL("/api/prose/generate"),
|
||||||
|
{
|
||||||
|
method: "POST",
|
||||||
|
headers: {
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
},
|
||||||
|
body: JSON.stringify({
|
||||||
|
prompt,
|
||||||
|
...options?.body,
|
||||||
|
}),
|
||||||
|
},
|
||||||
|
);
|
||||||
|
|
||||||
|
let fullText = "";
|
||||||
|
|
||||||
|
// Process the streaming response
|
||||||
|
for await (const chunk of response) {
|
||||||
|
fullText += chunk.data;
|
||||||
|
setCompletion(fullText);
|
||||||
|
}
|
||||||
|
|
||||||
|
setIsLoading(false);
|
||||||
|
return fullText;
|
||||||
|
} catch (e) {
|
||||||
|
const error = e instanceof Error ? e : new Error("An error occurred");
|
||||||
|
setError(error);
|
||||||
|
toast.error(error.message);
|
||||||
|
setIsLoading(false);
|
||||||
|
throw error;
|
||||||
|
}
|
||||||
|
},
|
||||||
|
[],
|
||||||
|
);
|
||||||
|
|
||||||
|
const reset = useCallback(() => {
|
||||||
|
setCompletion("");
|
||||||
|
setError(null);
|
||||||
|
setIsLoading(false);
|
||||||
|
}, []);
|
||||||
|
|
||||||
|
return {
|
||||||
|
completion,
|
||||||
|
complete,
|
||||||
|
isLoading,
|
||||||
|
error,
|
||||||
|
reset,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
export function AISelector({ onOpenChange }: AISelectorProps) {
|
export function AISelector({ onOpenChange }: AISelectorProps) {
|
||||||
const { editor } = useEditor();
|
const { editor } = useEditor();
|
||||||
const [inputValue, setInputValue] = useState("");
|
const [inputValue, setInputValue] = useState("");
|
||||||
|
|
||||||
const { completion, complete, isLoading } = useCompletion({
|
const { completion, complete, isLoading } = useProseCompletion();
|
||||||
// id: "novel",
|
|
||||||
api: "/api/generate",
|
|
||||||
onResponse: (response) => {
|
|
||||||
if (response.status === 429) {
|
|
||||||
toast.error("You have reached your request limit for the day.");
|
|
||||||
return;
|
|
||||||
}
|
|
||||||
},
|
|
||||||
onError: (e) => {
|
|
||||||
toast.error(e.message);
|
|
||||||
},
|
|
||||||
});
|
|
||||||
|
|
||||||
if (!editor) return null;
|
if (!editor) return null;
|
||||||
|
|
||||||
@@ -57,7 +107,7 @@ export function AISelector({ onOpenChange }: AISelectorProps) {
|
|||||||
)}
|
)}
|
||||||
|
|
||||||
{isLoading && (
|
{isLoading && (
|
||||||
<div className="text-muted-foreground flex h-12 w-full items-center px-4 text-sm font-medium text-purple-500">
|
<div className="flex h-12 w-full items-center px-4 text-sm font-medium text-purple-500">
|
||||||
<Magic className="mr-2 h-4 w-4 shrink-0" />
|
<Magic className="mr-2 h-4 w-4 shrink-0" />
|
||||||
AI is thinking
|
AI is thinking
|
||||||
<div className="mt-1 ml-2">
|
<div className="mt-1 ml-2">
|
||||||
|
|||||||
@@ -9,7 +9,7 @@ import { sleep } from "../utils";
|
|||||||
import { resolveServiceURL } from "./resolve-service-url";
|
import { resolveServiceURL } from "./resolve-service-url";
|
||||||
import type { ChatEvent } from "./types";
|
import type { ChatEvent } from "./types";
|
||||||
|
|
||||||
export function chatStream(
|
export async function* chatStream(
|
||||||
userMessage: string,
|
userMessage: string,
|
||||||
params: {
|
params: {
|
||||||
thread_id: string;
|
thread_id: string;
|
||||||
@@ -32,13 +32,19 @@ export function chatStream(
|
|||||||
if (location.search.includes("mock") || location.search.includes("replay=")) {
|
if (location.search.includes("mock") || location.search.includes("replay=")) {
|
||||||
return chatReplayStream(userMessage, params, options);
|
return chatReplayStream(userMessage, params, options);
|
||||||
}
|
}
|
||||||
return fetchStream<ChatEvent>(resolveServiceURL("chat/stream"), {
|
const stream = fetchStream(resolveServiceURL("chat/stream"), {
|
||||||
body: JSON.stringify({
|
body: JSON.stringify({
|
||||||
messages: [{ role: "user", content: userMessage }],
|
messages: [{ role: "user", content: userMessage }],
|
||||||
...params,
|
...params,
|
||||||
}),
|
}),
|
||||||
signal: options.abortSignal,
|
signal: options.abortSignal,
|
||||||
});
|
});
|
||||||
|
for await (const event of stream) {
|
||||||
|
yield {
|
||||||
|
type: event.event,
|
||||||
|
data: JSON.parse(event.data),
|
||||||
|
} as ChatEvent;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
async function* chatReplayStream(
|
async function* chatReplayStream(
|
||||||
|
|||||||
@@ -2,6 +2,6 @@
|
|||||||
// SPDX-License-Identifier: MIT
|
// SPDX-License-Identifier: MIT
|
||||||
|
|
||||||
export interface StreamEvent {
|
export interface StreamEvent {
|
||||||
type: string;
|
event: string;
|
||||||
data: object;
|
data: string;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,10 +3,10 @@
|
|||||||
|
|
||||||
import { type StreamEvent } from "./StreamEvent";
|
import { type StreamEvent } from "./StreamEvent";
|
||||||
|
|
||||||
export async function* fetchStream<T extends StreamEvent>(
|
export async function* fetchStream(
|
||||||
url: string,
|
url: string,
|
||||||
init: RequestInit,
|
init: RequestInit,
|
||||||
): AsyncIterable<T> {
|
): AsyncIterable<StreamEvent> {
|
||||||
const response = await fetch(url, {
|
const response = await fetch(url, {
|
||||||
method: "POST",
|
method: "POST",
|
||||||
headers: {
|
headers: {
|
||||||
@@ -39,7 +39,7 @@ export async function* fetchStream<T extends StreamEvent>(
|
|||||||
}
|
}
|
||||||
const chunk = buffer.slice(0, index);
|
const chunk = buffer.slice(0, index);
|
||||||
buffer = buffer.slice(index + 2);
|
buffer = buffer.slice(index + 2);
|
||||||
const event = parseEvent<T>(chunk);
|
const event = parseEvent(chunk);
|
||||||
if (event) {
|
if (event) {
|
||||||
yield event;
|
yield event;
|
||||||
}
|
}
|
||||||
@@ -47,9 +47,9 @@ export async function* fetchStream<T extends StreamEvent>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function parseEvent<T extends StreamEvent>(chunk: string) {
|
function parseEvent(chunk: string) {
|
||||||
let resultType = "message";
|
let resultEvent = "message";
|
||||||
let resultData: object | null = null;
|
let resultData: string | null = null;
|
||||||
for (const line of chunk.split("\n")) {
|
for (const line of chunk.split("\n")) {
|
||||||
const pos = line.indexOf(": ");
|
const pos = line.indexOf(": ");
|
||||||
if (pos === -1) {
|
if (pos === -1) {
|
||||||
@@ -58,16 +58,16 @@ function parseEvent<T extends StreamEvent>(chunk: string) {
|
|||||||
const key = line.slice(0, pos);
|
const key = line.slice(0, pos);
|
||||||
const value = line.slice(pos + 2);
|
const value = line.slice(pos + 2);
|
||||||
if (key === "event") {
|
if (key === "event") {
|
||||||
resultType = value;
|
resultEvent = value;
|
||||||
} else if (key === "data") {
|
} else if (key === "data") {
|
||||||
resultData = JSON.parse(value);
|
resultData = value;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (resultType === "message" && resultData === null) {
|
if (resultEvent === "message" && resultData === null) {
|
||||||
return undefined;
|
return undefined;
|
||||||
}
|
}
|
||||||
return {
|
return {
|
||||||
type: resultType,
|
event: resultEvent,
|
||||||
data: resultData,
|
data: resultData,
|
||||||
} as T;
|
} as StreamEvent;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user