chat.ts 21 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729
  1. import { trimTopic, getMessageTextContent } from "../utils";
  2. import Locale, { getLang } from "../locales";
  3. import { showToast } from "../components/ui-lib";
  4. import { ModelConfig, ModelType, useAppConfig } from "./config";
  5. import { createEmptyMask, Mask } from "./mask";
  6. import {
  7. DEFAULT_INPUT_TEMPLATE,
  8. DEFAULT_MODELS,
  9. DEFAULT_SYSTEM_TEMPLATE,
  10. KnowledgeCutOffDate,
  11. StoreKey,
  12. SUMMARIZE_MODEL,
  13. GEMINI_SUMMARIZE_MODEL,
  14. ServiceProvider,
  15. } from "../constant";
  16. import { getClientApi } from "../client/api";
  17. import type {
  18. ClientApi,
  19. RequestMessage,
  20. MultimodalContent,
  21. } from "../client/api";
  22. import { ChatControllerPool } from "../client/controller";
  23. import { prettyObject } from "../utils/format";
  24. import { estimateTokenLength } from "../utils/token";
  25. import { nanoid } from "nanoid";
  26. import { createPersistStore } from "../utils/store";
  27. import { collectModelsWithDefaultModel } from "../utils/model";
  28. import { useAccessStore } from "./access";
  29. import { isDalle3 } from "../utils";
  30. export type ChatMessage = RequestMessage & {
  31. date: string;
  32. streaming?: boolean;
  33. isError?: boolean;
  34. id: string;
  35. model?: ModelType;
  36. };
  37. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  38. return {
  39. id: nanoid(),
  40. date: new Date().toLocaleString(),
  41. role: "user",
  42. content: "",
  43. ...override,
  44. };
  45. }
  46. export interface ChatStat {
  47. tokenCount: number;
  48. wordCount: number;
  49. charCount: number;
  50. }
  51. export interface ChatSession {
  52. id: string;
  53. topic: string;
  54. memoryPrompt: string;
  55. messages: ChatMessage[];
  56. stat: ChatStat;
  57. lastUpdate: number;
  58. lastSummarizeIndex: number;
  59. clearContextIndex?: number;
  60. mask: Mask;
  61. }
  62. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  63. export const BOT_HELLO: ChatMessage = createMessage({
  64. role: "assistant",
  65. content: '我是「建科招聘智能助理」,请问您有什么招聘、薪酬或工作相关的问题需要咨询?我会尽力为您提供帮助。',
  66. });
  67. function createEmptySession(): ChatSession {
  68. return {
  69. id: nanoid(),
  70. topic: DEFAULT_TOPIC,
  71. memoryPrompt: "",
  72. messages: [],
  73. stat: {
  74. tokenCount: 0,
  75. wordCount: 0,
  76. charCount: 0,
  77. },
  78. lastUpdate: Date.now(),
  79. lastSummarizeIndex: 0,
  80. mask: createEmptyMask(),
  81. };
  82. }
  83. function getSummarizeModel(currentModel: string) {
  84. // if it is using gpt-* models, force to use 4o-mini to summarize
  85. if (currentModel.startsWith("gpt")) {
  86. const configStore = useAppConfig.getState();
  87. const accessStore = useAccessStore.getState();
  88. const allModel = collectModelsWithDefaultModel(
  89. configStore.models,
  90. [configStore.customModels, accessStore.customModels].join(","),
  91. accessStore.defaultModel,
  92. );
  93. const summarizeModel = allModel.find(
  94. (m) => m.name === SUMMARIZE_MODEL && m.available,
  95. );
  96. return summarizeModel?.name ?? currentModel;
  97. }
  98. if (currentModel.startsWith("gemini")) {
  99. return GEMINI_SUMMARIZE_MODEL;
  100. }
  101. return currentModel;
  102. }
  103. function countMessages(msgs: ChatMessage[]) {
  104. return msgs.reduce(
  105. (pre, cur) => pre + estimateTokenLength(getMessageTextContent(cur)),
  106. 0,
  107. );
  108. }
  109. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  110. const cutoff =
  111. KnowledgeCutOffDate[modelConfig.model] ?? KnowledgeCutOffDate.default;
  112. // Find the model in the DEFAULT_MODELS array that matches the modelConfig.model
  113. const modelInfo = DEFAULT_MODELS.find((m) => m.name === modelConfig.model);
  114. var serviceProvider = "OpenAI";
  115. if (modelInfo) {
  116. // TODO: auto detect the providerName from the modelConfig.model
  117. // Directly use the providerName from the modelInfo
  118. serviceProvider = modelInfo.provider.providerName;
  119. }
  120. const vars = {
  121. ServiceProvider: serviceProvider,
  122. cutoff,
  123. model: modelConfig.model,
  124. time: new Date().toString(),
  125. lang: getLang(),
  126. input: input,
  127. };
  128. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  129. // remove duplicate
  130. if (input.startsWith(output)) {
  131. output = "";
  132. }
  133. // must contains {{input}}
  134. const inputVar = "{{input}}";
  135. if (!output.includes(inputVar)) {
  136. output += "\n" + inputVar;
  137. }
  138. Object.entries(vars).forEach(([name, value]) => {
  139. const regex = new RegExp(`{{${name}}}`, "g");
  140. output = output.replace(regex, value.toString()); // Ensure value is a string
  141. });
  142. return output;
  143. }
  144. const DEFAULT_CHAT_STATE = {
  145. sessions: [createEmptySession()],
  146. currentSessionIndex: 0,
  147. };
  148. export const useChatStore = createPersistStore(
  149. DEFAULT_CHAT_STATE,
  150. (set, _get) => {
  151. function get() {
  152. return {
  153. ..._get(),
  154. ...methods,
  155. };
  156. }
  157. const methods = {
  158. clearSessions() {
  159. set(() => ({
  160. sessions: [createEmptySession()],
  161. currentSessionIndex: 0,
  162. }));
  163. },
  164. selectSession(index: number) {
  165. set({
  166. currentSessionIndex: index,
  167. });
  168. },
  169. moveSession(from: number, to: number) {
  170. set((state) => {
  171. const { sessions, currentSessionIndex: oldIndex } = state;
  172. // move the session
  173. const newSessions = [...sessions];
  174. const session = newSessions[from];
  175. newSessions.splice(from, 1);
  176. newSessions.splice(to, 0, session);
  177. // modify current session id
  178. let newIndex = oldIndex === from ? to : oldIndex;
  179. if (oldIndex > from && oldIndex <= to) {
  180. newIndex -= 1;
  181. } else if (oldIndex < from && oldIndex >= to) {
  182. newIndex += 1;
  183. }
  184. return {
  185. currentSessionIndex: newIndex,
  186. sessions: newSessions,
  187. };
  188. });
  189. },
  190. newSession(mask?: Mask) {
  191. const session = createEmptySession();
  192. if (mask) {
  193. const config = useAppConfig.getState();
  194. const globalModelConfig = config.modelConfig;
  195. session.mask = {
  196. ...mask,
  197. modelConfig: {
  198. ...globalModelConfig,
  199. ...mask.modelConfig,
  200. },
  201. };
  202. session.topic = mask.name;
  203. }
  204. set((state) => ({
  205. currentSessionIndex: 0,
  206. sessions: [session].concat(state.sessions),
  207. }));
  208. },
  209. nextSession(delta: number) {
  210. const n = get().sessions.length;
  211. const limit = (x: number) => (x + n) % n;
  212. const i = get().currentSessionIndex;
  213. get().selectSession(limit(i + delta));
  214. },
  215. deleteSession(index: number) {
  216. const deletingLastSession = get().sessions.length === 1;
  217. const deletedSession = get().sessions.at(index);
  218. if (!deletedSession) return;
  219. const sessions = get().sessions.slice();
  220. sessions.splice(index, 1);
  221. const currentIndex = get().currentSessionIndex;
  222. let nextIndex = Math.min(
  223. currentIndex - Number(index < currentIndex),
  224. sessions.length - 1,
  225. );
  226. if (deletingLastSession) {
  227. nextIndex = 0;
  228. sessions.push(createEmptySession());
  229. }
  230. // for undo delete action
  231. const restoreState = {
  232. currentSessionIndex: get().currentSessionIndex,
  233. sessions: get().sessions.slice(),
  234. };
  235. set(() => ({
  236. currentSessionIndex: nextIndex,
  237. sessions,
  238. }));
  239. showToast(
  240. Locale.Home.DeleteToast,
  241. {
  242. text: Locale.Home.Revert,
  243. onClick() {
  244. set(() => restoreState);
  245. },
  246. },
  247. 5000,
  248. );
  249. },
  250. currentSession() {
  251. let index = get().currentSessionIndex;
  252. const sessions = get().sessions;
  253. if (index < 0 || index >= sessions.length) {
  254. index = Math.min(sessions.length - 1, Math.max(0, index));
  255. set(() => ({ currentSessionIndex: index }));
  256. }
  257. const session = sessions[index];
  258. return session;
  259. },
  260. onNewMessage(message: ChatMessage) {
  261. get().updateCurrentSession((session) => {
  262. session.messages = session.messages.concat();
  263. session.lastUpdate = Date.now();
  264. });
  265. get().updateStat(message);
  266. get().summarizeSession();
  267. },
  268. async onUserInput(content: string, attachImages?: string[]) {
  269. const session = get().currentSession();
  270. const modelConfig = session.mask.modelConfig;
  271. const userContent = fillTemplateWith(content, modelConfig);
  272. console.log("[User Input] after template: ", userContent);
  273. let mContent: string | MultimodalContent[] = userContent;
  274. if (attachImages && attachImages.length > 0) {
  275. mContent = [
  276. {
  277. type: "text",
  278. text: userContent,
  279. },
  280. ];
  281. mContent = mContent.concat(
  282. attachImages.map((url) => {
  283. return {
  284. type: "image_url",
  285. image_url: {
  286. url: url,
  287. },
  288. };
  289. }),
  290. );
  291. }
  292. let userMessage: ChatMessage = createMessage({
  293. role: "user",
  294. content: mContent,
  295. });
  296. const botMessage: ChatMessage = createMessage({
  297. role: "assistant",
  298. streaming: true,
  299. model: modelConfig.model,
  300. });
  301. // get recent messages
  302. const recentMessages = get().getMessagesWithMemory();
  303. const sendMessages = recentMessages.concat(userMessage);
  304. const messageIndex = get().currentSession().messages.length + 1;
  305. // save user's and bot's message
  306. get().updateCurrentSession((session) => {
  307. const savedUserMessage = {
  308. ...userMessage,
  309. content: mContent,
  310. };
  311. session.messages = session.messages.concat([
  312. savedUserMessage,
  313. botMessage,
  314. ]);
  315. });
  316. // 固定使用大模型
  317. const providerName = 'BigModel' as ServiceProvider;
  318. const api: ClientApi = getClientApi(providerName);
  319. api.llm.chat({
  320. messages: sendMessages,
  321. config: { ...modelConfig, stream: true },
  322. onUpdate(message) {
  323. botMessage.streaming = true;
  324. if (message) {
  325. botMessage.content = message;
  326. }
  327. get().updateCurrentSession((session) => {
  328. session.messages = session.messages.concat();
  329. });
  330. },
  331. onFinish(message) {
  332. botMessage.streaming = false;
  333. if (message) {
  334. botMessage.content = message;
  335. get().onNewMessage(botMessage);
  336. }
  337. ChatControllerPool.remove(session.id, botMessage.id);
  338. },
  339. onError(error) {
  340. const isAborted = error.message.includes("aborted");
  341. botMessage.content +=
  342. "\n\n" +
  343. prettyObject({
  344. error: true,
  345. message: error.message,
  346. });
  347. botMessage.streaming = false;
  348. userMessage.isError = !isAborted;
  349. botMessage.isError = !isAborted;
  350. get().updateCurrentSession((session) => {
  351. session.messages = session.messages.concat();
  352. });
  353. ChatControllerPool.remove(
  354. session.id,
  355. botMessage.id ?? messageIndex,
  356. );
  357. console.error("[Chat] failed ", error);
  358. },
  359. onController(controller) {
  360. // collect controller for stop/retry
  361. ChatControllerPool.addController(
  362. session.id,
  363. botMessage.id ?? messageIndex,
  364. controller,
  365. );
  366. },
  367. });
  368. },
  369. getMemoryPrompt() {
  370. const session = get().currentSession();
  371. if (session.memoryPrompt.length) {
  372. return {
  373. role: "system",
  374. content: Locale.Store.Prompt.History(session.memoryPrompt),
  375. date: "",
  376. } as ChatMessage;
  377. }
  378. },
  379. getMessagesWithMemory() {
  380. const session = get().currentSession();
  381. const modelConfig = session.mask.modelConfig;
  382. const clearContextIndex = session.clearContextIndex ?? 0;
  383. const messages = session.messages.slice();
  384. const totalMessageCount = session.messages.length;
  385. // in-context prompts
  386. const contextPrompts = session.mask.context.slice();
  387. // system prompts, to get close to OpenAI Web ChatGPT
  388. const shouldInjectSystemPrompts =
  389. modelConfig.enableInjectSystemPrompts &&
  390. session.mask.modelConfig.model.startsWith("gpt-");
  391. var systemPrompts: ChatMessage[] = [];
  392. systemPrompts = shouldInjectSystemPrompts
  393. ? [
  394. createMessage({
  395. role: "system",
  396. content: fillTemplateWith("", {
  397. ...modelConfig,
  398. template: DEFAULT_SYSTEM_TEMPLATE,
  399. }),
  400. }),
  401. ]
  402. : [];
  403. if (shouldInjectSystemPrompts) {
  404. console.log(
  405. "[Global System Prompt] ",
  406. systemPrompts.at(0)?.content ?? "empty",
  407. );
  408. }
  409. const memoryPrompt = get().getMemoryPrompt();
  410. // long term memory
  411. const shouldSendLongTermMemory =
  412. modelConfig.sendMemory &&
  413. session.memoryPrompt &&
  414. session.memoryPrompt.length > 0 &&
  415. session.lastSummarizeIndex > clearContextIndex;
  416. const longTermMemoryPrompts =
  417. shouldSendLongTermMemory && memoryPrompt ? [memoryPrompt] : [];
  418. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  419. // short term memory
  420. const shortTermMemoryStartIndex = Math.max(
  421. 0,
  422. totalMessageCount - modelConfig.historyMessageCount,
  423. );
  424. // lets concat send messages, including 4 parts:
  425. // 0. system prompt: to get close to OpenAI Web ChatGPT
  426. // 1. long term memory: summarized memory messages
  427. // 2. pre-defined in-context prompts
  428. // 3. short term memory: latest n messages
  429. // 4. newest input message
  430. const memoryStartIndex = shouldSendLongTermMemory
  431. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  432. : shortTermMemoryStartIndex;
  433. // and if user has cleared history messages, we should exclude the memory too.
  434. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  435. const maxTokenThreshold = modelConfig.max_tokens;
  436. // get recent messages as much as possible
  437. const reversedRecentMessages = [];
  438. for (
  439. let i = totalMessageCount - 1, tokenCount = 0;
  440. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  441. i -= 1
  442. ) {
  443. const msg = messages[i];
  444. if (!msg || msg.isError) continue;
  445. tokenCount += estimateTokenLength(getMessageTextContent(msg));
  446. reversedRecentMessages.push(msg);
  447. }
  448. // concat all messages
  449. const recentMessages = [
  450. ...systemPrompts,
  451. ...longTermMemoryPrompts,
  452. ...contextPrompts,
  453. ...reversedRecentMessages.reverse(),
  454. ];
  455. return recentMessages;
  456. },
  457. updateMessage(
  458. sessionIndex: number,
  459. messageIndex: number,
  460. updater: (message?: ChatMessage) => void,
  461. ) {
  462. const sessions = get().sessions;
  463. const session = sessions.at(sessionIndex);
  464. const messages = session?.messages;
  465. updater(messages?.at(messageIndex));
  466. set(() => ({ sessions }));
  467. },
  468. resetSession() {
  469. get().updateCurrentSession((session) => {
  470. session.messages = [];
  471. session.memoryPrompt = "";
  472. });
  473. },
  474. summarizeSession() {
  475. const config = useAppConfig.getState();
  476. const session = get().currentSession();
  477. const modelConfig = session.mask.modelConfig;
  478. // skip summarize when using dalle3?
  479. if (isDalle3(modelConfig.model)) {
  480. return;
  481. }
  482. const providerName = modelConfig.providerName;
  483. const api: ClientApi = getClientApi(providerName);
  484. // remove error messages if any
  485. const messages = session.messages;
  486. // should summarize topic after chating more than 50 words
  487. const SUMMARIZE_MIN_LEN = 50;
  488. if (
  489. config.enableAutoGenerateTitle &&
  490. session.topic === DEFAULT_TOPIC &&
  491. countMessages(messages) >= SUMMARIZE_MIN_LEN
  492. ) {
  493. const topicMessages = messages.concat(
  494. createMessage({
  495. role: "user",
  496. content: Locale.Store.Prompt.Topic,
  497. }),
  498. );
  499. api.llm.chat({
  500. messages: topicMessages,
  501. config: {
  502. model: getSummarizeModel(session.mask.modelConfig.model),
  503. stream: false,
  504. providerName,
  505. },
  506. onFinish(message) {
  507. get().updateCurrentSession(
  508. (session) =>
  509. (session.topic =
  510. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  511. );
  512. },
  513. });
  514. }
  515. const summarizeIndex = Math.max(
  516. session.lastSummarizeIndex,
  517. session.clearContextIndex ?? 0,
  518. );
  519. let toBeSummarizedMsgs = messages
  520. .filter((msg) => !msg.isError)
  521. .slice(summarizeIndex);
  522. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  523. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  524. const n = toBeSummarizedMsgs.length;
  525. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  526. Math.max(0, n - modelConfig.historyMessageCount),
  527. );
  528. }
  529. const memoryPrompt = get().getMemoryPrompt();
  530. if (memoryPrompt) {
  531. // add memory prompt
  532. toBeSummarizedMsgs.unshift(memoryPrompt);
  533. }
  534. const lastSummarizeIndex = session.messages.length;
  535. console.log(
  536. "[Chat History] ",
  537. toBeSummarizedMsgs,
  538. historyMsgLength,
  539. modelConfig.compressMessageLengthThreshold,
  540. );
  541. if (
  542. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  543. modelConfig.sendMemory
  544. ) {
  545. /** Destruct max_tokens while summarizing
  546. * this param is just shit
  547. **/
  548. const { max_tokens, ...modelcfg } = modelConfig;
  549. api.llm.chat({
  550. messages: toBeSummarizedMsgs.concat(
  551. createMessage({
  552. role: "system",
  553. content: Locale.Store.Prompt.Summarize,
  554. date: "",
  555. }),
  556. ),
  557. config: {
  558. ...modelcfg,
  559. stream: true,
  560. model: getSummarizeModel(session.mask.modelConfig.model),
  561. },
  562. onUpdate(message) {
  563. session.memoryPrompt = message;
  564. },
  565. onFinish(message) {
  566. console.log("[Memory] ", message);
  567. get().updateCurrentSession((session) => {
  568. session.lastSummarizeIndex = lastSummarizeIndex;
  569. session.memoryPrompt = message; // Update the memory prompt for stored it in local storage
  570. });
  571. },
  572. onError(err) {
  573. console.error("[Summarize] ", err);
  574. },
  575. });
  576. }
  577. },
  578. updateStat(message: ChatMessage) {
  579. get().updateCurrentSession((session) => {
  580. session.stat.charCount += message.content.length;
  581. // TODO: should update chat count and word count
  582. });
  583. },
  584. updateCurrentSession(updater: (session: ChatSession) => void) {
  585. const sessions = get().sessions;
  586. const index = get().currentSessionIndex;
  587. updater(sessions[index]);
  588. set(() => ({ sessions }));
  589. },
  590. clearAllData() {
  591. localStorage.clear();
  592. location.reload();
  593. },
  594. };
  595. return methods;
  596. },
  597. {
  598. name: StoreKey.Chat,
  599. version: 3.1,
  600. migrate(persistedState, version) {
  601. const state = persistedState as any;
  602. const newState = JSON.parse(
  603. JSON.stringify(state),
  604. ) as typeof DEFAULT_CHAT_STATE;
  605. if (version < 2) {
  606. newState.sessions = [];
  607. const oldSessions = state.sessions;
  608. for (const oldSession of oldSessions) {
  609. const newSession = createEmptySession();
  610. newSession.topic = oldSession.topic;
  611. newSession.messages = [...oldSession.messages];
  612. newSession.mask.modelConfig.sendMemory = true;
  613. newSession.mask.modelConfig.historyMessageCount = 4;
  614. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  615. newState.sessions.push(newSession);
  616. }
  617. }
  618. if (version < 3) {
  619. // migrate id to nanoid
  620. newState.sessions.forEach((s) => {
  621. s.id = nanoid();
  622. s.messages.forEach((m) => (m.id = nanoid()));
  623. });
  624. }
  625. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  626. // Resolve issue of old sessions not automatically enabling.
  627. if (version < 3.1) {
  628. newState.sessions.forEach((s) => {
  629. if (
  630. // Exclude those already set by user
  631. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  632. ) {
  633. // Because users may have changed this configuration,
  634. // the user's current configuration is used instead of the default
  635. const config = useAppConfig.getState();
  636. s.mask.modelConfig.enableInjectSystemPrompts =
  637. config.modelConfig.enableInjectSystemPrompts;
  638. }
  639. });
  640. }
  641. return newState as any;
  642. },
  643. },
  644. );