chat.ts 19 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664
  1. import { trimTopic } from "../utils";
  2. import Locale, { getLang } from "../locales";
  3. import { showToast } from "../components/ui-lib";
  4. import { ModelConfig, ModelType, useAppConfig } from "./config";
  5. import { createEmptyMask, Mask } from "./mask";
  6. import {
  7. DEFAULT_INPUT_TEMPLATE,
  8. DEFAULT_SYSTEM_TEMPLATE,
  9. KnowledgeCutOffDate,
  10. ModelProvider,
  11. StoreKey,
  12. SUMMARIZE_MODEL,
  13. } from "../constant";
  14. import { ClientApi, RequestMessage } from "../client/api";
  15. import { ChatControllerPool } from "../client/controller";
  16. import { prettyObject } from "../utils/format";
  17. import { estimateTokenLength } from "../utils/token";
  18. import { nanoid } from "nanoid";
  19. import { createPersistStore } from "../utils/store";
  20. export type ChatMessage = RequestMessage & {
  21. date: string;
  22. streaming?: boolean;
  23. isError?: boolean;
  24. id: string;
  25. model?: ModelType;
  26. };
  27. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  28. return {
  29. id: nanoid(),
  30. date: new Date().toLocaleString(),
  31. role: "user",
  32. content: "",
  33. ...override,
  34. };
  35. }
  36. export interface ChatStat {
  37. tokenCount: number;
  38. wordCount: number;
  39. charCount: number;
  40. }
  41. export interface ChatSession {
  42. id: string;
  43. topic: string;
  44. memoryPrompt: string;
  45. messages: ChatMessage[];
  46. stat: ChatStat;
  47. lastUpdate: number;
  48. lastSummarizeIndex: number;
  49. clearContextIndex?: number;
  50. mask: Mask;
  51. }
  52. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  53. export const BOT_HELLO: ChatMessage = createMessage({
  54. role: "assistant",
  55. content: Locale.Store.BotHello,
  56. });
  57. function createEmptySession(): ChatSession {
  58. return {
  59. id: nanoid(),
  60. topic: DEFAULT_TOPIC,
  61. memoryPrompt: "",
  62. messages: [],
  63. stat: {
  64. tokenCount: 0,
  65. wordCount: 0,
  66. charCount: 0,
  67. },
  68. lastUpdate: Date.now(),
  69. lastSummarizeIndex: 0,
  70. mask: createEmptyMask(),
  71. };
  72. }
  73. function getSummarizeModel(currentModel: string) {
  74. // if it is using gpt-* models, force to use 3.5 to summarize
  75. return currentModel.startsWith("gpt") ? SUMMARIZE_MODEL : currentModel;
  76. }
  77. function countMessages(msgs: ChatMessage[]) {
  78. return msgs.reduce((pre, cur) => pre + estimateTokenLength(cur.content), 0);
  79. }
  80. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  81. let cutoff =
  82. KnowledgeCutOffDate[modelConfig.model] ?? KnowledgeCutOffDate.default;
  83. const vars = {
  84. cutoff,
  85. model: modelConfig.model,
  86. time: new Date().toLocaleString(),
  87. lang: getLang(),
  88. input: input,
  89. };
  90. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  91. // must contains {{input}}
  92. const inputVar = "{{input}}";
  93. if (!output.includes(inputVar)) {
  94. output += "\n" + inputVar;
  95. }
  96. Object.entries(vars).forEach(([name, value]) => {
  97. output = output.replaceAll(`{{${name}}}`, value);
  98. });
  99. return output;
  100. }
  101. const DEFAULT_CHAT_STATE = {
  102. sessions: [createEmptySession()],
  103. currentSessionIndex: 0,
  104. };
  105. export const useChatStore = createPersistStore(
  106. DEFAULT_CHAT_STATE,
  107. (set, _get) => {
  108. function get() {
  109. return {
  110. ..._get(),
  111. ...methods,
  112. };
  113. }
  114. const methods = {
  115. clearSessions() {
  116. set(() => ({
  117. sessions: [createEmptySession()],
  118. currentSessionIndex: 0,
  119. }));
  120. },
  121. selectSession(index: number) {
  122. set({
  123. currentSessionIndex: index,
  124. });
  125. },
  126. moveSession(from: number, to: number) {
  127. set((state) => {
  128. const { sessions, currentSessionIndex: oldIndex } = state;
  129. // move the session
  130. const newSessions = [...sessions];
  131. const session = newSessions[from];
  132. newSessions.splice(from, 1);
  133. newSessions.splice(to, 0, session);
  134. // modify current session id
  135. let newIndex = oldIndex === from ? to : oldIndex;
  136. if (oldIndex > from && oldIndex <= to) {
  137. newIndex -= 1;
  138. } else if (oldIndex < from && oldIndex >= to) {
  139. newIndex += 1;
  140. }
  141. return {
  142. currentSessionIndex: newIndex,
  143. sessions: newSessions,
  144. };
  145. });
  146. },
  147. newSession(mask?: Mask) {
  148. const session = createEmptySession();
  149. if (mask) {
  150. const config = useAppConfig.getState();
  151. const globalModelConfig = config.modelConfig;
  152. session.mask = {
  153. ...mask,
  154. modelConfig: {
  155. ...globalModelConfig,
  156. ...mask.modelConfig,
  157. },
  158. };
  159. session.topic = mask.name;
  160. }
  161. set((state) => ({
  162. currentSessionIndex: 0,
  163. sessions: [session].concat(state.sessions),
  164. }));
  165. },
  166. nextSession(delta: number) {
  167. const n = get().sessions.length;
  168. const limit = (x: number) => (x + n) % n;
  169. const i = get().currentSessionIndex;
  170. get().selectSession(limit(i + delta));
  171. },
  172. deleteSession(index: number) {
  173. const deletingLastSession = get().sessions.length === 1;
  174. const deletedSession = get().sessions.at(index);
  175. if (!deletedSession) return;
  176. const sessions = get().sessions.slice();
  177. sessions.splice(index, 1);
  178. const currentIndex = get().currentSessionIndex;
  179. let nextIndex = Math.min(
  180. currentIndex - Number(index < currentIndex),
  181. sessions.length - 1,
  182. );
  183. if (deletingLastSession) {
  184. nextIndex = 0;
  185. sessions.push(createEmptySession());
  186. }
  187. // for undo delete action
  188. const restoreState = {
  189. currentSessionIndex: get().currentSessionIndex,
  190. sessions: get().sessions.slice(),
  191. };
  192. set(() => ({
  193. currentSessionIndex: nextIndex,
  194. sessions,
  195. }));
  196. showToast(
  197. Locale.Home.DeleteToast,
  198. {
  199. text: Locale.Home.Revert,
  200. onClick() {
  201. set(() => restoreState);
  202. },
  203. },
  204. 5000,
  205. );
  206. },
  207. currentSession() {
  208. let index = get().currentSessionIndex;
  209. const sessions = get().sessions;
  210. if (index < 0 || index >= sessions.length) {
  211. index = Math.min(sessions.length - 1, Math.max(0, index));
  212. set(() => ({ currentSessionIndex: index }));
  213. }
  214. const session = sessions[index];
  215. return session;
  216. },
  217. onNewMessage(message: ChatMessage) {
  218. get().updateCurrentSession((session) => {
  219. session.messages = session.messages.concat();
  220. session.lastUpdate = Date.now();
  221. });
  222. get().updateStat(message);
  223. get().summarizeSession();
  224. },
  225. async onUserInput(content: string) {
  226. const session = get().currentSession();
  227. const modelConfig = session.mask.modelConfig;
  228. const userContent = fillTemplateWith(content, modelConfig);
  229. console.log("[User Input] after template: ", userContent);
  230. const userMessage: ChatMessage = createMessage({
  231. role: "user",
  232. content: userContent,
  233. });
  234. const botMessage: ChatMessage = createMessage({
  235. role: "assistant",
  236. streaming: true,
  237. model: modelConfig.model,
  238. });
  239. // get recent messages
  240. const recentMessages = get().getMessagesWithMemory();
  241. const sendMessages = recentMessages.concat(userMessage);
  242. const messageIndex = get().currentSession().messages.length + 1;
  243. // save user's and bot's message
  244. get().updateCurrentSession((session) => {
  245. const savedUserMessage = {
  246. ...userMessage,
  247. content,
  248. };
  249. session.messages = session.messages.concat([
  250. savedUserMessage,
  251. botMessage,
  252. ]);
  253. });
  254. var api: ClientApi;
  255. if (modelConfig.model === "gemini") {
  256. api = new ClientApi(ModelProvider.Gemini);
  257. } else {
  258. api = new ClientApi(ModelProvider.GPT);
  259. }
  260. // make request
  261. api.llm.chat({
  262. messages: sendMessages,
  263. config: { ...modelConfig, stream: true },
  264. onUpdate(message) {
  265. botMessage.streaming = true;
  266. if (message) {
  267. botMessage.content = message;
  268. }
  269. get().updateCurrentSession((session) => {
  270. session.messages = session.messages.concat();
  271. });
  272. },
  273. onFinish(message) {
  274. botMessage.streaming = false;
  275. if (message) {
  276. botMessage.content = message;
  277. get().onNewMessage(botMessage);
  278. }
  279. ChatControllerPool.remove(session.id, botMessage.id);
  280. },
  281. onError(error) {
  282. const isAborted = error.message.includes("aborted");
  283. botMessage.content +=
  284. "\n\n" +
  285. prettyObject({
  286. error: true,
  287. message: error.message,
  288. });
  289. botMessage.streaming = false;
  290. userMessage.isError = !isAborted;
  291. botMessage.isError = !isAborted;
  292. get().updateCurrentSession((session) => {
  293. session.messages = session.messages.concat();
  294. });
  295. ChatControllerPool.remove(
  296. session.id,
  297. botMessage.id ?? messageIndex,
  298. );
  299. console.error("[Chat] failed ", error);
  300. },
  301. onController(controller) {
  302. // collect controller for stop/retry
  303. ChatControllerPool.addController(
  304. session.id,
  305. botMessage.id ?? messageIndex,
  306. controller,
  307. );
  308. },
  309. });
  310. },
  311. getMemoryPrompt() {
  312. const session = get().currentSession();
  313. return {
  314. role: "system",
  315. content:
  316. session.memoryPrompt.length > 0
  317. ? Locale.Store.Prompt.History(session.memoryPrompt)
  318. : "",
  319. date: "",
  320. } as ChatMessage;
  321. },
  322. getMessagesWithMemory() {
  323. const session = get().currentSession();
  324. const modelConfig = session.mask.modelConfig;
  325. const clearContextIndex = session.clearContextIndex ?? 0;
  326. const messages = session.messages.slice();
  327. const totalMessageCount = session.messages.length;
  328. // in-context prompts
  329. const contextPrompts = session.mask.context.slice();
  330. // system prompts, to get close to OpenAI Web ChatGPT
  331. const shouldInjectSystemPrompts = modelConfig.enableInjectSystemPrompts;
  332. var systemPrompts: ChatMessage[] = [];
  333. if (modelConfig.model !== "gemini") {
  334. systemPrompts = shouldInjectSystemPrompts
  335. ? [
  336. createMessage({
  337. role: "system",
  338. content: fillTemplateWith("", {
  339. ...modelConfig,
  340. template: DEFAULT_SYSTEM_TEMPLATE,
  341. }),
  342. }),
  343. ]
  344. : [];
  345. if (shouldInjectSystemPrompts) {
  346. console.log(
  347. "[Global System Prompt] ",
  348. systemPrompts.at(0)?.content ?? "empty",
  349. );
  350. }
  351. }
  352. // long term memory
  353. const shouldSendLongTermMemory =
  354. modelConfig.sendMemory &&
  355. session.memoryPrompt &&
  356. session.memoryPrompt.length > 0 &&
  357. session.lastSummarizeIndex > clearContextIndex;
  358. const longTermMemoryPrompts = shouldSendLongTermMemory
  359. ? [get().getMemoryPrompt()]
  360. : [];
  361. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  362. // short term memory
  363. const shortTermMemoryStartIndex = Math.max(
  364. 0,
  365. totalMessageCount - modelConfig.historyMessageCount,
  366. );
  367. // lets concat send messages, including 4 parts:
  368. // 0. system prompt: to get close to OpenAI Web ChatGPT
  369. // 1. long term memory: summarized memory messages
  370. // 2. pre-defined in-context prompts
  371. // 3. short term memory: latest n messages
  372. // 4. newest input message
  373. const memoryStartIndex = shouldSendLongTermMemory
  374. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  375. : shortTermMemoryStartIndex;
  376. // and if user has cleared history messages, we should exclude the memory too.
  377. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  378. const maxTokenThreshold = modelConfig.max_tokens;
  379. // get recent messages as much as possible
  380. const reversedRecentMessages = [];
  381. for (
  382. let i = totalMessageCount - 1, tokenCount = 0;
  383. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  384. i -= 1
  385. ) {
  386. const msg = messages[i];
  387. if (!msg || msg.isError) continue;
  388. tokenCount += estimateTokenLength(msg.content);
  389. reversedRecentMessages.push(msg);
  390. }
  391. // concat all messages
  392. const recentMessages = [
  393. ...systemPrompts,
  394. ...longTermMemoryPrompts,
  395. ...contextPrompts,
  396. ...reversedRecentMessages.reverse(),
  397. ];
  398. return recentMessages;
  399. },
  400. updateMessage(
  401. sessionIndex: number,
  402. messageIndex: number,
  403. updater: (message?: ChatMessage) => void,
  404. ) {
  405. const sessions = get().sessions;
  406. const session = sessions.at(sessionIndex);
  407. const messages = session?.messages;
  408. updater(messages?.at(messageIndex));
  409. set(() => ({ sessions }));
  410. },
  411. resetSession() {
  412. get().updateCurrentSession((session) => {
  413. session.messages = [];
  414. session.memoryPrompt = "";
  415. });
  416. },
  417. summarizeSession() {
  418. const config = useAppConfig.getState();
  419. const session = get().currentSession();
  420. const modelConfig = session.mask.modelConfig;
  421. var api: ClientApi;
  422. if (modelConfig.model === "gemini") {
  423. api = new ClientApi(ModelProvider.Gemini);
  424. } else {
  425. api = new ClientApi(ModelProvider.GPT);
  426. }
  427. // remove error messages if any
  428. const messages = session.messages;
  429. // should summarize topic after chating more than 50 words
  430. const SUMMARIZE_MIN_LEN = 50;
  431. if (
  432. config.enableAutoGenerateTitle &&
  433. session.topic === DEFAULT_TOPIC &&
  434. countMessages(messages) >= SUMMARIZE_MIN_LEN
  435. ) {
  436. const topicMessages = messages.concat(
  437. createMessage({
  438. role: "user",
  439. content: Locale.Store.Prompt.Topic,
  440. }),
  441. );
  442. api.llm.chat({
  443. messages: topicMessages,
  444. config: {
  445. model: getSummarizeModel(session.mask.modelConfig.model),
  446. },
  447. onFinish(message) {
  448. get().updateCurrentSession(
  449. (session) =>
  450. (session.topic =
  451. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  452. );
  453. },
  454. });
  455. }
  456. const summarizeIndex = Math.max(
  457. session.lastSummarizeIndex,
  458. session.clearContextIndex ?? 0,
  459. );
  460. let toBeSummarizedMsgs = messages
  461. .filter((msg) => !msg.isError)
  462. .slice(summarizeIndex);
  463. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  464. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  465. const n = toBeSummarizedMsgs.length;
  466. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  467. Math.max(0, n - modelConfig.historyMessageCount),
  468. );
  469. }
  470. // add memory prompt
  471. toBeSummarizedMsgs.unshift(get().getMemoryPrompt());
  472. const lastSummarizeIndex = session.messages.length;
  473. console.log(
  474. "[Chat History] ",
  475. toBeSummarizedMsgs,
  476. historyMsgLength,
  477. modelConfig.compressMessageLengthThreshold,
  478. );
  479. if (
  480. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  481. modelConfig.sendMemory
  482. ) {
  483. api.llm.chat({
  484. messages: toBeSummarizedMsgs.concat(
  485. createMessage({
  486. role: "system",
  487. content: Locale.Store.Prompt.Summarize,
  488. date: "",
  489. }),
  490. ),
  491. config: {
  492. ...modelConfig,
  493. stream: true,
  494. model: getSummarizeModel(session.mask.modelConfig.model),
  495. },
  496. onUpdate(message) {
  497. session.memoryPrompt = message;
  498. },
  499. onFinish(message) {
  500. console.log("[Memory] ", message);
  501. get().updateCurrentSession((session) => {
  502. session.lastSummarizeIndex = lastSummarizeIndex;
  503. session.memoryPrompt = message; // Update the memory prompt for stored it in local storage
  504. });
  505. },
  506. onError(err) {
  507. console.error("[Summarize] ", err);
  508. },
  509. });
  510. }
  511. },
  512. updateStat(message: ChatMessage) {
  513. get().updateCurrentSession((session) => {
  514. session.stat.charCount += message.content.length;
  515. // TODO: should update chat count and word count
  516. });
  517. },
  518. updateCurrentSession(updater: (session: ChatSession) => void) {
  519. const sessions = get().sessions;
  520. const index = get().currentSessionIndex;
  521. updater(sessions[index]);
  522. set(() => ({ sessions }));
  523. },
  524. clearAllData() {
  525. localStorage.clear();
  526. location.reload();
  527. },
  528. };
  529. return methods;
  530. },
  531. {
  532. name: StoreKey.Chat,
  533. version: 3.1,
  534. migrate(persistedState, version) {
  535. const state = persistedState as any;
  536. const newState = JSON.parse(
  537. JSON.stringify(state),
  538. ) as typeof DEFAULT_CHAT_STATE;
  539. if (version < 2) {
  540. newState.sessions = [];
  541. const oldSessions = state.sessions;
  542. for (const oldSession of oldSessions) {
  543. const newSession = createEmptySession();
  544. newSession.topic = oldSession.topic;
  545. newSession.messages = [...oldSession.messages];
  546. newSession.mask.modelConfig.sendMemory = true;
  547. newSession.mask.modelConfig.historyMessageCount = 4;
  548. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  549. newState.sessions.push(newSession);
  550. }
  551. }
  552. if (version < 3) {
  553. // migrate id to nanoid
  554. newState.sessions.forEach((s) => {
  555. s.id = nanoid();
  556. s.messages.forEach((m) => (m.id = nanoid()));
  557. });
  558. }
  559. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  560. // Resolve issue of old sessions not automatically enabling.
  561. if (version < 3.1) {
  562. newState.sessions.forEach((s) => {
  563. if (
  564. // Exclude those already set by user
  565. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  566. ) {
  567. // Because users may have changed this configuration,
  568. // the user's current configuration is used instead of the default
  569. const config = useAppConfig.getState();
  570. s.mask.modelConfig.enableInjectSystemPrompts =
  571. config.modelConfig.enableInjectSystemPrompts;
  572. }
  573. });
  574. }
  575. return newState as any;
  576. },
  577. },
  578. );