chat.ts 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677
  1. import { trimTopic } from "../utils";
  2. import Locale, { getLang } from "../locales";
  3. import { showToast } from "../components/ui-lib";
  4. import { ModelConfig, ModelType, useAppConfig } from "./config";
  5. import { createEmptyMask, Mask } from "./mask";
  6. import {
  7. DEFAULT_INPUT_TEMPLATE,
  8. DEFAULT_MODELS,
  9. DEFAULT_SYSTEM_TEMPLATE,
  10. KnowledgeCutOffDate,
  11. ModelProvider,
  12. StoreKey,
  13. SUMMARIZE_MODEL,
  14. } from "../constant";
  15. import { ClientApi, RequestMessage } from "../client/api";
  16. import { ChatControllerPool } from "../client/controller";
  17. import { prettyObject } from "../utils/format";
  18. import { estimateTokenLength } from "../utils/token";
  19. import { nanoid } from "nanoid";
  20. import { createPersistStore } from "../utils/store";
  21. export type ChatMessage = RequestMessage & {
  22. date: string;
  23. streaming?: boolean;
  24. isError?: boolean;
  25. id: string;
  26. model?: ModelType;
  27. };
  28. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  29. return {
  30. id: nanoid(),
  31. date: new Date().toLocaleString(),
  32. role: "user",
  33. content: "",
  34. ...override,
  35. };
  36. }
  37. export interface ChatStat {
  38. tokenCount: number;
  39. wordCount: number;
  40. charCount: number;
  41. }
  42. export interface ChatSession {
  43. id: string;
  44. topic: string;
  45. memoryPrompt: string;
  46. messages: ChatMessage[];
  47. stat: ChatStat;
  48. lastUpdate: number;
  49. lastSummarizeIndex: number;
  50. clearContextIndex?: number;
  51. mask: Mask;
  52. }
  53. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  54. export const BOT_HELLO: ChatMessage = createMessage({
  55. role: "assistant",
  56. content: Locale.Store.BotHello,
  57. });
  58. function createEmptySession(): ChatSession {
  59. return {
  60. id: nanoid(),
  61. topic: DEFAULT_TOPIC,
  62. memoryPrompt: "",
  63. messages: [],
  64. stat: {
  65. tokenCount: 0,
  66. wordCount: 0,
  67. charCount: 0,
  68. },
  69. lastUpdate: Date.now(),
  70. lastSummarizeIndex: 0,
  71. mask: createEmptyMask(),
  72. };
  73. }
  74. function getSummarizeModel(currentModel: string) {
  75. // if it is using gpt-* models, force to use 3.5 to summarize
  76. return currentModel.startsWith("gpt") ? SUMMARIZE_MODEL : currentModel;
  77. }
  78. function countMessages(msgs: ChatMessage[]) {
  79. return msgs.reduce((pre, cur) => pre + estimateTokenLength(cur.content), 0);
  80. }
  81. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  82. const cutoff =
  83. KnowledgeCutOffDate[modelConfig.model] ?? KnowledgeCutOffDate.default;
  84. // Find the model in the DEFAULT_MODELS array that matches the modelConfig.model
  85. const modelInfo = DEFAULT_MODELS.find((m) => m.name === modelConfig.model);
  86. var serviceProvider = "OpenAI";
  87. if (modelInfo) {
  88. // TODO: auto detect the providerName from the modelConfig.model
  89. // Directly use the providerName from the modelInfo
  90. serviceProvider = modelInfo.provider.providerName;
  91. }
  92. const vars = {
  93. ServiceProvider: serviceProvider,
  94. cutoff,
  95. model: modelConfig.model,
  96. time: new Date().toLocaleString(),
  97. lang: getLang(),
  98. input: input,
  99. };
  100. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  101. // must contains {{input}}
  102. const inputVar = "{{input}}";
  103. if (!output.includes(inputVar)) {
  104. output += "\n" + inputVar;
  105. }
  106. Object.entries(vars).forEach(([name, value]) => {
  107. const regex = new RegExp(`{{${name}}}`, "g");
  108. output = output.replace(regex, value.toString()); // Ensure value is a string
  109. });
  110. return output;
  111. }
  112. const DEFAULT_CHAT_STATE = {
  113. sessions: [createEmptySession()],
  114. currentSessionIndex: 0,
  115. };
  116. export const useChatStore = createPersistStore(
  117. DEFAULT_CHAT_STATE,
  118. (set, _get) => {
  119. function get() {
  120. return {
  121. ..._get(),
  122. ...methods,
  123. };
  124. }
  125. const methods = {
  126. clearSessions() {
  127. set(() => ({
  128. sessions: [createEmptySession()],
  129. currentSessionIndex: 0,
  130. }));
  131. },
  132. selectSession(index: number) {
  133. set({
  134. currentSessionIndex: index,
  135. });
  136. },
  137. moveSession(from: number, to: number) {
  138. set((state) => {
  139. const { sessions, currentSessionIndex: oldIndex } = state;
  140. // move the session
  141. const newSessions = [...sessions];
  142. const session = newSessions[from];
  143. newSessions.splice(from, 1);
  144. newSessions.splice(to, 0, session);
  145. // modify current session id
  146. let newIndex = oldIndex === from ? to : oldIndex;
  147. if (oldIndex > from && oldIndex <= to) {
  148. newIndex -= 1;
  149. } else if (oldIndex < from && oldIndex >= to) {
  150. newIndex += 1;
  151. }
  152. return {
  153. currentSessionIndex: newIndex,
  154. sessions: newSessions,
  155. };
  156. });
  157. },
  158. newSession(mask?: Mask) {
  159. const session = createEmptySession();
  160. if (mask) {
  161. const config = useAppConfig.getState();
  162. const globalModelConfig = config.modelConfig;
  163. session.mask = {
  164. ...mask,
  165. modelConfig: {
  166. ...globalModelConfig,
  167. ...mask.modelConfig,
  168. },
  169. };
  170. session.topic = mask.name;
  171. }
  172. set((state) => ({
  173. currentSessionIndex: 0,
  174. sessions: [session].concat(state.sessions),
  175. }));
  176. },
  177. nextSession(delta: number) {
  178. const n = get().sessions.length;
  179. const limit = (x: number) => (x + n) % n;
  180. const i = get().currentSessionIndex;
  181. get().selectSession(limit(i + delta));
  182. },
  183. deleteSession(index: number) {
  184. const deletingLastSession = get().sessions.length === 1;
  185. const deletedSession = get().sessions.at(index);
  186. if (!deletedSession) return;
  187. const sessions = get().sessions.slice();
  188. sessions.splice(index, 1);
  189. const currentIndex = get().currentSessionIndex;
  190. let nextIndex = Math.min(
  191. currentIndex - Number(index < currentIndex),
  192. sessions.length - 1,
  193. );
  194. if (deletingLastSession) {
  195. nextIndex = 0;
  196. sessions.push(createEmptySession());
  197. }
  198. // for undo delete action
  199. const restoreState = {
  200. currentSessionIndex: get().currentSessionIndex,
  201. sessions: get().sessions.slice(),
  202. };
  203. set(() => ({
  204. currentSessionIndex: nextIndex,
  205. sessions,
  206. }));
  207. showToast(
  208. Locale.Home.DeleteToast,
  209. {
  210. text: Locale.Home.Revert,
  211. onClick() {
  212. set(() => restoreState);
  213. },
  214. },
  215. 5000,
  216. );
  217. },
  218. currentSession() {
  219. let index = get().currentSessionIndex;
  220. const sessions = get().sessions;
  221. if (index < 0 || index >= sessions.length) {
  222. index = Math.min(sessions.length - 1, Math.max(0, index));
  223. set(() => ({ currentSessionIndex: index }));
  224. }
  225. const session = sessions[index];
  226. return session;
  227. },
  228. onNewMessage(message: ChatMessage) {
  229. get().updateCurrentSession((session) => {
  230. session.messages = session.messages.concat();
  231. session.lastUpdate = Date.now();
  232. });
  233. get().updateStat(message);
  234. get().summarizeSession();
  235. },
  236. async onUserInput(content: string) {
  237. const session = get().currentSession();
  238. const modelConfig = session.mask.modelConfig;
  239. const userContent = fillTemplateWith(content, modelConfig);
  240. console.log("[User Input] after template: ", userContent);
  241. const userMessage: ChatMessage = createMessage({
  242. role: "user",
  243. content: userContent,
  244. });
  245. const botMessage: ChatMessage = createMessage({
  246. role: "assistant",
  247. streaming: true,
  248. model: modelConfig.model,
  249. });
  250. // get recent messages
  251. const recentMessages = get().getMessagesWithMemory();
  252. const sendMessages = recentMessages.concat(userMessage);
  253. const messageIndex = get().currentSession().messages.length + 1;
  254. // save user's and bot's message
  255. get().updateCurrentSession((session) => {
  256. const savedUserMessage = {
  257. ...userMessage,
  258. content,
  259. };
  260. session.messages = session.messages.concat([
  261. savedUserMessage,
  262. botMessage,
  263. ]);
  264. });
  265. var api: ClientApi;
  266. if (modelConfig.model.startsWith("gemini")) {
  267. api = new ClientApi(ModelProvider.GeminiPro);
  268. } else {
  269. api = new ClientApi(ModelProvider.GPT);
  270. }
  271. // make request
  272. api.llm.chat({
  273. messages: sendMessages,
  274. config: { ...modelConfig, stream: true },
  275. onUpdate(message) {
  276. botMessage.streaming = true;
  277. if (message) {
  278. botMessage.content = message;
  279. }
  280. get().updateCurrentSession((session) => {
  281. session.messages = session.messages.concat();
  282. });
  283. },
  284. onFinish(message) {
  285. botMessage.streaming = false;
  286. if (message) {
  287. botMessage.content = message;
  288. get().onNewMessage(botMessage);
  289. }
  290. ChatControllerPool.remove(session.id, botMessage.id);
  291. },
  292. onError(error) {
  293. const isAborted = error.message.includes("aborted");
  294. botMessage.content +=
  295. "\n\n" +
  296. prettyObject({
  297. error: true,
  298. message: error.message,
  299. });
  300. botMessage.streaming = false;
  301. userMessage.isError = !isAborted;
  302. botMessage.isError = !isAborted;
  303. get().updateCurrentSession((session) => {
  304. session.messages = session.messages.concat();
  305. });
  306. ChatControllerPool.remove(
  307. session.id,
  308. botMessage.id ?? messageIndex,
  309. );
  310. console.error("[Chat] failed ", error);
  311. },
  312. onController(controller) {
  313. // collect controller for stop/retry
  314. ChatControllerPool.addController(
  315. session.id,
  316. botMessage.id ?? messageIndex,
  317. controller,
  318. );
  319. },
  320. });
  321. },
  322. getMemoryPrompt() {
  323. const session = get().currentSession();
  324. return {
  325. role: "system",
  326. content:
  327. session.memoryPrompt.length > 0
  328. ? Locale.Store.Prompt.History(session.memoryPrompt)
  329. : "",
  330. date: "",
  331. } as ChatMessage;
  332. },
  333. getMessagesWithMemory() {
  334. const session = get().currentSession();
  335. const modelConfig = session.mask.modelConfig;
  336. const clearContextIndex = session.clearContextIndex ?? 0;
  337. const messages = session.messages.slice();
  338. const totalMessageCount = session.messages.length;
  339. // in-context prompts
  340. const contextPrompts = session.mask.context.slice();
  341. // system prompts, to get close to OpenAI Web ChatGPT
  342. const shouldInjectSystemPrompts =
  343. modelConfig.enableInjectSystemPrompts &&
  344. session.mask.modelConfig.model.startsWith("gpt-");
  345. var systemPrompts: ChatMessage[] = [];
  346. systemPrompts = shouldInjectSystemPrompts
  347. ? [
  348. createMessage({
  349. role: "system",
  350. content: fillTemplateWith("", {
  351. ...modelConfig,
  352. template: DEFAULT_SYSTEM_TEMPLATE,
  353. }),
  354. }),
  355. ]
  356. : [];
  357. if (shouldInjectSystemPrompts) {
  358. console.log(
  359. "[Global System Prompt] ",
  360. systemPrompts.at(0)?.content ?? "empty",
  361. );
  362. }
  363. // long term memory
  364. const shouldSendLongTermMemory =
  365. modelConfig.sendMemory &&
  366. session.memoryPrompt &&
  367. session.memoryPrompt.length > 0 &&
  368. session.lastSummarizeIndex > clearContextIndex;
  369. const longTermMemoryPrompts = shouldSendLongTermMemory
  370. ? [get().getMemoryPrompt()]
  371. : [];
  372. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  373. // short term memory
  374. const shortTermMemoryStartIndex = Math.max(
  375. 0,
  376. totalMessageCount - modelConfig.historyMessageCount,
  377. );
  378. // lets concat send messages, including 4 parts:
  379. // 0. system prompt: to get close to OpenAI Web ChatGPT
  380. // 1. long term memory: summarized memory messages
  381. // 2. pre-defined in-context prompts
  382. // 3. short term memory: latest n messages
  383. // 4. newest input message
  384. const memoryStartIndex = shouldSendLongTermMemory
  385. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  386. : shortTermMemoryStartIndex;
  387. // and if user has cleared history messages, we should exclude the memory too.
  388. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  389. const maxTokenThreshold = modelConfig.max_tokens;
  390. // get recent messages as much as possible
  391. const reversedRecentMessages = [];
  392. for (
  393. let i = totalMessageCount - 1, tokenCount = 0;
  394. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  395. i -= 1
  396. ) {
  397. const msg = messages[i];
  398. if (!msg || msg.isError) continue;
  399. tokenCount += estimateTokenLength(msg.content);
  400. reversedRecentMessages.push(msg);
  401. }
  402. // concat all messages
  403. const recentMessages = [
  404. ...systemPrompts,
  405. ...longTermMemoryPrompts,
  406. ...contextPrompts,
  407. ...reversedRecentMessages.reverse(),
  408. ];
  409. return recentMessages;
  410. },
  411. updateMessage(
  412. sessionIndex: number,
  413. messageIndex: number,
  414. updater: (message?: ChatMessage) => void,
  415. ) {
  416. const sessions = get().sessions;
  417. const session = sessions.at(sessionIndex);
  418. const messages = session?.messages;
  419. updater(messages?.at(messageIndex));
  420. set(() => ({ sessions }));
  421. },
  422. resetSession() {
  423. get().updateCurrentSession((session) => {
  424. session.messages = [];
  425. session.memoryPrompt = "";
  426. });
  427. },
  428. summarizeSession() {
  429. const config = useAppConfig.getState();
  430. const session = get().currentSession();
  431. const modelConfig = session.mask.modelConfig;
  432. var api: ClientApi;
  433. if (modelConfig.model.startsWith("gemini")) {
  434. api = new ClientApi(ModelProvider.GeminiPro);
  435. } else {
  436. api = new ClientApi(ModelProvider.GPT);
  437. }
  438. // remove error messages if any
  439. const messages = session.messages;
  440. // should summarize topic after chating more than 50 words
  441. const SUMMARIZE_MIN_LEN = 50;
  442. if (
  443. config.enableAutoGenerateTitle &&
  444. session.topic === DEFAULT_TOPIC &&
  445. countMessages(messages) >= SUMMARIZE_MIN_LEN
  446. ) {
  447. const topicMessages = messages.concat(
  448. createMessage({
  449. role: "user",
  450. content: Locale.Store.Prompt.Topic,
  451. }),
  452. );
  453. api.llm.chat({
  454. messages: topicMessages,
  455. config: {
  456. model: getSummarizeModel(session.mask.modelConfig.model),
  457. },
  458. onFinish(message) {
  459. get().updateCurrentSession(
  460. (session) =>
  461. (session.topic =
  462. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  463. );
  464. },
  465. });
  466. }
  467. const summarizeIndex = Math.max(
  468. session.lastSummarizeIndex,
  469. session.clearContextIndex ?? 0,
  470. );
  471. let toBeSummarizedMsgs = messages
  472. .filter((msg) => !msg.isError)
  473. .slice(summarizeIndex);
  474. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  475. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  476. const n = toBeSummarizedMsgs.length;
  477. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  478. Math.max(0, n - modelConfig.historyMessageCount),
  479. );
  480. }
  481. // add memory prompt
  482. toBeSummarizedMsgs.unshift(get().getMemoryPrompt());
  483. const lastSummarizeIndex = session.messages.length;
  484. console.log(
  485. "[Chat History] ",
  486. toBeSummarizedMsgs,
  487. historyMsgLength,
  488. modelConfig.compressMessageLengthThreshold,
  489. );
  490. if (
  491. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  492. modelConfig.sendMemory
  493. ) {
  494. api.llm.chat({
  495. messages: toBeSummarizedMsgs.concat(
  496. createMessage({
  497. role: "system",
  498. content: Locale.Store.Prompt.Summarize,
  499. date: "",
  500. }),
  501. ),
  502. config: {
  503. ...modelConfig,
  504. stream: true,
  505. model: getSummarizeModel(session.mask.modelConfig.model),
  506. },
  507. onUpdate(message) {
  508. session.memoryPrompt = message;
  509. },
  510. onFinish(message) {
  511. console.log("[Memory] ", message);
  512. get().updateCurrentSession((session) => {
  513. session.lastSummarizeIndex = lastSummarizeIndex;
  514. session.memoryPrompt = message; // Update the memory prompt for stored it in local storage
  515. });
  516. },
  517. onError(err) {
  518. console.error("[Summarize] ", err);
  519. },
  520. });
  521. }
  522. },
  523. updateStat(message: ChatMessage) {
  524. get().updateCurrentSession((session) => {
  525. session.stat.charCount += message.content.length;
  526. // TODO: should update chat count and word count
  527. });
  528. },
  529. updateCurrentSession(updater: (session: ChatSession) => void) {
  530. const sessions = get().sessions;
  531. const index = get().currentSessionIndex;
  532. updater(sessions[index]);
  533. set(() => ({ sessions }));
  534. },
  535. clearAllData() {
  536. localStorage.clear();
  537. location.reload();
  538. },
  539. };
  540. return methods;
  541. },
  542. {
  543. name: StoreKey.Chat,
  544. version: 3.1,
  545. migrate(persistedState, version) {
  546. const state = persistedState as any;
  547. const newState = JSON.parse(
  548. JSON.stringify(state),
  549. ) as typeof DEFAULT_CHAT_STATE;
  550. if (version < 2) {
  551. newState.sessions = [];
  552. const oldSessions = state.sessions;
  553. for (const oldSession of oldSessions) {
  554. const newSession = createEmptySession();
  555. newSession.topic = oldSession.topic;
  556. newSession.messages = [...oldSession.messages];
  557. newSession.mask.modelConfig.sendMemory = true;
  558. newSession.mask.modelConfig.historyMessageCount = 4;
  559. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  560. newState.sessions.push(newSession);
  561. }
  562. }
  563. if (version < 3) {
  564. // migrate id to nanoid
  565. newState.sessions.forEach((s) => {
  566. s.id = nanoid();
  567. s.messages.forEach((m) => (m.id = nanoid()));
  568. });
  569. }
  570. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  571. // Resolve issue of old sessions not automatically enabling.
  572. if (version < 3.1) {
  573. newState.sessions.forEach((s) => {
  574. if (
  575. // Exclude those already set by user
  576. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  577. ) {
  578. // Because users may have changed this configuration,
  579. // the user's current configuration is used instead of the default
  580. const config = useAppConfig.getState();
  581. s.mask.modelConfig.enableInjectSystemPrompts =
  582. config.modelConfig.enableInjectSystemPrompts;
  583. }
  584. });
  585. }
  586. return newState as any;
  587. },
  588. },
  589. );