chat.ts 21 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720
  1. import { trimTopic, getMessageTextContent } from "../utils";
  2. import Locale, { getLang } from "../locales";
  3. import { showToast } from "../components/ui-lib";
  4. import { ModelConfig, ModelType, useAppConfig } from "./config";
  5. import { createEmptyMask, Mask } from "./mask";
  6. import {
  7. DEFAULT_INPUT_TEMPLATE,
  8. DEFAULT_MODELS,
  9. DEFAULT_SYSTEM_TEMPLATE,
  10. KnowledgeCutOffDate,
  11. ModelProvider,
  12. StoreKey,
  13. SUMMARIZE_MODEL,
  14. GEMINI_SUMMARIZE_MODEL,
  15. } from "../constant";
  16. import { ClientApi, RequestMessage, MultimodalContent } from "../client/api";
  17. import { ChatControllerPool } from "../client/controller";
  18. import { prettyObject } from "../utils/format";
  19. import { estimateTokenLength } from "../utils/token";
  20. import { nanoid } from "nanoid";
  21. import { createPersistStore } from "../utils/store";
  22. export type ChatMessage = RequestMessage & {
  23. date: string;
  24. streaming?: boolean;
  25. isError?: boolean;
  26. id: string;
  27. model?: ModelType;
  28. };
  29. export function createMessage(override: Partial<ChatMessage>): ChatMessage {
  30. return {
  31. id: nanoid(),
  32. date: new Date().toLocaleString(),
  33. role: "user",
  34. content: "",
  35. ...override,
  36. };
  37. }
  38. export interface ChatStat {
  39. tokenCount: number;
  40. wordCount: number;
  41. charCount: number;
  42. }
  43. export interface ChatSession {
  44. id: string;
  45. topic: string;
  46. memoryPrompt: string;
  47. messages: ChatMessage[];
  48. stat: ChatStat;
  49. lastUpdate: number;
  50. lastSummarizeIndex: number;
  51. clearContextIndex?: number;
  52. mask: Mask;
  53. }
  54. export const DEFAULT_TOPIC = Locale.Store.DefaultTopic;
  55. export const BOT_HELLO: ChatMessage = createMessage({
  56. role: "assistant",
  57. content: Locale.Store.BotHello,
  58. });
  59. function createEmptySession(): ChatSession {
  60. return {
  61. id: nanoid(),
  62. topic: DEFAULT_TOPIC,
  63. memoryPrompt: "",
  64. messages: [],
  65. stat: {
  66. tokenCount: 0,
  67. wordCount: 0,
  68. charCount: 0,
  69. },
  70. lastUpdate: Date.now(),
  71. lastSummarizeIndex: 0,
  72. mask: createEmptyMask(),
  73. };
  74. }
  75. function getSummarizeModel(currentModel: string) {
  76. // if it is using gpt-* models, force to use 3.5 to summarize
  77. if (currentModel.startsWith("gpt")) {
  78. return SUMMARIZE_MODEL;
  79. }
  80. if (currentModel.startsWith("gemini-pro")) {
  81. return GEMINI_SUMMARIZE_MODEL;
  82. }
  83. return currentModel;
  84. }
  85. function countMessages(msgs: ChatMessage[]) {
  86. return msgs.reduce(
  87. (pre, cur) => pre + estimateTokenLength(getMessageTextContent(cur)),
  88. 0,
  89. );
  90. }
  91. function fillTemplateWith(input: string, modelConfig: ModelConfig) {
  92. const cutoff =
  93. KnowledgeCutOffDate[modelConfig.model] ?? KnowledgeCutOffDate.default;
  94. // Find the model in the DEFAULT_MODELS array that matches the modelConfig.model
  95. const modelInfo = DEFAULT_MODELS.find((m) => m.name === modelConfig.model);
  96. var serviceProvider = "OpenAI";
  97. if (modelInfo) {
  98. // TODO: auto detect the providerName from the modelConfig.model
  99. // Directly use the providerName from the modelInfo
  100. serviceProvider = modelInfo.provider.providerName;
  101. }
  102. const vars = {
  103. ServiceProvider: serviceProvider,
  104. cutoff,
  105. model: modelConfig.model,
  106. time: new Date().toLocaleString(),
  107. lang: getLang(),
  108. input: input,
  109. };
  110. let output = modelConfig.template ?? DEFAULT_INPUT_TEMPLATE;
  111. // remove duplicate
  112. if (input.startsWith(output)) {
  113. output = "";
  114. }
  115. // must contains {{input}}
  116. const inputVar = "{{input}}";
  117. if (!output.includes(inputVar)) {
  118. output += "\n" + inputVar;
  119. }
  120. Object.entries(vars).forEach(([name, value]) => {
  121. const regex = new RegExp(`{{${name}}}`, "g");
  122. output = output.replace(regex, value.toString()); // Ensure value is a string
  123. });
  124. return output;
  125. }
  126. const DEFAULT_CHAT_STATE = {
  127. sessions: [createEmptySession()],
  128. currentSessionIndex: 0,
  129. };
  130. export const useChatStore = createPersistStore(
  131. DEFAULT_CHAT_STATE,
  132. (set, _get) => {
  133. function get() {
  134. return {
  135. ..._get(),
  136. ...methods,
  137. };
  138. }
  139. const methods = {
  140. clearSessions() {
  141. set(() => ({
  142. sessions: [createEmptySession()],
  143. currentSessionIndex: 0,
  144. }));
  145. },
  146. selectSession(index: number) {
  147. set({
  148. currentSessionIndex: index,
  149. });
  150. },
  151. moveSession(from: number, to: number) {
  152. set((state) => {
  153. const { sessions, currentSessionIndex: oldIndex } = state;
  154. // move the session
  155. const newSessions = [...sessions];
  156. const session = newSessions[from];
  157. newSessions.splice(from, 1);
  158. newSessions.splice(to, 0, session);
  159. // modify current session id
  160. let newIndex = oldIndex === from ? to : oldIndex;
  161. if (oldIndex > from && oldIndex <= to) {
  162. newIndex -= 1;
  163. } else if (oldIndex < from && oldIndex >= to) {
  164. newIndex += 1;
  165. }
  166. return {
  167. currentSessionIndex: newIndex,
  168. sessions: newSessions,
  169. };
  170. });
  171. },
  172. newSession(mask?: Mask) {
  173. const session = createEmptySession();
  174. if (mask) {
  175. const config = useAppConfig.getState();
  176. const globalModelConfig = config.modelConfig;
  177. session.mask = {
  178. ...mask,
  179. modelConfig: {
  180. ...globalModelConfig,
  181. ...mask.modelConfig,
  182. },
  183. };
  184. session.topic = mask.name;
  185. }
  186. set((state) => ({
  187. currentSessionIndex: 0,
  188. sessions: [session].concat(state.sessions),
  189. }));
  190. },
  191. nextSession(delta: number) {
  192. const n = get().sessions.length;
  193. const limit = (x: number) => (x + n) % n;
  194. const i = get().currentSessionIndex;
  195. get().selectSession(limit(i + delta));
  196. },
  197. deleteSession(index: number) {
  198. const deletingLastSession = get().sessions.length === 1;
  199. const deletedSession = get().sessions.at(index);
  200. if (!deletedSession) return;
  201. const sessions = get().sessions.slice();
  202. sessions.splice(index, 1);
  203. const currentIndex = get().currentSessionIndex;
  204. let nextIndex = Math.min(
  205. currentIndex - Number(index < currentIndex),
  206. sessions.length - 1,
  207. );
  208. if (deletingLastSession) {
  209. nextIndex = 0;
  210. sessions.push(createEmptySession());
  211. }
  212. // for undo delete action
  213. const restoreState = {
  214. currentSessionIndex: get().currentSessionIndex,
  215. sessions: get().sessions.slice(),
  216. };
  217. set(() => ({
  218. currentSessionIndex: nextIndex,
  219. sessions,
  220. }));
  221. showToast(
  222. Locale.Home.DeleteToast,
  223. {
  224. text: Locale.Home.Revert,
  225. onClick() {
  226. set(() => restoreState);
  227. },
  228. },
  229. 5000,
  230. );
  231. },
  232. currentSession() {
  233. let index = get().currentSessionIndex;
  234. const sessions = get().sessions;
  235. if (index < 0 || index >= sessions.length) {
  236. index = Math.min(sessions.length - 1, Math.max(0, index));
  237. set(() => ({ currentSessionIndex: index }));
  238. }
  239. const session = sessions[index];
  240. return session;
  241. },
  242. onNewMessage(message: ChatMessage) {
  243. get().updateCurrentSession((session) => {
  244. session.messages = session.messages.concat();
  245. session.lastUpdate = Date.now();
  246. });
  247. get().updateStat(message);
  248. get().summarizeSession();
  249. },
  250. async onUserInput(content: string, attachImages?: string[]) {
  251. const session = get().currentSession();
  252. const modelConfig = session.mask.modelConfig;
  253. const userContent = fillTemplateWith(content, modelConfig);
  254. console.log("[User Input] after template: ", userContent);
  255. let mContent: string | MultimodalContent[] = userContent;
  256. if (attachImages && attachImages.length > 0) {
  257. mContent = [
  258. {
  259. type: "text",
  260. text: userContent,
  261. },
  262. ];
  263. mContent = mContent.concat(
  264. attachImages.map((url) => {
  265. return {
  266. type: "image_url",
  267. image_url: {
  268. url: url,
  269. },
  270. };
  271. }),
  272. );
  273. }
  274. let userMessage: ChatMessage = createMessage({
  275. role: "user",
  276. content: mContent,
  277. });
  278. const botMessage: ChatMessage = createMessage({
  279. role: "assistant",
  280. streaming: true,
  281. model: modelConfig.model,
  282. });
  283. // get recent messages
  284. const recentMessages = get().getMessagesWithMemory();
  285. const sendMessages = recentMessages.concat(userMessage);
  286. const messageIndex = get().currentSession().messages.length + 1;
  287. // save user's and bot's message
  288. get().updateCurrentSession((session) => {
  289. const savedUserMessage = {
  290. ...userMessage,
  291. content: mContent,
  292. };
  293. session.messages = session.messages.concat([
  294. savedUserMessage,
  295. botMessage,
  296. ]);
  297. });
  298. var api: ClientApi;
  299. if (modelConfig.model.startsWith("gemini")) {
  300. api = new ClientApi(ModelProvider.GeminiPro);
  301. } else if (modelConfig.model.startsWith("claude")) {
  302. api = new ClientApi(ModelProvider.Claude);
  303. } else {
  304. api = new ClientApi(ModelProvider.GPT);
  305. }
  306. // make request
  307. api.llm.chat({
  308. messages: sendMessages,
  309. config: { ...modelConfig, stream: true },
  310. onUpdate(message) {
  311. botMessage.streaming = true;
  312. if (message) {
  313. botMessage.content = message;
  314. }
  315. get().updateCurrentSession((session) => {
  316. session.messages = session.messages.concat();
  317. });
  318. },
  319. onFinish(message) {
  320. botMessage.streaming = false;
  321. if (message) {
  322. botMessage.content = message;
  323. get().onNewMessage(botMessage);
  324. }
  325. ChatControllerPool.remove(session.id, botMessage.id);
  326. },
  327. onError(error) {
  328. const isAborted = error.message.includes("aborted");
  329. botMessage.content +=
  330. "\n\n" +
  331. prettyObject({
  332. error: true,
  333. message: error.message,
  334. });
  335. botMessage.streaming = false;
  336. userMessage.isError = !isAborted;
  337. botMessage.isError = !isAborted;
  338. get().updateCurrentSession((session) => {
  339. session.messages = session.messages.concat();
  340. });
  341. ChatControllerPool.remove(
  342. session.id,
  343. botMessage.id ?? messageIndex,
  344. );
  345. console.error("[Chat] failed ", error);
  346. },
  347. onController(controller) {
  348. // collect controller for stop/retry
  349. ChatControllerPool.addController(
  350. session.id,
  351. botMessage.id ?? messageIndex,
  352. controller,
  353. );
  354. },
  355. });
  356. },
  357. getMemoryPrompt() {
  358. const session = get().currentSession();
  359. return {
  360. role: "system",
  361. content:
  362. session.memoryPrompt.length > 0
  363. ? Locale.Store.Prompt.History(session.memoryPrompt)
  364. : "",
  365. date: "",
  366. } as ChatMessage;
  367. },
  368. getMessagesWithMemory() {
  369. const session = get().currentSession();
  370. const modelConfig = session.mask.modelConfig;
  371. const clearContextIndex = session.clearContextIndex ?? 0;
  372. const messages = session.messages.slice();
  373. const totalMessageCount = session.messages.length;
  374. // in-context prompts
  375. const contextPrompts = session.mask.context.slice();
  376. // system prompts, to get close to OpenAI Web ChatGPT
  377. const shouldInjectSystemPrompts =
  378. modelConfig.enableInjectSystemPrompts &&
  379. session.mask.modelConfig.model.startsWith("gpt-");
  380. var systemPrompts: ChatMessage[] = [];
  381. systemPrompts = shouldInjectSystemPrompts
  382. ? [
  383. createMessage({
  384. role: "system",
  385. content: fillTemplateWith("", {
  386. ...modelConfig,
  387. template: DEFAULT_SYSTEM_TEMPLATE,
  388. }),
  389. }),
  390. ]
  391. : [];
  392. if (shouldInjectSystemPrompts) {
  393. console.log(
  394. "[Global System Prompt] ",
  395. systemPrompts.at(0)?.content ?? "empty",
  396. );
  397. }
  398. // long term memory
  399. const shouldSendLongTermMemory =
  400. modelConfig.sendMemory &&
  401. session.memoryPrompt &&
  402. session.memoryPrompt.length > 0 &&
  403. session.lastSummarizeIndex > clearContextIndex;
  404. const longTermMemoryPrompts = shouldSendLongTermMemory
  405. ? [get().getMemoryPrompt()]
  406. : [];
  407. const longTermMemoryStartIndex = session.lastSummarizeIndex;
  408. // short term memory
  409. const shortTermMemoryStartIndex = Math.max(
  410. 0,
  411. totalMessageCount - modelConfig.historyMessageCount,
  412. );
  413. // lets concat send messages, including 4 parts:
  414. // 0. system prompt: to get close to OpenAI Web ChatGPT
  415. // 1. long term memory: summarized memory messages
  416. // 2. pre-defined in-context prompts
  417. // 3. short term memory: latest n messages
  418. // 4. newest input message
  419. const memoryStartIndex = shouldSendLongTermMemory
  420. ? Math.min(longTermMemoryStartIndex, shortTermMemoryStartIndex)
  421. : shortTermMemoryStartIndex;
  422. // and if user has cleared history messages, we should exclude the memory too.
  423. const contextStartIndex = Math.max(clearContextIndex, memoryStartIndex);
  424. const maxTokenThreshold = modelConfig.max_tokens;
  425. // get recent messages as much as possible
  426. const reversedRecentMessages = [];
  427. for (
  428. let i = totalMessageCount - 1, tokenCount = 0;
  429. i >= contextStartIndex && tokenCount < maxTokenThreshold;
  430. i -= 1
  431. ) {
  432. const msg = messages[i];
  433. if (!msg || msg.isError) continue;
  434. tokenCount += estimateTokenLength(getMessageTextContent(msg));
  435. reversedRecentMessages.push(msg);
  436. }
  437. // concat all messages
  438. const recentMessages = [
  439. ...systemPrompts,
  440. ...longTermMemoryPrompts,
  441. ...contextPrompts,
  442. ...reversedRecentMessages.reverse(),
  443. ];
  444. return recentMessages;
  445. },
  446. updateMessage(
  447. sessionIndex: number,
  448. messageIndex: number,
  449. updater: (message?: ChatMessage) => void,
  450. ) {
  451. const sessions = get().sessions;
  452. const session = sessions.at(sessionIndex);
  453. const messages = session?.messages;
  454. updater(messages?.at(messageIndex));
  455. set(() => ({ sessions }));
  456. },
  457. resetSession() {
  458. get().updateCurrentSession((session) => {
  459. session.messages = [];
  460. session.memoryPrompt = "";
  461. });
  462. },
  463. summarizeSession() {
  464. const config = useAppConfig.getState();
  465. const session = get().currentSession();
  466. const modelConfig = session.mask.modelConfig;
  467. var api: ClientApi;
  468. if (modelConfig.model.startsWith("gemini")) {
  469. api = new ClientApi(ModelProvider.GeminiPro);
  470. } else if (modelConfig.model.startsWith("claude")) {
  471. api = new ClientApi(ModelProvider.Claude);
  472. } else {
  473. api = new ClientApi(ModelProvider.GPT);
  474. }
  475. // remove error messages if any
  476. const messages = session.messages;
  477. // should summarize topic after chating more than 50 words
  478. const SUMMARIZE_MIN_LEN = 50;
  479. if (
  480. config.enableAutoGenerateTitle &&
  481. session.topic === DEFAULT_TOPIC &&
  482. countMessages(messages) >= SUMMARIZE_MIN_LEN
  483. ) {
  484. const topicMessages = messages.concat(
  485. createMessage({
  486. role: "user",
  487. content: Locale.Store.Prompt.Topic,
  488. }),
  489. );
  490. api.llm.chat({
  491. messages: topicMessages,
  492. config: {
  493. model: getSummarizeModel(session.mask.modelConfig.model),
  494. stream: false,
  495. },
  496. onFinish(message) {
  497. get().updateCurrentSession(
  498. (session) =>
  499. (session.topic =
  500. message.length > 0 ? trimTopic(message) : DEFAULT_TOPIC),
  501. );
  502. },
  503. });
  504. }
  505. const summarizeIndex = Math.max(
  506. session.lastSummarizeIndex,
  507. session.clearContextIndex ?? 0,
  508. );
  509. let toBeSummarizedMsgs = messages
  510. .filter((msg) => !msg.isError)
  511. .slice(summarizeIndex);
  512. const historyMsgLength = countMessages(toBeSummarizedMsgs);
  513. if (historyMsgLength > modelConfig?.max_tokens ?? 4000) {
  514. const n = toBeSummarizedMsgs.length;
  515. toBeSummarizedMsgs = toBeSummarizedMsgs.slice(
  516. Math.max(0, n - modelConfig.historyMessageCount),
  517. );
  518. }
  519. // add memory prompt
  520. toBeSummarizedMsgs.unshift(get().getMemoryPrompt());
  521. const lastSummarizeIndex = session.messages.length;
  522. console.log(
  523. "[Chat History] ",
  524. toBeSummarizedMsgs,
  525. historyMsgLength,
  526. modelConfig.compressMessageLengthThreshold,
  527. );
  528. if (
  529. historyMsgLength > modelConfig.compressMessageLengthThreshold &&
  530. modelConfig.sendMemory
  531. ) {
  532. /** Destruct max_tokens while summarizing
  533. * this param is just shit
  534. **/
  535. const { max_tokens, ...modelcfg } = modelConfig;
  536. api.llm.chat({
  537. messages: toBeSummarizedMsgs.concat(
  538. createMessage({
  539. role: "system",
  540. content: Locale.Store.Prompt.Summarize,
  541. date: "",
  542. }),
  543. ),
  544. config: {
  545. ...modelcfg,
  546. stream: true,
  547. model: getSummarizeModel(session.mask.modelConfig.model),
  548. },
  549. onUpdate(message) {
  550. session.memoryPrompt = message;
  551. },
  552. onFinish(message) {
  553. console.log("[Memory] ", message);
  554. get().updateCurrentSession((session) => {
  555. session.lastSummarizeIndex = lastSummarizeIndex;
  556. session.memoryPrompt = message; // Update the memory prompt for stored it in local storage
  557. });
  558. },
  559. onError(err) {
  560. console.error("[Summarize] ", err);
  561. },
  562. });
  563. }
  564. },
  565. updateStat(message: ChatMessage) {
  566. get().updateCurrentSession((session) => {
  567. session.stat.charCount += message.content.length;
  568. // TODO: should update chat count and word count
  569. });
  570. },
  571. updateCurrentSession(updater: (session: ChatSession) => void) {
  572. const sessions = get().sessions;
  573. const index = get().currentSessionIndex;
  574. updater(sessions[index]);
  575. set(() => ({ sessions }));
  576. },
  577. clearAllData() {
  578. localStorage.clear();
  579. location.reload();
  580. },
  581. };
  582. return methods;
  583. },
  584. {
  585. name: StoreKey.Chat,
  586. version: 3.1,
  587. migrate(persistedState, version) {
  588. const state = persistedState as any;
  589. const newState = JSON.parse(
  590. JSON.stringify(state),
  591. ) as typeof DEFAULT_CHAT_STATE;
  592. if (version < 2) {
  593. newState.sessions = [];
  594. const oldSessions = state.sessions;
  595. for (const oldSession of oldSessions) {
  596. const newSession = createEmptySession();
  597. newSession.topic = oldSession.topic;
  598. newSession.messages = [...oldSession.messages];
  599. newSession.mask.modelConfig.sendMemory = true;
  600. newSession.mask.modelConfig.historyMessageCount = 4;
  601. newSession.mask.modelConfig.compressMessageLengthThreshold = 1000;
  602. newState.sessions.push(newSession);
  603. }
  604. }
  605. if (version < 3) {
  606. // migrate id to nanoid
  607. newState.sessions.forEach((s) => {
  608. s.id = nanoid();
  609. s.messages.forEach((m) => (m.id = nanoid()));
  610. });
  611. }
  612. // Enable `enableInjectSystemPrompts` attribute for old sessions.
  613. // Resolve issue of old sessions not automatically enabling.
  614. if (version < 3.1) {
  615. newState.sessions.forEach((s) => {
  616. if (
  617. // Exclude those already set by user
  618. !s.mask.modelConfig.hasOwnProperty("enableInjectSystemPrompts")
  619. ) {
  620. // Because users may have changed this configuration,
  621. // the user's current configuration is used instead of the default
  622. const config = useAppConfig.getState();
  623. s.mask.modelConfig.enableInjectSystemPrompts =
  624. config.modelConfig.enableInjectSystemPrompts;
  625. }
  626. });
  627. }
  628. return newState as any;
  629. },
  630. },
  631. );