// AI.TS // by ihatenodejs/Aidan // // ----------------------------------------------------------------------- // // This is free and unencumbered software released into the public domain. // // Anyone is free to copy, modify, publish, use, compile, sell, or // distribute this software, either in source code form or as a compiled // binary, for any purpose, commercial or non-commercial, and by any // means. // // In jurisdictions that recognize copyright laws, the author or authors // of this software dedicate any and all copyright interest in the // software to the public domain. We make this dedication for the benefit // of the public at large and to the detriment of our heirs and // successors. We intend this dedication to be an overt act of // relinquishment in perpetuity of all present and future rights to this // software under copyright law. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, // EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF // MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. // IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR // OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, // ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR // OTHER DEALINGS IN THE SOFTWARE. // // For more information, please refer to import { isOnSpamWatch } from "../spamwatch/spamwatch" import spamwatchMiddlewareModule from "../spamwatch/Middleware" import { Telegraf, Context } from "telegraf" import type { Message } from "telegraf/types" import { replyToMessageId } from "../utils/reply-to-message-id" import { getStrings } from "../plugins/checklang" import { languageCode } from "../utils/language-code" import axios from "axios" import { rateLimiter } from "../utils/rate-limiter" import { logger } from "../utils/log" const spamwatchMiddleware = spamwatchMiddlewareModule(isOnSpamWatch) export const flash_model = "gemma3:4b" export const thinking_model = "deepseek-r1:1.5b" type TextContext = Context & { message: Message.TextMessage } export function sanitizeForJson(text: string): string { return text .replace(/\\/g, '\\\\') .replace(/"/g, '\\"') .replace(/\n/g, '\\n') .replace(/\r/g, '\\r') .replace(/\t/g, '\\t') } export async function preChecks() { const envs = [ "ollamaApi", ] for (const env of envs) { if (!process.env[env]) { console.error(`[✨ AI | !] ❌ ${env} not set!`) return false } } console.log("[✨ AI] Pre-checks passed\n") return true } async function getResponse(prompt: string, ctx: TextContext, replyGenerating: Message, model: string) { const Strings = getStrings(languageCode(ctx)) if (!ctx.chat) return { "success": false, "error": Strings.unexpectedErr.replace("{error}", "No chat found"), } try { const aiResponse = await axios.post(`${process.env.ollamaApi}/api/generate`, { model: model, prompt: prompt, stream: true, }, { responseType: "stream", }) let fullResponse = "" let thoughts = "" let lastUpdate = Date.now() for await (const chunk of aiResponse.data) { const lines = chunk.toString().split('\n') for (const line of lines) { if (!line.trim()) continue let ln = JSON.parse(line) if (model === thinking_model && ln.response.includes('')) { const thinkMatch = ln.response.match(/([\s\S]*?)<\/think>/) if (thinkMatch) { const innerContent = thinkMatch[1] if (innerContent.trim().length > 0) { logger.logThinking(ctx.chat.id, replyGenerating.message_id, true) } } else { logger.logThinking(ctx.chat.id, replyGenerating.message_id, true) } } else if (model === thinking_model && ln.response.includes('')) { logger.logThinking(ctx.chat.id, replyGenerating.message_id, false) } try { const now = Date.now() if (ln.response) { if (model === thinking_model) { let patchedThoughts = ln.response // TODO: hide blank thinking chunks const thinkTagRx = /([\s\S]*?)<\/think>/g patchedThoughts = patchedThoughts.replace(thinkTagRx, (match, p1) => { if (p1.trim().length > 0) { console.log(p1) return '`Thinking...`' + p1 + '`Finished thinking`' } else { return '' } }) patchedThoughts = patchedThoughts.replace(//g, '`Thinking...`') patchedThoughts = patchedThoughts.replace(/<\/think>/g, '`Finished thinking`') thoughts += patchedThoughts fullResponse += patchedThoughts } else { fullResponse += ln.response } if (now - lastUpdate >= 1000) { await rateLimiter.editMessageWithRetry( ctx, ctx.chat.id, replyGenerating.message_id, thoughts, { parse_mode: 'Markdown' } ) lastUpdate = now } } } catch (e) { console.error("[✨ AI | !] Error parsing chunk:", e) } } } return { "success": true, "response": fullResponse, } } catch (error: any) { let shouldPullModel = false if (error.response?.data?.error) { if (error.response.data.error.includes(`model '${model}' not found`) || error.status === 404) { shouldPullModel = true } else { console.error("[✨ AI | !] Error zone 1:", error.response.data.error) return { "success": false, "error": error.response.data.error, } } } else if (error.status === 404) { shouldPullModel = true } if (shouldPullModel) { ctx.telegram.editMessageText(ctx.chat.id, replyGenerating.message_id, undefined, `🔄 Pulling ${model} from ollama...\n\nThis may take a few minutes...`) console.log(`[✨ AI | i] Pulling ${model} from ollama...`) let pullModelStream: any try { pullModelStream = await axios.post(`${process.env.ollamaApi}/api/pull`, { model: model, stream: false, timeout: process.env.ollamaApiTimeout || 10000, }) } catch (e: any) { console.error("[✨ AI | !] Something went wrong:", e.response.data.error) return { "success": false, "error": `❌ Something went wrong while pulling ${model}, please try your command again!`, } } console.log(`[✨ AI | i] ${model} pulled successfully`) return { "success": true, "response": `✅ Pulled ${model} successfully, please retry the command.`, } } if (error.response) { console.error("[✨ AI | !] Error zone 2:", error.response) return { "success": false, "error": error.response, } } if (error.statusText) { console.error("[✨ AI | !] Error zone 3:", error.statusText) return { "success": false, "error": error.statusText, } } return { "success": false, "error": "An unexpected error occurred", } } } export default (bot: Telegraf) => { const botName = bot.botInfo?.first_name && bot.botInfo?.last_name ? `${bot.botInfo.first_name} ${bot.botInfo.last_name}` : "Kowalski" bot.command(["ask", "think"], spamwatchMiddleware, async (ctx) => { if (!ctx.message || !('text' in ctx.message)) return; const isAsk = ctx.message.text.startsWith("/ask") const model = isAsk ? flash_model : thinking_model const textCtx = ctx as TextContext; const reply_to_message_id = replyToMessageId(textCtx) const Strings = getStrings(languageCode(textCtx)) const message = textCtx.message.text const author = ("@" + ctx.from?.username) || ctx.from?.first_name logger.logCmdStart( author, model === flash_model ? "ask" : "think" ) if (!process.env.ollamaApi) { await ctx.reply(Strings.aiDisabled, { parse_mode: 'Markdown', ...({ reply_to_message_id }) }) return } const replyGenerating = await ctx.reply(Strings.askGenerating.replace("{model}", model), { parse_mode: 'Markdown', ...({ reply_to_message_id }) }) const fixedMsg = message.replace(/\/ask /, "").replace(/\/think /, "") if (fixedMsg.length < 1) { await ctx.reply(Strings.askNoMessage, { parse_mode: 'Markdown', ...({ reply_to_message_id }) }) return } logger.logPrompt(fixedMsg) const prompt = sanitizeForJson( `You are a plaintext-only, helpful assistant called ${botName}. Current Date/Time (UTC): ${new Date().toLocaleString()} --- Respond to the user's message: ${fixedMsg}`) const aiResponse = await getResponse(prompt, textCtx, replyGenerating, model) if (!aiResponse) return if (aiResponse.success && aiResponse.response) { if (!ctx.chat) return await rateLimiter.editMessageWithRetry( ctx, ctx.chat.id, replyGenerating.message_id, aiResponse.response, { parse_mode: 'Markdown' } ) } else { if (!ctx.chat) return const error = Strings.unexpectedErr.replace("{error}", aiResponse.error) await rateLimiter.editMessageWithRetry( ctx, ctx.chat.id, replyGenerating.message_id, error, { parse_mode: 'Markdown' } ) } }) }