diff --git a/src/bot.ts b/src/bot.ts index bf20d8b3..e15c4a8a 100644 --- a/src/bot.ts +++ b/src/bot.ts @@ -527,7 +527,6 @@ bot.command('support', async (ctx) => { bot.command('models', async (ctx) => { const models = llmModelManager.generateTelegramOutput() - console.log(models) writeCommandLog(ctx as OnMessageContext).catch(logErrorHandler) return await ctx.reply(models, { parse_mode: 'Markdown', diff --git a/src/modules/llms/dalleBot.ts b/src/modules/llms/dalleBot.ts index 0348f3ce..03c10483 100644 --- a/src/modules/llms/dalleBot.ts +++ b/src/modules/llms/dalleBot.ts @@ -160,6 +160,7 @@ export class DalleBot extends LlmsBase { ctx.transient.analytics.module = this.module const session = this.getSession(ctx) const isSupportedEvent = this.isSupportedEvent(ctx) + if (!isSupportedEvent && ctx.chat?.type !== 'private') { this.logger.warn(`### unsupported command ${ctx.message?.text}`) return diff --git a/src/modules/llms/utils/llmsData.ts b/src/modules/llms/utils/llmsData.ts index 4d6a8356..4b5a4f60 100644 --- a/src/modules/llms/utils/llmsData.ts +++ b/src/modules/llms/utils/llmsData.ts @@ -110,7 +110,7 @@ export const llmData: LLMData = { fullName: 'GPT-4 Vision', botName: 'OpenAIBot', version: 'gpt-4-vision-preview', - commands: ['vision, v'], + commands: ['vision', 'v'], prefix: ['v. '], apiSpec: 'https://platform.openai.com/docs/guides/vision', inputPrice: 0.03,