Skip to content

Files

116 lines (59 loc) · 6.94 KB

usage.md

File metadata and controls

116 lines (59 loc) · 6.94 KB

📖 Usage

This document covers how to use the bot in a room.

The 🌟 Features page also includes details about how each feature works and can be configured.

💬 Text Generation

This is related to the 💬 Text Generation feature.

If there's a text-generation handler agent configured, the bot may respond to messages sent in the room.

Some models also support vision, so you may be able to mix text and images in the same conversation.

See screenshots of:

Whether the bot responds depends on:

Room messages start a threaded conversation where you can continue back-and-forth communication with the bot. Using on-demand involvement, you can can also mention the bot to provoke it to get involved in any conversation thread or reply chain.

Unless you've enabled the ♻️ Context Management feature, all messages will be sent to the agent's API each time. If the context management feature is enabled, older messages may be dropped.

🗣️ Text-to-Speech

This is related to the 🗣️ Text-to-Speech feature.

If there's a text-to-speech handler agent configured, the bot may convert text messages sent to the room to audio (voice).

See:

By default, the bot:

🦻 Speech-to-Text

This is related to the 🦻 Speech-to-Text feature.

If there's a speech-to-text handler agent configured, the bot may transcribe voice messages sent to the room to text.

See a 🖼️ Screenshot of the default flow for Speech-to-Text and Text-Generation.

The speech-to-text feature triggers automatically by default, but can be adjusted via the 🦻 Speech-to-Text / 🪄 Flow Type setting.

If all your messages are in the same language, you can improve accuracy & latency by configuring the language (see 🦻 Speech-to-Text / 🔤 Language).

Image Generation

This feature is not configurable at the moment. The configuration (size, quality, style) specified at the 🤖 agent level will be used.

Capabilities depend on the ☁️ provider and model used.

🖌️ Creating images

Simply send a command like !bai image create A beautiful sunset over the ocean and the bot will start a threaded conversation and post an image based on your prompt.

See a 🖼️ Screenshot of the Image Creation feature.

You can then respond in the same message thread with:

  • more messages, to add more criteria to your prompt.
  • a message saying again, to generate one more image with the current prompt.

🎨 Editing images

Simply send a command like !bai image edit Turn the following image into an anime-style drawing and the bot will start a threaded conversation asking for more details.

See a 🖼️ Screenshot of the Image Editing feature (manipulating a single image) and a 🖼️ Screenshot of the Image Editing feature (manipulating multiple images).

You can then respond in the same message thread with:

  • more messages, to add more criteria to your prompt.
  • one or more images, to provide the images that the bot will operate on.
  • a message saying go, to start the image generation process.
  • a message saying again, to prompt the bot to generate one more image edit with the current prompt.

🫵 Creating stickers

A variation of creating images is creating "sticker images".

See a 🖼️ Screenshot of the Sticker Creation feature.

To create a sticker, send a command like !bai sticker A huge ramen bowl with lots of chashu and a mountain of beansprouts on top.

The difference from creating images is that the bot will:

  • generate a smaller-resolution image (currently hardcoded to 256x256) - smaller/quicker, but still good enough for a sticker
  • potentially switch to a different (cheaper or otherwise more suitable) model, if available
  • post the image directly to the room (as a reply to your message), without starting a threaded conversation

Some models (like OpenAI's Dall-E-3) can only generate larger images (1024x1024, etc., for a higher charge), so we switching to a smaller/cheaper model (like Dall-E-2) is a way to generate a sticker cheaply.