Create different instructions for Docker and bare metal
This commit is contained in:
parent
a84b68ec32
commit
e912042490
60
BARE.md
Normal file
60
BARE.md
Normal file
|
@ -0,0 +1,60 @@
|
|||
# Running Tailchat Assistant normally
|
||||
|
||||
1. Clone the project and install dependencies:
|
||||
```bash
|
||||
git clone https://git.povario.com/powermaker450/Tailchat-Assistant && cd Tailchat-Assistant
|
||||
|
||||
npm i
|
||||
```
|
||||
2. Create the `.env` file from the example file:
|
||||
```bash
|
||||
mv .env.example .env
|
||||
```
|
||||
3. **Make sure that the "Openapi Platform Plugin" and "App Integration" plugins are enabled.** Then [follow the instructions here to create an App.](https://tailchat.msgbyte.com/docs/advanced-usage/openapp/create)
|
||||
4. Open up the `.env` file with your preffered text editor and populate the file:
|
||||
```bash
|
||||
HOST=http://localhost:11000
|
||||
# wherever your Tailchat server is running
|
||||
|
||||
ID=
|
||||
# your bot ID
|
||||
|
||||
SECRET=
|
||||
# Your bot secret
|
||||
|
||||
API_ENDPOINT=http://localhost:8080/v1
|
||||
# The OpenAI-compatible endpoint for the bot to send messsages to. Defaults to "http://localhost:8080/v1"
|
||||
# e.x. "http://localhost:8080/v1", "https://api.openai.com/v1"
|
||||
|
||||
API_KEY=none
|
||||
# Your API key here for OpenAI/LocalAI. Defaults to the string "none".
|
||||
# MAKE SURE to fill this in with your OpenAI API Key or a key you may have set for LocalAI.
|
||||
# If you didn't set the API key for LocalAI, you may leave this blank.
|
||||
|
||||
TEXT_MODEL=gpt-4
|
||||
# The model to query when sending text messages. Defaults to "gpt-4"
|
||||
# e.x. "gpt-3", "gpt-4"
|
||||
|
||||
CREATE_IMAGE_MODEL=stablediffusion-cpp
|
||||
# The model to use when creating images. Defaults to "stablediffusion-cpp",
|
||||
# e.x. "dall-e-3", "stablediffusion-cpp"
|
||||
|
||||
ANALYZE_IMAGE_MODEL=gpt-4-vision-preview
|
||||
# The model to use when analyzing images. Defaults to "gpt4-vision-preview".
|
||||
# e.x. "gpt-4-vision-preview", "llava"
|
||||
|
||||
ALLOWED_CHAT=
|
||||
# The ID of the channel that the bot is allowed to respond in.
|
||||
# The bot will always respond to Direct Messages.
|
||||
|
||||
SAFE_WORD=\
|
||||
# When this character/string is detected anywhere in a message, the bot won't respond to it.
|
||||
# Defaults to "\".
|
||||
```
|
||||
|
||||
5. After completing these steps and making any desired changes, build and run the project:
|
||||
```bash
|
||||
npm run build
|
||||
|
||||
npm run start
|
||||
```
|
61
DOCKER.md
Normal file
61
DOCKER.md
Normal file
|
@ -0,0 +1,61 @@
|
|||
# Running Tailchat Assistant in Docker
|
||||
|
||||
1. Clone the project:
|
||||
```bash
|
||||
git clone https://git.povario.com/powermaker450/Tailchat-Assistant && cd Tailchat-Assistant
|
||||
```
|
||||
|
||||
2. Create the `.env` file from the example file:
|
||||
```bash
|
||||
mv .env.example .env
|
||||
```
|
||||
3. **Make sure that the "Openapi Platform Plugin" and "App Integration" plugins are enabled.** Then [follow the instructions here to create an App.](https://tailchat.msgbyte.com/docs/advanced-usage/openapp/create)
|
||||
4. Open up the `.env` file with your preffered text editor and populate the file:
|
||||
**NOTE: Due to the way Docker organizes containers and networks, your `HOST` and `API_ENDPOINT` variables will likely not be `localhost`. Use the container names or IP's instead.**
|
||||
|
||||
```bash
|
||||
HOST=http://tailchat-core:11000
|
||||
# wherever your Tailchat server is running
|
||||
|
||||
ID=
|
||||
# your bot ID
|
||||
|
||||
SECRET=
|
||||
# Your bot secret
|
||||
|
||||
API_ENDPOINT=http://localai:8080/v1
|
||||
# The OpenAI-compatible endpoint for the bot to send messsages to. Defaults to "http://localhost:8080/v1"
|
||||
# e.x. "http://localhost:8080/v1", "https://api.openai.com/v1"
|
||||
|
||||
API_KEY=none
|
||||
# Your API key here for OpenAI/LocalAI. Defaults to the string "none".
|
||||
# MAKE SURE to fill this in with your OpenAI API Key or a key you may have set for LocalAI.
|
||||
# If you didn't set the API key for LocalAI, you may leave this blank.
|
||||
|
||||
TEXT_MODEL=gpt-4
|
||||
# The model to query when sending text messages. Defaults to "gpt-4"
|
||||
# e.x. "gpt-3", "gpt-4"
|
||||
|
||||
CREATE_IMAGE_MODEL=stablediffusion-cpp
|
||||
# The model to use when creating images. Defaults to "stablediffusion-cpp",
|
||||
# e.x. "dall-e-3", "stablediffusion-cpp"
|
||||
|
||||
ANALYZE_IMAGE_MODEL=gpt-4-vision-preview
|
||||
# The model to use when analyzing images. Defaults to "gpt4-vision-preview".
|
||||
# e.x. "gpt-4-vision-preview", "llava"
|
||||
|
||||
ALLOWED_CHAT=
|
||||
# The ID of the channel that the bot is allowed to respond in.
|
||||
# The bot will always respond to Direct Messages.
|
||||
|
||||
SAFE_WORD=\
|
||||
# When this character/string is detected anywhere in a message, the bot won't respond to it.
|
||||
# Defaults to "\".
|
||||
```
|
||||
|
||||
5. After populating the .env file, start the container:
|
||||
```bash
|
||||
docker compose up -d
|
||||
```
|
||||
|
||||
Docker will build the container image and start the project. Use `docker compose logs -f` to make sure everything is working and no errors have occured.
|
61
README.md
61
README.md
|
@ -28,61 +28,6 @@ Reach out to me at [contact@povario.com](mailto:contact@povario.com) if you enco
|
|||
---
|
||||
|
||||
# Getting started
|
||||
1. Clone the project and install dependencies:
|
||||
```bash
|
||||
git clone https://git.povario.com/powermaker450/Tailchat-Assistant && cd Tailchat-Assistant
|
||||
|
||||
npm i
|
||||
```
|
||||
2. Create the `.env` file from the example file:
|
||||
```bash
|
||||
mv .env.example .env
|
||||
```
|
||||
3. **Make sure that the "Openapi Platform Plugin" and "App Integration" plugins are enabled.** Then [follow the instructions here to create an App.](https://tailchat.msgbyte.com/docs/advanced-usage/openapp/create)
|
||||
4. Open up the `.env` file with your preffered text editor and populate the file:
|
||||
```bash
|
||||
HOST=http://localhost:11000
|
||||
# wherever your Tailchat server is running
|
||||
|
||||
ID=
|
||||
# your bot ID
|
||||
|
||||
SECRET=
|
||||
# Your bot secret
|
||||
|
||||
API_ENDPOINT=http://localhost:8080/v1
|
||||
# The OpenAI-compatible endpoint for the bot to send messsages to. Defaults to "http://localhost:8080/v1"
|
||||
# e.x. "http://localhost:8080/v1", "https://api.openai.com/v1"
|
||||
|
||||
API_KEY=none
|
||||
# Your API key here for OpenAI/LocalAI. Defaults to the string "none".
|
||||
# MAKE SURE to fill this in with your OpenAI API Key or a key you may have set for LocalAI.
|
||||
# If you didn't set the API key for LocalAI, you may leave this blank.
|
||||
|
||||
TEXT_MODEL=gpt-4
|
||||
# The model to query when sending text messages. Defaults to "gpt-4"
|
||||
# e.x. "gpt-3", "gpt-4"
|
||||
|
||||
CREATE_IMAGE_MODEL=stablediffusion-cpp
|
||||
# The model to use when creating images. Defaults to "stablediffusion-cpp",
|
||||
# e.x. "dall-e-3", "stablediffusion-cpp"
|
||||
|
||||
ANALYZE_IMAGE_MODEL=gpt-4-vision-preview
|
||||
# The model to use when analyzing images. Defaults to "gpt4-vision-preview".
|
||||
# e.x. "gpt-4-vision-preview", "llava"
|
||||
|
||||
ALLOWED_CHAT=
|
||||
# The ID of the channel that the bot is allowed to respond in.
|
||||
# The bot will always respond to Direct Messages.
|
||||
|
||||
SAFE_WORD=\
|
||||
# When this character/string is detected anywhere in a message, the bot won't respond to it.
|
||||
# Defaults to "\".
|
||||
```
|
||||
|
||||
5. After completing these steps and making any desired changes, build and run the project:
|
||||
```bash
|
||||
npm run build
|
||||
|
||||
npm run start
|
||||
```
|
||||
**Choose how you want to run the bot:**
|
||||
- [Normally (with `npm`)](https://git.povario.com/powermaker450/Tailchat-Assistant/src/branch/main/BARE.md)
|
||||
- [Docker](https://git.povario.com/powermaker450/Tailchat-Assistant/src/branch/main/DOCKER.md)
|
||||
|
|
Loading…
Reference in a new issue