From 2ead99a06bda5cc9942523c8d926df4951d2f037 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sat, 5 Aug 2023 23:11:23 +0800
Subject: [PATCH 01/15] refactor code structure and remove unused

---
 .dockerignore                          |  47 +-
 .env.example                           |   5 +-
 .gitignore                             | 279 +++++----
 .vscode/settings.json                  |   3 -
 CHANGELOG.md                           |   7 +
 Dockerfile                             |  32 +-
 LICENSE                                |  42 +-
 README.md                              |  86 ++-
 bard.py                                | 104 ---
 bing.py                                |  64 --
 compose.yaml                           |   8 -
 config.json.example                    |  19 +-
 requirements.txt                       |  31 +-
 BingImageGen.py => src/BingImageGen.py | 330 +++++-----
 askgpt.py => src/askgpt.py             |  92 +--
 bot.py => src/bot.py                   | 836 ++++++++++++-------------
 log.py => src/log.py                   |  60 +-
 main.py => src/main.py                 | 123 ++--
 pandora.py => src/pandora.py           |  27 +-
 v3.py                                  | 324 ----------
 20 files changed, 1004 insertions(+), 1515 deletions(-)
 delete mode 100644 .vscode/settings.json
 create mode 100644 CHANGELOG.md
 delete mode 100644 bard.py
 delete mode 100644 bing.py
 rename BingImageGen.py => src/BingImageGen.py (97%)
 rename askgpt.py => src/askgpt.py (88%)
 rename bot.py => src/bot.py (64%)
 rename log.py => src/log.py (96%)
 rename main.py => src/main.py (64%)
 rename pandora.py => src/pandora.py (86%)
 delete mode 100644 v3.py

diff --git a/.dockerignore b/.dockerignore
index 964c822..7495bae 100644
--- a/.dockerignore
+++ b/.dockerignore
@@ -1,21 +1,26 @@
-.gitignore
-images
-*.md
-Dockerfile
-Dockerfile-dev
-.dockerignore
-config.json
-config.json.sample
-.vscode
-bot.log
-venv
-.venv
-*.yaml
-*.yml
-.git
-.idea
-__pycache__
-.env
-.env.example
-.github
-settings.js
\ No newline at end of file
+.gitignore
+images
+*.md
+Dockerfile
+Dockerfile-dev
+compose.yaml
+compose-dev.yaml
+.dockerignore
+config.json
+config.json.sample
+.vscode
+bot.log
+venv
+.venv
+*.yaml
+*.yml
+.git
+.idea
+__pycache__
+src/__pycache__
+.env
+.env.example
+.github
+settings.js
+mattermost-server
+tests
\ No newline at end of file
diff --git a/.env.example b/.env.example
index 1b175cc..65f8e92 100644
--- a/.env.example
+++ b/.env.example
@@ -2,8 +2,7 @@ SERVER_URL="xxxxx.xxxxxx.xxxxxxxxx"
 ACCESS_TOKEN="xxxxxxxxxxxxxxxxx"
 USERNAME="@chatgpt"
 OPENAI_API_KEY="sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
-BING_API_ENDPOINT="http://api:3000/conversation"
-BARD_TOKEN="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx."
 BING_AUTH_COOKIE="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
 PANDORA_API_ENDPOINT="http://pandora:8008"
-PANDORA_API_MODEL="text-davinci-002-render-sha-mobile"
\ No newline at end of file
+PANDORA_API_MODEL="text-davinci-002-render-sha-mobile"
+GPT_ENGINE="gpt-3.5-turbo"
\ No newline at end of file
diff --git a/.gitignore b/.gitignore
index 8827e10..bd51f1a 100644
--- a/.gitignore
+++ b/.gitignore
@@ -1,139 +1,140 @@
-# Byte-compiled / optimized / DLL files
-__pycache__/
-*.py[cod]
-*$py.class
-
-# C extensions
-*.so
-
-# Distribution / packaging
-.Python
-build/
-develop-eggs/
-dist/
-downloads/
-eggs/
-.eggs/
-lib/
-lib64/
-parts/
-sdist/
-var/
-wheels/
-pip-wheel-metadata/
-share/python-wheels/
-*.egg-info/
-.installed.cfg
-*.egg
-MANIFEST
-
-# PyInstaller
-#  Usually these files are written by a python script from a template
-#  before PyInstaller builds the exe, so as to inject date/other infos into it.
-*.manifest
-*.spec
-
-# Installer logs
-pip-log.txt
-pip-delete-this-directory.txt
-
-# Unit test / coverage reports
-htmlcov/
-.tox/
-.nox/
-.coverage
-.coverage.*
-.cache
-nosetests.xml
-coverage.xml
-*.cover
-*.py,cover
-.hypothesis/
-.pytest_cache/
-
-# Translations
-*.mo
-*.pot
-
-# Django stuff:
-*.log
-local_settings.py
-db.sqlite3
-db.sqlite3-journal
-
-# Flask stuff:
-instance/
-.webassets-cache
-
-# Scrapy stuff:
-.scrapy
-
-# Sphinx documentation
-docs/_build/
-
-# PyBuilder
-target/
-
-# Jupyter Notebook
-.ipynb_checkpoints
-
-# IPython
-profile_default/
-ipython_config.py
-
-# pyenv
-.python-version
-
-# pipenv
-#   According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
-#   However, in case of collaboration, if having platform-specific dependencies or dependencies
-#   having no cross-platform support, pipenv may install dependencies that don't work, or not
-#   install all needed dependencies.
-#Pipfile.lock
-
-# PEP 582; used by e.g. github.com/David-OConnor/pyflow
-__pypackages__/
-
-# Celery stuff
-celerybeat-schedule
-celerybeat.pid
-
-# SageMath parsed files
-*.sage.py
-
-# custom path
-images
-Dockerfile-dev
-compose-dev.yaml
-settings.js
-
-# Environments
-.env
-.venv
-env/
-venv/
-ENV/
-env.bak/
-venv.bak/
-config.json
-
-# Spyder project settings
-.spyderproject
-.spyproject
-
-# Rope project settings
-.ropeproject
-
-# mkdocs documentation
-/site
-
-# mypy
-.mypy_cache/
-.dmypy.json
-dmypy.json
-
-# Pyre type checker
-.pyre/
-
-# custom
-compose-local-dev.yaml
\ No newline at end of file
+# Byte-compiled / optimized / DLL files
+__pycache__/
+*.py[cod]
+*$py.class
+
+# C extensions
+*.so
+
+# Distribution / packaging
+.Python
+build/
+develop-eggs/
+dist/
+downloads/
+eggs/
+.eggs/
+lib/
+lib64/
+parts/
+sdist/
+var/
+wheels/
+pip-wheel-metadata/
+share/python-wheels/
+*.egg-info/
+.installed.cfg
+*.egg
+MANIFEST
+
+# PyInstaller
+#  Usually these files are written by a python script from a template
+#  before PyInstaller builds the exe, so as to inject date/other infos into it.
+*.manifest
+*.spec
+
+# Installer logs
+pip-log.txt
+pip-delete-this-directory.txt
+
+# Unit test / coverage reports
+htmlcov/
+.tox/
+.nox/
+.coverage
+.coverage.*
+.cache
+nosetests.xml
+coverage.xml
+*.cover
+*.py,cover
+.hypothesis/
+.pytest_cache/
+
+# Translations
+*.mo
+*.pot
+
+# Django stuff:
+*.log
+local_settings.py
+db.sqlite3
+db.sqlite3-journal
+
+# Flask stuff:
+instance/
+.webassets-cache
+
+# Scrapy stuff:
+.scrapy
+
+# Sphinx documentation
+docs/_build/
+
+# PyBuilder
+target/
+
+# Jupyter Notebook
+.ipynb_checkpoints
+
+# IPython
+profile_default/
+ipython_config.py
+
+# pyenv
+.python-version
+
+# pipenv
+#   According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
+#   However, in case of collaboration, if having platform-specific dependencies or dependencies
+#   having no cross-platform support, pipenv may install dependencies that don't work, or not
+#   install all needed dependencies.
+#Pipfile.lock
+
+# PEP 582; used by e.g. github.com/David-OConnor/pyflow
+__pypackages__/
+
+# Celery stuff
+celerybeat-schedule
+celerybeat.pid
+
+# SageMath parsed files
+*.sage.py
+
+# custom path
+images
+Dockerfile-dev
+compose-dev.yaml
+settings.js
+
+# Environments
+.env
+.venv
+env/
+venv/
+ENV/
+env.bak/
+venv.bak/
+config.json
+
+# Spyder project settings
+.spyderproject
+.spyproject
+
+# Rope project settings
+.ropeproject
+
+# mkdocs documentation
+/site
+
+# mypy
+.mypy_cache/
+.dmypy.json
+dmypy.json
+
+# Pyre type checker
+.pyre/
+
+# custom
+compose-dev.yaml
+mattermost-server
\ No newline at end of file
diff --git a/.vscode/settings.json b/.vscode/settings.json
deleted file mode 100644
index 324d91c..0000000
--- a/.vscode/settings.json
+++ /dev/null
@@ -1,3 +0,0 @@
-{
-    "python.formatting.provider": "black"
-}
\ No newline at end of file
diff --git a/CHANGELOG.md b/CHANGELOG.md
new file mode 100644
index 0000000..64e9ef2
--- /dev/null
+++ b/CHANGELOG.md
@@ -0,0 +1,7 @@
+# Changelog
+
+## v1.0.4
+
+- refactor code structure and remove unused
+- remove Bing AI and Google Bard due to technical problems
+- bug fix and improvement
\ No newline at end of file
diff --git a/Dockerfile b/Dockerfile
index 688f7a7..d96a624 100644
--- a/Dockerfile
+++ b/Dockerfile
@@ -1,16 +1,16 @@
-FROM python:3.11-alpine as base
-
-FROM base as builder
-# RUN sed -i 's|v3\.\d*|edge|' /etc/apk/repositories
-RUN apk update && apk add --no-cache gcc musl-dev libffi-dev git
-COPY requirements.txt .
-RUN pip install -U pip setuptools wheel && pip install --user -r ./requirements.txt && rm ./requirements.txt
-
-FROM base as runner
-RUN apk update && apk add --no-cache libffi-dev
-COPY --from=builder /root/.local /usr/local
-COPY . /app
-
-FROM runner
-WORKDIR /app
-CMD ["python", "main.py"]
+FROM python:3.11-alpine as base
+
+FROM base as builder
+# RUN sed -i 's|v3\.\d*|edge|' /etc/apk/repositories
+RUN apk update && apk add --no-cache gcc musl-dev libffi-dev git
+COPY requirements.txt .
+RUN pip install -U pip setuptools wheel && pip install --user -r ./requirements.txt && rm ./requirements.txt
+
+FROM base as runner
+RUN apk update && apk add --no-cache libffi-dev
+COPY --from=builder /root/.local /usr/local
+COPY . /app
+
+FROM runner
+WORKDIR /app
+CMD ["python", "src/main.py"]
diff --git a/LICENSE b/LICENSE
index 92612e2..7d299c6 100644
--- a/LICENSE
+++ b/LICENSE
@@ -1,21 +1,21 @@
-MIT License
-
-Copyright (c) 2023 BobMaster
-
-Permission is hereby granted, free of charge, to any person obtaining a copy
-of this software and associated documentation files (the "Software"), to deal
-in the Software without restriction, including without limitation the rights
-to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
-copies of the Software, and to permit persons to whom the Software is
-furnished to do so, subject to the following conditions:
-
-The above copyright notice and this permission notice shall be included in all
-copies or substantial portions of the Software.
-
-THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
-IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
-FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
-AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
-LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
-OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
-SOFTWARE.
+MIT License
+
+Copyright (c) 2023 BobMaster
+
+Permission is hereby granted, free of charge, to any person obtaining a copy
+of this software and associated documentation files (the "Software"), to deal
+in the Software without restriction, including without limitation the rights
+to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
+copies of the Software, and to permit persons to whom the Software is
+furnished to do so, subject to the following conditions:
+
+The above copyright notice and this permission notice shall be included in all
+copies or substantial portions of the Software.
+
+THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
+AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+SOFTWARE.
diff --git a/README.md b/README.md
index cbd01eb..be8493a 100644
--- a/README.md
+++ b/README.md
@@ -1,44 +1,42 @@
-## Introduction
-
-This is a simple Mattermost Bot that uses OpenAI's GPT API and Bing AI and Google Bard to generate responses to user inputs. The bot responds to these commands: `!gpt`, `!chat` and `!bing` and `!pic` and `!bard` and `!talk` and `!goon` and `!new` and  `!help` depending on the first word of the prompt.
-
-## Feature
-
-1. Support Openai ChatGPT and Bing AI and Google Bard
-2. Support Bing Image Creator
-3. [pandora](https://github.com/pengzhile/pandora) with Session isolation support
-
-## Installation and Setup
-
-See https://github.com/hibobmaster/mattermost_bot/wiki
-
-Edit `config.json` or `.env` with proper values
-
-```sh
-docker compose up -d
-```
-
-## Commands
-
-- `!help` help message
-- `!gpt + [prompt]` generate a one time response from chatGPT
-- `!chat + [prompt]` chat using official chatGPT api with context conversation
-- `!bing + [prompt]` chat with Bing AI with context conversation
-- `!bard + [prompt]` chat with Google's Bard
-- `!pic + [prompt]` generate an image from Bing Image Creator
-
-The following commands need pandora http api: https://github.com/pengzhile/pandora/blob/master/doc/wiki_en.md#http-restful-api
-- `!talk + [prompt]` chat using chatGPT web with context conversation
-- `!goon` ask chatGPT to complete the missing part from previous conversation
-- `!new` start a new converstaion 
-
-## Demo
-
-![demo1](https://i.imgur.com/XRAQB4B.jpg)
-![demo2](https://i.imgur.com/if72kyH.jpg)
-![demo3](https://i.imgur.com/GHczfkv.jpg)
-
-## Thanks
-<a href="https://jb.gg/OpenSourceSupport" target="_blank">
-<img src="https://resources.jetbrains.com/storage/products/company/brand/logos/jb_beam.png" alt="JetBrains Logo (Main) logo." width="200" height="200">
-</a>
+## Introduction
+
+This is a simple Mattermost Bot that uses OpenAI's GPT API to generate responses to user inputs. The bot responds to these commands: `!gpt`, `!chat` and `!talk` and `!goon` and `!new` and  `!help` depending on the first word of the prompt.
+
+## Feature
+
+1. Support Openai ChatGPT
+3. ChatGPT web ([pandora](https://github.com/pengzhile/pandora))
+## Installation and Setup
+
+See https://github.com/hibobmaster/mattermost_bot/wiki
+
+Edit `config.json` or `.env` with proper values
+
+```sh
+docker compose up -d
+```
+
+## Commands
+
+- `!help` help message
+- `!gpt + [prompt]` generate a one time response from chatGPT
+- `!chat + [prompt]` chat using official chatGPT api with context conversation
+- `!bing + [prompt]` chat with Bing AI with context conversation
+- `!bard + [prompt]` chat with Google's Bard
+- `!pic + [prompt]` generate an image from Bing Image Creator
+
+The following commands need pandora http api: https://github.com/pengzhile/pandora/blob/master/doc/wiki_en.md#http-restful-api
+- `!talk + [prompt]` chat using chatGPT web with context conversation
+- `!goon` ask chatGPT to complete the missing part from previous conversation
+- `!new` start a new converstaion 
+
+## Demo
+Remove support for Bing AI, Google Bard due to technical problems.
+![demo1](https://i.imgur.com/XRAQB4B.jpg)
+![demo2](https://i.imgur.com/if72kyH.jpg)
+![demo3](https://i.imgur.com/GHczfkv.jpg)
+
+## Thanks
+<a href="https://jb.gg/OpenSourceSupport" target="_blank">
+<img src="https://resources.jetbrains.com/storage/products/company/brand/logos/jb_beam.png" alt="JetBrains Logo (Main) logo." width="200" height="200">
+</a>
diff --git a/bard.py b/bard.py
deleted file mode 100644
index ef83325..0000000
--- a/bard.py
+++ /dev/null
@@ -1,104 +0,0 @@
-"""
-Code derived from: https://github.com/acheong08/Bard/blob/main/src/Bard.py
-"""
-
-import random
-import string
-import re
-import json
-import requests
-
-
-class Bardbot:
-    """
-    A class to interact with Google Bard.
-    Parameters
-        session_id: str
-            The __Secure-1PSID cookie.
-    """
-
-    __slots__ = [
-        "headers",
-        "_reqid",
-        "SNlM0e",
-        "conversation_id",
-        "response_id",
-        "choice_id",
-        "session",
-    ]
-
-    def __init__(self, session_id):
-        headers = {
-            "Host": "bard.google.com",
-            "X-Same-Domain": "1",
-            "User-Agent": "Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/91.0.4472.114 Safari/537.36",
-            "Content-Type": "application/x-www-form-urlencoded;charset=UTF-8",
-            "Origin": "https://bard.google.com",
-            "Referer": "https://bard.google.com/",
-        }
-        self._reqid = int("".join(random.choices(string.digits, k=4)))
-        self.conversation_id = ""
-        self.response_id = ""
-        self.choice_id = ""
-        self.session = requests.Session()
-        self.session.headers = headers
-        self.session.cookies.set("__Secure-1PSID", session_id)
-        self.SNlM0e = self.__get_snlm0e()
-
-    def __get_snlm0e(self):
-        resp = self.session.get(url="https://bard.google.com/", timeout=10)
-        # Find "SNlM0e":"<ID>"
-        if resp.status_code != 200:
-            raise Exception("Could not get Google Bard")
-        SNlM0e = re.search(r"SNlM0e\":\"(.*?)\"", resp.text).group(1)
-        return SNlM0e
-
-    def ask(self, message: str) -> dict:
-        """
-        Send a message to Google Bard and return the response.
-        :param message: The message to send to Google Bard.
-        :return: A dict containing the response from Google Bard.
-        """
-        # url params
-        params = {
-            "bl": "boq_assistant-bard-web-server_20230326.21_p0",
-            "_reqid": str(self._reqid),
-            "rt": "c",
-        }
-
-        # message arr -> data["f.req"]. Message is double json stringified
-        message_struct = [
-            [message],
-            None,
-            [self.conversation_id, self.response_id, self.choice_id],
-        ]
-        data = {
-            "f.req": json.dumps([None, json.dumps(message_struct)]),
-            "at": self.SNlM0e,
-        }
-
-        # do the request!
-        resp = self.session.post(
-            "https://bard.google.com/_/BardChatUi/data/assistant.lamda.BardFrontendService/StreamGenerate",
-            params=params,
-            data=data,
-            timeout=120,
-        )
-
-        chat_data = json.loads(resp.content.splitlines()[3])[0][2]
-        if not chat_data:
-            return {"content": f"Google Bard encountered an error: {resp.content}."}
-        json_chat_data = json.loads(chat_data)
-        results = {
-            "content": json_chat_data[0][0],
-            "conversation_id": json_chat_data[1][0],
-            "response_id": json_chat_data[1][1],
-            "factualityQueries": json_chat_data[3],
-            "textQuery": json_chat_data[2][0] if json_chat_data[2] is not None else "",
-            "choices": [{"id": i[0], "content": i[1]} for i in json_chat_data[4]],
-        }
-        self.conversation_id = results["conversation_id"]
-        self.response_id = results["response_id"]
-        self.choice_id = results["choices"][0]["id"]
-        self._reqid += 100000
-        return results
diff --git a/bing.py b/bing.py
deleted file mode 100644
index 2573759..0000000
--- a/bing.py
+++ /dev/null
@@ -1,64 +0,0 @@
-import aiohttp
-import json
-import asyncio
-from log import getlogger
-
-# api_endpoint = "http://localhost:3000/conversation"
-from log import getlogger
-
-logger = getlogger()
-
-
-class BingBot:
-    def __init__(
-        self,
-        session: aiohttp.ClientSession,
-        bing_api_endpoint: str,
-        jailbreakEnabled: bool = True,
-    ):
-        self.data = {
-            "clientOptions.clientToUse": "bing",
-        }
-        self.bing_api_endpoint = bing_api_endpoint
-
-        self.session = session
-
-        self.jailbreakEnabled = jailbreakEnabled
-
-        if self.jailbreakEnabled:
-            self.data["jailbreakConversationId"] = True
-
-    async def ask_bing(self, prompt) -> str:
-        self.data["message"] = prompt
-        max_try = 2
-        while max_try > 0:
-            try:
-                resp = await self.session.post(
-                    url=self.bing_api_endpoint, json=self.data, timeout=120
-                )
-                status_code = resp.status
-                body = await resp.read()
-                if not status_code == 200:
-                    # print failed reason
-                    logger.warning(str(resp.reason))
-                    max_try = max_try - 1
-                    await asyncio.sleep(2)
-                    continue
-                json_body = json.loads(body)
-                if self.jailbreakEnabled:
-                    self.data["jailbreakConversationId"] = json_body[
-                        "jailbreakConversationId"
-                    ]
-                    self.data["parentMessageId"] = json_body["messageId"]
-                else:
-                    self.data["conversationSignature"] = json_body[
-                        "conversationSignature"
-                    ]
-                    self.data["conversationId"] = json_body["conversationId"]
-                    self.data["clientId"] = json_body["clientId"]
-                    self.data["invocationId"] = json_body["invocationId"]
-                return json_body["details"]["adaptiveCards"][0]["body"][0]["text"]
-            except Exception as e:
-                logger.error("Error Exception", exc_info=True)
-
-        return "Error, please retry"
diff --git a/compose.yaml b/compose.yaml
index 8a5d9a7..d3500df 100644
--- a/compose.yaml
+++ b/compose.yaml
@@ -11,14 +11,6 @@ services:
     networks:
       - mattermost_network
 
-  # api:
-  #   image: hibobmaster/node-chatgpt-api:latest
-  #   container_name: node-chatgpt-api
-  #   volumes:
-  #     - ./settings.js:/var/chatgpt-api/settings.js
-  #   networks:
-  #     - mattermost_network
-
   # pandora:
   #   image: pengzhile/pandora
   #   container_name: pandora
diff --git a/config.json.example b/config.json.example
index a391781..b6258bc 100644
--- a/config.json.example
+++ b/config.json.example
@@ -1,11 +1,10 @@
-{
-    "server_url": "xxxx.xxxx.xxxxx",
-    "access_token": "xxxxxxxxxxxxxxxxxxxxxx",
-    "username": "@chatgpt",
-    "openai_api_key": "sk-xxxxxxxxxxxxxxxxxxx",
-    "bing_api_endpoint": "http://api:3000/conversation",
-    "bard_token": "xxxxxxxxxxxxxxxxxxxxxxxxxxxxx.",
-    "bing_auth_cookie": "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
-    "pandora_api_endpoint": "http://127.0.0.1:8008",
-    "pandora_api_model": "text-davinci-002-render-sha-mobile"
+{
+    "server_url": "xxxx.xxxx.xxxxx",
+    "access_token": "xxxxxxxxxxxxxxxxxxxxxx",
+    "username": "@chatgpt",
+    "openai_api_key": "sk-xxxxxxxxxxxxxxxxxxx",
+    "gpt_engine": "gpt-3.5-turbo",
+    "bing_auth_cookie": "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
+    "pandora_api_endpoint": "http://127.0.0.1:8008",
+    "pandora_api_model": "text-davinci-002-render-sha-mobile"
 }
\ No newline at end of file
diff --git a/requirements.txt b/requirements.txt
index af9086a..42932c4 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -1,27 +1,4 @@
-aiohttp==3.8.4
-aiosignal==1.3.1
-anyio==3.6.2
-async-timeout==4.0.2
-attrs==23.1.0
-certifi==2022.12.7
-charset-normalizer==3.1.0
-click==8.1.3
-colorama==0.4.6
-frozenlist==1.3.3
-h11==0.14.0
-httpcore==0.17.0
-httpx==0.24.0
-idna==3.4
-mattermostdriver @ git+https://github.com/hibobmaster/python-mattermost-driver
-multidict==6.0.4
-mypy-extensions==1.0.0
-packaging==23.1
-pathspec==0.11.1
-platformdirs==3.2.0
-regex==2023.3.23
-requests==2.28.2
-sniffio==1.3.0
-tiktoken==0.3.3
-urllib3==1.26.15
-websockets==11.0.1
-yarl==1.8.2
+aiohttp
+httpx
+mattermostdriver @ git+https://github.com/hibobmaster/python-mattermost-driver
+revChatGPT>=6.8.6
\ No newline at end of file
diff --git a/BingImageGen.py b/src/BingImageGen.py
similarity index 97%
rename from BingImageGen.py
rename to src/BingImageGen.py
index 878a1a6..979346c 100644
--- a/BingImageGen.py
+++ b/src/BingImageGen.py
@@ -1,165 +1,165 @@
-"""
-Code derived from:
-https://github.com/acheong08/EdgeGPT/blob/f940cecd24a4818015a8b42a2443dd97c3c2a8f4/src/ImageGen.py
-"""
-from log import getlogger
-from uuid import uuid4
-import os
-import contextlib
-import aiohttp
-import asyncio
-import random
-import requests
-import regex
-
-logger = getlogger()
-
-BING_URL = "https://www.bing.com"
-# Generate random IP between range 13.104.0.0/14
-FORWARDED_IP = (
-    f"13.{random.randint(104, 107)}.{random.randint(0, 255)}.{random.randint(0, 255)}"
-)
-HEADERS = {
-    "accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7",
-    "accept-language": "en-US,en;q=0.9",
-    "cache-control": "max-age=0",
-    "content-type": "application/x-www-form-urlencoded",
-    "referrer": "https://www.bing.com/images/create/",
-    "origin": "https://www.bing.com",
-    "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/110.0.0.0 Safari/537.36 Edg/110.0.1587.63",
-    "x-forwarded-for": FORWARDED_IP,
-}
-
-
-class ImageGenAsync:
-    """
-    Image generation by Microsoft Bing
-    Parameters:
-        auth_cookie: str
-    """
-
-    def __init__(self, auth_cookie: str, quiet: bool = True) -> None:
-        self.session = aiohttp.ClientSession(
-            headers=HEADERS,
-            cookies={"_U": auth_cookie},
-        )
-        self.quiet = quiet
-
-    async def __aenter__(self):
-        return self
-
-    async def __aexit__(self, *excinfo) -> None:
-        await self.session.close()
-
-    def __del__(self):
-        try:
-            loop = asyncio.get_running_loop()
-        except RuntimeError:
-            loop = asyncio.new_event_loop()
-            asyncio.set_event_loop(loop)
-        loop.run_until_complete(self._close())
-
-    async def _close(self):
-        await self.session.close()
-
-    async def get_images(self, prompt: str) -> list:
-        """
-        Fetches image links from Bing
-        Parameters:
-            prompt: str
-        """
-        if not self.quiet:
-            print("Sending request...")
-        url_encoded_prompt = requests.utils.quote(prompt)
-        # https://www.bing.com/images/create?q=<PROMPT>&rt=3&FORM=GENCRE
-        url = f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=4&FORM=GENCRE"
-        async with self.session.post(url, allow_redirects=False) as response:
-            content = await response.text()
-            if "this prompt has been blocked" in content.lower():
-                raise Exception(
-                    "Your prompt has been blocked by Bing. Try to change any bad words and try again.",
-                )
-            if response.status != 302:
-                # if rt4 fails, try rt3
-                url = (
-                    f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=3&FORM=GENCRE"
-                )
-                async with self.session.post(
-                    url,
-                    allow_redirects=False,
-                    timeout=200,
-                ) as response3:
-                    if response3.status != 302:
-                        print(f"ERROR: {response3.text}")
-                        raise Exception("Redirect failed")
-                    response = response3
-        # Get redirect URL
-        redirect_url = response.headers["Location"].replace("&nfy=1", "")
-        request_id = redirect_url.split("id=")[-1]
-        await self.session.get(f"{BING_URL}{redirect_url}")
-        # https://www.bing.com/images/create/async/results/{ID}?q={PROMPT}
-        polling_url = f"{BING_URL}/images/create/async/results/{request_id}?q={url_encoded_prompt}"
-        # Poll for results
-        if not self.quiet:
-            print("Waiting for results...")
-        while True:
-            if not self.quiet:
-                print(".", end="", flush=True)
-            # By default, timeout is 300s, change as needed
-            response = await self.session.get(polling_url)
-            if response.status != 200:
-                raise Exception("Could not get results")
-            content = await response.text()
-            if content and content.find("errorMessage") == -1:
-                break
-
-            await asyncio.sleep(1)
-            continue
-        # Use regex to search for src=""
-        image_links = regex.findall(r'src="([^"]+)"', content)
-        # Remove size limit
-        normal_image_links = [link.split("?w=")[0] for link in image_links]
-        # Remove duplicates
-        normal_image_links = list(set(normal_image_links))
-
-        # Bad images
-        bad_images = [
-            "https://r.bing.com/rp/in-2zU3AJUdkgFe7ZKv19yPBHVs.png",
-            "https://r.bing.com/rp/TX9QuO3WzcCJz1uaaSwQAz39Kb0.jpg",
-        ]
-        for im in normal_image_links:
-            if im in bad_images:
-                raise Exception("Bad images")
-        # No images
-        if not normal_image_links:
-            raise Exception("No images")
-        return normal_image_links
-
-    async def save_images(self, links: list, output_dir: str) -> str:
-        """
-        Saves images to output directory
-        """
-        if not self.quiet:
-            print("\nDownloading images...")
-        with contextlib.suppress(FileExistsError):
-            os.mkdir(output_dir)
-
-        # image name
-        image_name = str(uuid4())
-        # we just need one image for better display in chat room
-        if links:
-            link = links.pop()
-
-        image_path = os.path.join(output_dir, f"{image_name}.jpeg")
-        try:
-            async with self.session.get(link, raise_for_status=True) as response:
-                # save response to file
-                with open(image_path, "wb") as output_file:
-                    async for chunk in response.content.iter_chunked(8192):
-                        output_file.write(chunk)
-            return f"{output_dir}/{image_name}.jpeg"
-
-        except aiohttp.client_exceptions.InvalidURL as url_exception:
-            raise Exception(
-                "Inappropriate contents found in the generated images. Please try again or try another prompt.",
-            ) from url_exception
+"""
+Code derived from:
+https://github.com/acheong08/EdgeGPT/blob/f940cecd24a4818015a8b42a2443dd97c3c2a8f4/src/ImageGen.py
+"""
+from log import getlogger
+from uuid import uuid4
+import os
+import contextlib
+import aiohttp
+import asyncio
+import random
+import requests
+import regex
+
+logger = getlogger()
+
+BING_URL = "https://www.bing.com"
+# Generate random IP between range 13.104.0.0/14
+FORWARDED_IP = (
+    f"13.{random.randint(104, 107)}.{random.randint(0, 255)}.{random.randint(0, 255)}"
+)
+HEADERS = {
+    "accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7",
+    "accept-language": "en-US,en;q=0.9",
+    "cache-control": "max-age=0",
+    "content-type": "application/x-www-form-urlencoded",
+    "referrer": "https://www.bing.com/images/create/",
+    "origin": "https://www.bing.com",
+    "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/110.0.0.0 Safari/537.36 Edg/110.0.1587.63",
+    "x-forwarded-for": FORWARDED_IP,
+}
+
+
+class ImageGenAsync:
+    """
+    Image generation by Microsoft Bing
+    Parameters:
+        auth_cookie: str
+    """
+
+    def __init__(self, auth_cookie: str, quiet: bool = True) -> None:
+        self.session = aiohttp.ClientSession(
+            headers=HEADERS,
+            cookies={"_U": auth_cookie},
+        )
+        self.quiet = quiet
+
+    async def __aenter__(self):
+        return self
+
+    async def __aexit__(self, *excinfo) -> None:
+        await self.session.close()
+
+    def __del__(self):
+        try:
+            loop = asyncio.get_running_loop()
+        except RuntimeError:
+            loop = asyncio.new_event_loop()
+            asyncio.set_event_loop(loop)
+        loop.run_until_complete(self._close())
+
+    async def _close(self):
+        await self.session.close()
+
+    async def get_images(self, prompt: str) -> list:
+        """
+        Fetches image links from Bing
+        Parameters:
+            prompt: str
+        """
+        if not self.quiet:
+            print("Sending request...")
+        url_encoded_prompt = requests.utils.quote(prompt)
+        # https://www.bing.com/images/create?q=<PROMPT>&rt=3&FORM=GENCRE
+        url = f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=4&FORM=GENCRE"
+        async with self.session.post(url, allow_redirects=False) as response:
+            content = await response.text()
+            if "this prompt has been blocked" in content.lower():
+                raise Exception(
+                    "Your prompt has been blocked by Bing. Try to change any bad words and try again.",
+                )
+            if response.status != 302:
+                # if rt4 fails, try rt3
+                url = (
+                    f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=3&FORM=GENCRE"
+                )
+                async with self.session.post(
+                    url,
+                    allow_redirects=False,
+                    timeout=200,
+                ) as response3:
+                    if response3.status != 302:
+                        print(f"ERROR: {response3.text}")
+                        raise Exception("Redirect failed")
+                    response = response3
+        # Get redirect URL
+        redirect_url = response.headers["Location"].replace("&nfy=1", "")
+        request_id = redirect_url.split("id=")[-1]
+        await self.session.get(f"{BING_URL}{redirect_url}")
+        # https://www.bing.com/images/create/async/results/{ID}?q={PROMPT}
+        polling_url = f"{BING_URL}/images/create/async/results/{request_id}?q={url_encoded_prompt}"
+        # Poll for results
+        if not self.quiet:
+            print("Waiting for results...")
+        while True:
+            if not self.quiet:
+                print(".", end="", flush=True)
+            # By default, timeout is 300s, change as needed
+            response = await self.session.get(polling_url)
+            if response.status != 200:
+                raise Exception("Could not get results")
+            content = await response.text()
+            if content and content.find("errorMessage") == -1:
+                break
+
+            await asyncio.sleep(1)
+            continue
+        # Use regex to search for src=""
+        image_links = regex.findall(r'src="([^"]+)"', content)
+        # Remove size limit
+        normal_image_links = [link.split("?w=")[0] for link in image_links]
+        # Remove duplicates
+        normal_image_links = list(set(normal_image_links))
+
+        # Bad images
+        bad_images = [
+            "https://r.bing.com/rp/in-2zU3AJUdkgFe7ZKv19yPBHVs.png",
+            "https://r.bing.com/rp/TX9QuO3WzcCJz1uaaSwQAz39Kb0.jpg",
+        ]
+        for im in normal_image_links:
+            if im in bad_images:
+                raise Exception("Bad images")
+        # No images
+        if not normal_image_links:
+            raise Exception("No images")
+        return normal_image_links
+
+    async def save_images(self, links: list, output_dir: str) -> str:
+        """
+        Saves images to output directory
+        """
+        if not self.quiet:
+            print("\nDownloading images...")
+        with contextlib.suppress(FileExistsError):
+            os.mkdir(output_dir)
+
+        # image name
+        image_name = str(uuid4())
+        # we just need one image for better display in chat room
+        if links:
+            link = links.pop()
+
+        image_path = os.path.join(output_dir, f"{image_name}.jpeg")
+        try:
+            async with self.session.get(link, raise_for_status=True) as response:
+                # save response to file
+                with open(image_path, "wb") as output_file:
+                    async for chunk in response.content.iter_chunked(8192):
+                        output_file.write(chunk)
+            return f"{output_dir}/{image_name}.jpeg"
+
+        except aiohttp.client_exceptions.InvalidURL as url_exception:
+            raise Exception(
+                "Inappropriate contents found in the generated images. Please try again or try another prompt.",
+            ) from url_exception
diff --git a/askgpt.py b/src/askgpt.py
similarity index 88%
rename from askgpt.py
rename to src/askgpt.py
index 6b9a92d..0a9f0fc 100644
--- a/askgpt.py
+++ b/src/askgpt.py
@@ -1,46 +1,46 @@
-import aiohttp
-import asyncio
-import json
-
-from log import getlogger
-
-logger = getlogger()
-
-
-class askGPT:
-    def __init__(
-        self, session: aiohttp.ClientSession, api_endpoint: str, headers: str
-    ) -> None:
-        self.session = session
-        self.api_endpoint = api_endpoint
-        self.headers = headers
-
-    async def oneTimeAsk(self, prompt: str) -> str:
-        jsons = {
-            "model": "gpt-3.5-turbo",
-            "messages": [
-                {
-                    "role": "user",
-                    "content": prompt,
-                },
-            ],
-        }
-        max_try = 2
-        while max_try > 0:
-            try:
-                async with self.session.post(
-                    url=self.api_endpoint, json=jsons, headers=self.headers, timeout=120
-                ) as response:
-                    status_code = response.status
-                    if not status_code == 200:
-                        # print failed reason
-                        logger.warning(str(response.reason))
-                        max_try = max_try - 1
-                        # wait 2s
-                        await asyncio.sleep(2)
-                        continue
-
-                    resp = await response.read()
-                    return json.loads(resp)["choices"][0]["message"]["content"]
-            except Exception as e:
-                raise Exception(e)
+import aiohttp
+import asyncio
+import json
+
+from log import getlogger
+
+logger = getlogger()
+
+
+class askGPT:
+    def __init__(
+        self, session: aiohttp.ClientSession, headers: str
+    ) -> None:
+        self.session = session
+        self.api_endpoint = "https://api.openai.com/v1/chat/completions"
+        self.headers = headers
+
+    async def oneTimeAsk(self, prompt: str) -> str:
+        jsons = {
+            "model": "gpt-3.5-turbo",
+            "messages": [
+                {
+                    "role": "user",
+                    "content": prompt,
+                },
+            ],
+        }
+        max_try = 2
+        while max_try > 0:
+            try:
+                async with self.session.post(
+                    url=self.api_endpoint, json=jsons, headers=self.headers, timeout=120
+                ) as response:
+                    status_code = response.status
+                    if not status_code == 200:
+                        # print failed reason
+                        logger.warning(str(response.reason))
+                        max_try = max_try - 1
+                        # wait 2s
+                        await asyncio.sleep(2)
+                        continue
+
+                    resp = await response.read()
+                    return json.loads(resp)["choices"][0]["message"]["content"]
+            except Exception as e:
+                raise Exception(e)
diff --git a/bot.py b/src/bot.py
similarity index 64%
rename from bot.py
rename to src/bot.py
index ca2cfa4..5b385f2 100644
--- a/bot.py
+++ b/src/bot.py
@@ -1,426 +1,410 @@
-from mattermostdriver import Driver
-from typing import Optional
-import json
-import asyncio
-import re
-import os
-import aiohttp
-from askgpt import askGPT
-from v3 import Chatbot
-from bing import BingBot
-from bard import Bardbot
-from BingImageGen import ImageGenAsync
-from log import getlogger
-from pandora import Pandora
-import uuid
-
-logger = getlogger()
-
-
-class Bot:
-    def __init__(
-        self,
-        server_url: str,
-        username: str,
-        access_token: Optional[str] = None,
-        login_id: Optional[str] = None,
-        password: Optional[str] = None,
-        openai_api_key: Optional[str] = None,
-        openai_api_endpoint: Optional[str] = None,
-        bing_api_endpoint: Optional[str] = None,
-        pandora_api_endpoint: Optional[str] = None,
-        pandora_api_model: Optional[str] = None,
-        bard_token: Optional[str] = None,
-        bing_auth_cookie: Optional[str] = None,
-        port: int = 443,
-        timeout: int = 30,
-    ) -> None:
-        if server_url is None:
-            raise ValueError("server url must be provided")
-
-        if port is None:
-            self.port = 443
-
-        if timeout is None:
-            self.timeout = 30
-
-        # login relative info
-        if access_token is None and password is None:
-            raise ValueError("Either token or password must be provided")
-
-        if access_token is not None:
-            self.driver = Driver(
-                {
-                    "token": access_token,
-                    "url": server_url,
-                    "port": self.port,
-                    "request_timeout": self.timeout,
-                }
-            )
-        else:
-            self.driver = Driver(
-                {
-                    "login_id": login_id,
-                    "password": password,
-                    "url": server_url,
-                    "port": self.port,
-                    "request_timeout": self.timeout,
-                }
-            )
-
-        # @chatgpt
-        if username is None:
-            raise ValueError("username must be provided")
-        else:
-            self.username = username
-
-        # openai_api_endpoint
-        if openai_api_endpoint is None:
-            self.openai_api_endpoint = "https://api.openai.com/v1/chat/completions"
-        else:
-            self.openai_api_endpoint = openai_api_endpoint
-
-        # aiohttp session
-        self.session = aiohttp.ClientSession()
-
-        self.openai_api_key = openai_api_key
-        # initialize chatGPT class
-        if self.openai_api_key is not None:
-            # request header for !gpt command
-            self.headers = {
-                "Content-Type": "application/json",
-                "Authorization": f"Bearer {self.openai_api_key}",
-            }
-
-            self.askgpt = askGPT(
-                self.session,
-                self.openai_api_endpoint,
-                self.headers,
-            )
-
-            self.chatbot = Chatbot(api_key=self.openai_api_key)
-        else:
-            logger.warning(
-                "openai_api_key is not provided, !gpt and !chat command will not work"
-            )
-
-        self.bing_api_endpoint = bing_api_endpoint
-        # initialize bingbot
-        if self.bing_api_endpoint is not None:
-            self.bingbot = BingBot(
-                session=self.session,
-                bing_api_endpoint=self.bing_api_endpoint,
-            )
-        else:
-            logger.warning(
-                "bing_api_endpoint is not provided, !bing command will not work"
-            )
-
-        self.pandora_api_endpoint = pandora_api_endpoint
-        # initialize pandora
-        if pandora_api_endpoint is not None:
-            self.pandora = Pandora(
-                api_endpoint=pandora_api_endpoint,
-                clientSession=self.session
-            )
-            if pandora_api_model is None:
-                self.pandora_api_model = "text-davinci-002-render-sha-mobile"
-            else:
-                self.pandora_api_model = pandora_api_model
-        self.pandora_data = {}
-        
-        self.bard_token = bard_token
-        # initialize bard
-        if self.bard_token is not None:
-            self.bardbot = Bardbot(session_id=self.bard_token)
-        else:
-            logger.warning("bard_token is not provided, !bard command will not work")
-
-        self.bing_auth_cookie = bing_auth_cookie
-        # initialize image generator
-        if self.bing_auth_cookie is not None:
-            self.imagegen = ImageGenAsync(auth_cookie=self.bing_auth_cookie)
-        else:
-            logger.warning(
-                "bing_auth_cookie is not provided, !pic command will not work"
-            )
-
-        # regular expression to match keyword 
-        self.gpt_prog = re.compile(r"^\s*!gpt\s*(.+)$")
-        self.chat_prog = re.compile(r"^\s*!chat\s*(.+)$")
-        self.bing_prog = re.compile(r"^\s*!bing\s*(.+)$")
-        self.bard_prog = re.compile(r"^\s*!bard\s*(.+)$")
-        self.pic_prog = re.compile(r"^\s*!pic\s*(.+)$")
-        self.help_prog = re.compile(r"^\s*!help\s*.*$")
-        self.talk_prog = re.compile(r"^\s*!talk\s*(.+)$")
-        self.goon_prog = re.compile(r"^\s*!goon\s*.*$")
-        self.new_prog = re.compile(r"^\s*!new\s*.*$")
-
-    # close session
-    def __del__(self) -> None:
-        self.driver.disconnect()
-
-    def login(self) -> None:
-        self.driver.login()
-
-    def pandora_init(self, user_id: str) -> None:
-        self.pandora_data[user_id] = {
-            "conversation_id": None,
-            "parent_message_id": str(uuid.uuid4()),
-            "first_time": True
-        }
-        
-    async def run(self) -> None:
-        await self.driver.init_websocket(self.websocket_handler)
-
-    # websocket handler
-    async def websocket_handler(self, message) -> None:
-        print(message)
-        response = json.loads(message)
-        if "event" in response:
-            event_type = response["event"]
-            if event_type == "posted":
-                raw_data = response["data"]["post"]
-                raw_data_dict = json.loads(raw_data)
-                user_id = raw_data_dict["user_id"]
-                channel_id = raw_data_dict["channel_id"]
-                sender_name = response["data"]["sender_name"]
-                raw_message = raw_data_dict["message"]
-                
-                if user_id not in self.pandora_data:
-                    self.pandora_init(user_id)
-
-                try:
-                    asyncio.create_task(
-                        self.message_callback(
-                            raw_message, channel_id, user_id, sender_name
-                        )
-                    )
-                except Exception as e:
-                    await asyncio.to_thread(self.send_message, channel_id, f"{e}")
-
-    # message callback
-    async def message_callback(
-        self, raw_message: str, channel_id: str, user_id: str, sender_name: str
-    ) -> None:
-        # prevent command trigger loop
-        if sender_name != self.username:
-            message = raw_message
-
-            if self.openai_api_key is not None:
-                # !gpt command trigger handler
-                if self.gpt_prog.match(message):
-                    prompt = self.gpt_prog.match(message).group(1)
-                    try:
-                        response = await self.gpt(prompt)
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, f"{response}"
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-                # !chat command trigger handler
-                elif self.chat_prog.match(message):
-                    prompt = self.chat_prog.match(message).group(1)
-                    try:
-                        response = await self.chat(prompt)
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, f"{response}"
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-            if self.bing_api_endpoint is not None:
-                # !bing command trigger handler
-                if self.bing_prog.match(message):
-                    prompt = self.bing_prog.match(message).group(1)
-                    try:
-                        response = await self.bingbot.ask_bing(prompt)
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, f"{response}"
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-            if self.pandora_api_endpoint is not None:
-                # !talk command trigger handler
-                if self.talk_prog.match(message):
-                    prompt = self.talk_prog.match(message).group(1)
-                    try:
-                        if self.pandora_data[user_id]["conversation_id"] is not None:
-                            data = {
-                                "prompt": prompt,
-                                "model": self.pandora_api_model,
-                                "parent_message_id": self.pandora_data[user_id]["parent_message_id"],
-                                "conversation_id": self.pandora_data[user_id]["conversation_id"],
-                                "stream": False,
-                            }
-                        else:
-                            data = {
-                                "prompt": prompt,
-                                "model": self.pandora_api_model,
-                                "parent_message_id": self.pandora_data[user_id]["parent_message_id"],
-                                "stream": False,
-                            }
-                        response = await self.pandora.talk(data)
-                        self.pandora_data[user_id]["conversation_id"] = response['conversation_id']
-                        self.pandora_data[user_id]["parent_message_id"] = response['message']['id']
-                        content = response['message']['content']['parts'][0]
-                        if self.pandora_data[user_id]["first_time"]:
-                            self.pandora_data[user_id]["first_time"] = False
-                            data = {
-                                "model": self.pandora_api_model,
-                                "message_id": self.pandora_data[user_id]["parent_message_id"],
-                            }
-                            await self.pandora.gen_title(data, self.pandora_data[user_id]["conversation_id"])
-                        
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, f"{content}"
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-                    
-                # !goon command trigger handler
-                if self.goon_prog.match(message) and self.pandora_data[user_id]["conversation_id"] is not None:
-                    try:
-                        data = {
-                            "model": self.pandora_api_model,
-                            "parent_message_id": self.pandora_data[user_id]["parent_message_id"],
-                            "conversation_id": self.pandora_data[user_id]["conversation_id"],
-                            "stream": False,
-                        }
-                        response = await self.pandora.goon(data)
-                        self.pandora_data[user_id]["conversation_id"] = response['conversation_id']
-                        self.pandora_data[user_id]["parent_message_id"] = response['message']['id']
-                        content = response['message']['content']['parts'][0]
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, f"{content}"
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-                # !new command trigger handler
-                if self.new_prog.match(message):
-                    self.pandora_init(user_id)
-                    try:
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, "New conversation created, please use !talk to start chatting!"
-                        )
-                    except Exception:
-                        pass
-
-            if self.bard_token is not None:
-                # !bard command trigger handler
-                if self.bard_prog.match(message):
-                    prompt = self.bard_prog.match(message).group(1)
-                    try:
-                        # response is dict object
-                        response = await self.bard(prompt)
-                        content = str(response["content"]).strip()
-                        await asyncio.to_thread(
-                            self.send_message, channel_id, f"{content}"
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-            if self.bing_auth_cookie is not None:
-                # !pic command trigger handler
-                if self.pic_prog.match(message):
-                    prompt = self.pic_prog.match(message).group(1)
-                    # generate image
-                    try:
-                        links = await self.imagegen.get_images(prompt)
-                        image_path = await self.imagegen.save_images(links, "images")
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-                    # send image
-                    try:
-                        await asyncio.to_thread(
-                            self.send_file, channel_id, prompt, image_path
-                        )
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-            # !help command trigger handler
-            if self.help_prog.match(message):
-                try:
-                    await asyncio.to_thread(self.send_message, channel_id, self.help())
-                except Exception as e:
-                    logger.error(e, exc_info=True)
-
-    # send message to room
-    def send_message(self, channel_id: str, message: str) -> None:
-        self.driver.posts.create_post(
-            options={
-                "channel_id": channel_id,
-                "message": message
-            }
-        )
-
-    # send file to room
-    def send_file(self, channel_id: str, message: str, filepath: str) -> None:
-        filename = os.path.split(filepath)[-1]
-        try:
-            file_id = self.driver.files.upload_file(
-                channel_id=channel_id,
-                files={
-                    "files": (filename, open(filepath, "rb")),
-                },
-            )["file_infos"][0]["id"]
-        except Exception as e:
-            logger.error(e, exc_info=True)
-            raise Exception(e)
-
-        try:
-            self.driver.posts.create_post(
-                options={
-                    "channel_id": channel_id,
-                    "message": message,
-                    "file_ids": [file_id],
-                }
-            )
-            # remove image after posting
-            os.remove(filepath)
-        except Exception as e:
-            logger.error(e, exc_info=True)
-            raise Exception(e)
-
-    # !gpt command function
-    async def gpt(self, prompt: str) -> str:
-        return await self.askgpt.oneTimeAsk(prompt)
-
-    # !chat command function
-    async def chat(self, prompt: str) -> str:
-        return await self.chatbot.ask_async(prompt)
-
-    # !bing command function
-    async def bing(self, prompt: str) -> str:
-        return await self.bingbot.ask_bing(prompt)
-
-    # !bard command function
-    async def bard(self, prompt: str) -> str:
-        return await asyncio.to_thread(self.bardbot.ask, prompt)
-
-    # !help command function
-    def help(self) -> str:
-        help_info = (
-            "!gpt [content], generate response without context conversation\n"
-            + "!chat [content], chat with context conversation\n"
-            + "!bing [content], chat with context conversation powered by Bing AI\n"
-            + "!bard [content], chat with Google's Bard\n"
-            + "!pic [prompt], Image generation by Microsoft Bing\n"
-            + "!talk [content], talk using chatgpt web\n"
-            + "!goon, continue the incomplete conversation\n"
-            + "!new, start a new conversation\n"
-            + "!help, help message"
-        )
-        return help_info
+from mattermostdriver import AsyncDriver
+from typing import Optional
+import json
+import asyncio
+import re
+import os
+import aiohttp
+from askgpt import askGPT
+from revChatGPT.V3 import Chatbot as GPTChatBot
+from BingImageGen import ImageGenAsync
+from log import getlogger
+from pandora import Pandora
+import uuid
+
+logger = getlogger()
+
+ENGINES = [
+    "gpt-3.5-turbo",
+    "gpt-3.5-turbo-16k",
+    "gpt-3.5-turbo-0301",
+    "gpt-3.5-turbo-0613",
+    "gpt-3.5-turbo-16k-0613",
+    "gpt-4",
+    "gpt-4-0314",
+    "gpt-4-32k",
+    "gpt-4-32k-0314",
+    "gpt-4-0613",
+    "gpt-4-32k-0613",
+]
+
+
+class Bot:
+    def __init__(
+        self,
+        server_url: str,
+        username: str,
+        access_token: Optional[str] = None,
+        login_id: Optional[str] = None,
+        password: Optional[str] = None,
+        openai_api_key: Optional[str] = None,
+        pandora_api_endpoint: Optional[str] = None,
+        pandora_api_model: Optional[str] = None,
+        bing_auth_cookie: Optional[str] = None,
+        port: int = 443,
+        scheme: str = "https",
+        timeout: int = 30,
+        gpt_engine: str = "gpt-3.5-turbo",
+    ) -> None:
+        if server_url is None:
+            raise ValueError("server url must be provided")
+
+        if port is None:
+            self.port = 443
+        else:
+            if port < 0 or port > 65535:
+                raise ValueError("port must be between 0 and 65535")
+            self.port = port
+
+        if scheme is None:
+            self.scheme = "https"
+        else:
+            if scheme.strip().lower() not in ["http", "https"]:
+                raise ValueError("scheme must be either http or https")
+            self.scheme = scheme
+
+        if timeout is None:
+            self.timeout = 30
+        else:
+            self.timeout = timeout
+
+        if gpt_engine is None:
+            self.gpt_engine = "gpt-3.5-turbo"
+        else:
+            if gpt_engine not in ENGINES:
+                raise ValueError("gpt_engine must be one of {}".format(ENGINES))
+            self.gpt_engine = gpt_engine
+
+        # login relative info
+        if access_token is None and password is None:
+            raise ValueError("Either token or password must be provided")
+
+        if access_token is not None:
+            self.driver = AsyncDriver(
+                {
+                    "token": access_token,
+                    "url": server_url,
+                    "port": self.port,
+                    "request_timeout": self.timeout,
+                    "scheme": self.scheme,
+                }
+            )
+        else:
+            self.driver = AsyncDriver(
+                {
+                    "login_id": login_id,
+                    "password": password,
+                    "url": server_url,
+                    "port": self.port,
+                    "request_timeout": self.timeout,
+                    "scheme": self.scheme,
+                }
+            )
+
+        # @chatgpt
+        if username is None:
+            raise ValueError("username must be provided")
+        else:
+            self.username = username
+
+        # aiohttp session
+        self.session = aiohttp.ClientSession()
+
+        # initialize chatGPT class
+        self.openai_api_key = openai_api_key
+        if openai_api_key is not None:
+            # request header for !gpt command
+            self.headers = {
+                "Content-Type": "application/json",
+                "Authorization": f"Bearer {self.openai_api_key}",
+            }
+
+            self.askgpt = askGPT(
+                self.session,
+                self.headers,
+            )
+
+            self.gptchatbot = GPTChatBot(
+                api_key=self.openai_api_key, engine=self.gpt_engine
+            )
+        else:
+            logger.warning(
+                "openai_api_key is not provided, !gpt and !chat command will not work"
+            )
+
+        # initialize pandora
+        self.pandora_api_endpoint = pandora_api_endpoint
+        if pandora_api_endpoint is not None:
+            self.pandora = Pandora(
+                api_endpoint=pandora_api_endpoint, clientSession=self.session
+            )
+            if pandora_api_model is None:
+                self.pandora_api_model = "text-davinci-002-render-sha-mobile"
+            else:
+                self.pandora_api_model = pandora_api_model
+        self.pandora_data = {}
+
+        # initialize image generator
+        self.bing_auth_cookie = bing_auth_cookie
+        if bing_auth_cookie is not None:
+            self.imagegen = ImageGenAsync(auth_cookie=self.bing_auth_cookie)
+        else:
+            logger.warning(
+                "bing_auth_cookie is not provided, !pic command will not work"
+            )
+
+        # regular expression to match keyword
+        self.gpt_prog = re.compile(r"^\s*!gpt\s*(.+)$")
+        self.chat_prog = re.compile(r"^\s*!chat\s*(.+)$")
+        self.pic_prog = re.compile(r"^\s*!pic\s*(.+)$")
+        self.help_prog = re.compile(r"^\s*!help\s*.*$")
+        self.talk_prog = re.compile(r"^\s*!talk\s*(.+)$")
+        self.goon_prog = re.compile(r"^\s*!goon\s*.*$")
+        self.new_prog = re.compile(r"^\s*!new\s*.*$")
+
+    # close session
+    async def close(self, task: asyncio.Task) -> None:
+        await self.session.close()
+        self.driver.disconnect()
+        task.cancel()
+
+    async def login(self) -> None:
+        await self.driver.login()
+
+    def pandora_init(self, user_id: str) -> None:
+        self.pandora_data[user_id] = {
+            "conversation_id": None,
+            "parent_message_id": str(uuid.uuid4()),
+            "first_time": True,
+        }
+
+    async def run(self) -> None:
+        await self.driver.init_websocket(self.websocket_handler)
+
+    # websocket handler
+    async def websocket_handler(self, message) -> None:
+        logger.info(message)
+        response = json.loads(message)
+        if "event" in response:
+            event_type = response["event"]
+            if event_type == "posted":
+                raw_data = response["data"]["post"]
+                raw_data_dict = json.loads(raw_data)
+                user_id = raw_data_dict["user_id"]
+                channel_id = raw_data_dict["channel_id"]
+                sender_name = response["data"]["sender_name"]
+                raw_message = raw_data_dict["message"]
+
+                if user_id not in self.pandora_data:
+                    self.pandora_init(user_id)
+
+                try:
+                    asyncio.create_task(
+                        self.message_callback(
+                            raw_message, channel_id, user_id, sender_name
+                        )
+                    )
+                except Exception as e:
+                    await self.send_message(channel_id, f"{e}")
+
+    # message callback
+    async def message_callback(
+        self, raw_message: str, channel_id: str, user_id: str, sender_name: str
+    ) -> None:
+        # prevent command trigger loop
+        if sender_name != self.username:
+            message = raw_message
+
+            if self.openai_api_key is not None:
+                # !gpt command trigger handler
+                if self.gpt_prog.match(message):
+                    prompt = self.gpt_prog.match(message).group(1)
+                    try:
+                        response = await self.gpt(prompt)
+                        await self.send_message(channel_id, f"{response}")
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
+
+                # !chat command trigger handler
+                elif self.chat_prog.match(message):
+                    prompt = self.chat_prog.match(message).group(1)
+                    try:
+                        response = await self.chat(prompt)
+                        await self.send_message(channel_id, f"{response}")
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
+
+            if self.pandora_api_endpoint is not None:
+                # !talk command trigger handler
+                if self.talk_prog.match(message):
+                    prompt = self.talk_prog.match(message).group(1)
+                    try:
+                        if self.pandora_data[user_id]["conversation_id"] is not None:
+                            data = {
+                                "prompt": prompt,
+                                "model": self.pandora_api_model,
+                                "parent_message_id": self.pandora_data[user_id][
+                                    "parent_message_id"
+                                ],
+                                "conversation_id": self.pandora_data[user_id][
+                                    "conversation_id"
+                                ],
+                                "stream": False,
+                            }
+                        else:
+                            data = {
+                                "prompt": prompt,
+                                "model": self.pandora_api_model,
+                                "parent_message_id": self.pandora_data[user_id][
+                                    "parent_message_id"
+                                ],
+                                "stream": False,
+                            }
+                        response = await self.pandora.talk(data)
+                        self.pandora_data[user_id]["conversation_id"] = response[
+                            "conversation_id"
+                        ]
+                        self.pandora_data[user_id]["parent_message_id"] = response[
+                            "message"
+                        ]["id"]
+                        content = response["message"]["content"]["parts"][0]
+                        if self.pandora_data[user_id]["first_time"]:
+                            self.pandora_data[user_id]["first_time"] = False
+                            data = {
+                                "model": self.pandora_api_model,
+                                "message_id": self.pandora_data[user_id][
+                                    "parent_message_id"
+                                ],
+                            }
+                            await self.pandora.gen_title(
+                                data, self.pandora_data[user_id]["conversation_id"]
+                            )
+
+                        await self.send_message(channel_id, f"{content}")
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
+
+                # !goon command trigger handler
+                if (
+                    self.goon_prog.match(message)
+                    and self.pandora_data[user_id]["conversation_id"] is not None
+                ):
+                    try:
+                        data = {
+                            "model": self.pandora_api_model,
+                            "parent_message_id": self.pandora_data[user_id][
+                                "parent_message_id"
+                            ],
+                            "conversation_id": self.pandora_data[user_id][
+                                "conversation_id"
+                            ],
+                            "stream": False,
+                        }
+                        response = await self.pandora.goon(data)
+                        self.pandora_data[user_id]["conversation_id"] = response[
+                            "conversation_id"
+                        ]
+                        self.pandora_data[user_id]["parent_message_id"] = response[
+                            "message"
+                        ]["id"]
+                        content = response["message"]["content"]["parts"][0]
+                        await self.send_message(channel_id, f"{content}")
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
+
+                # !new command trigger handler
+                if self.new_prog.match(message):
+                    self.pandora_init(user_id)
+                    try:
+                        await self.send_message(
+                            channel_id,
+                            "New conversation created, " +
+                            "please use !talk to start chatting!",
+                        )
+                    except Exception:
+                        pass
+
+            if self.bing_auth_cookie is not None:
+                # !pic command trigger handler
+                if self.pic_prog.match(message):
+                    prompt = self.pic_prog.match(message).group(1)
+                    # generate image
+                    try:
+                        links = await self.imagegen.get_images(prompt)
+                        image_path = await self.imagegen.save_images(links, "images")
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
+
+                    # send image
+                    try:
+                        await self.send_file(channel_id, prompt, image_path)
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
+
+            # !help command trigger handler
+            if self.help_prog.match(message):
+                try:
+                    await self.send_message(channel_id, self.help())
+                except Exception as e:
+                    logger.error(e, exc_info=True)
+
+    # send message to room
+    async def send_message(self, channel_id: str, message: str) -> None:
+        await self.driver.posts.create_post(
+            options={"channel_id": channel_id, "message": message}
+        )
+
+    # send file to room
+    async def send_file(self, channel_id: str, message: str, filepath: str) -> None:
+        filename = os.path.split(filepath)[-1]
+        try:
+            file_id = await self.driver.files.upload_file(
+                channel_id=channel_id,
+                files={
+                    "files": (filename, open(filepath, "rb")),
+                },
+            )["file_infos"][0]["id"]
+        except Exception as e:
+            logger.error(e, exc_info=True)
+            raise Exception(e)
+
+        try:
+            await self.driver.posts.create_post(
+                options={
+                    "channel_id": channel_id,
+                    "message": message,
+                    "file_ids": [file_id],
+                }
+            )
+            # remove image after posting
+            os.remove(filepath)
+        except Exception as e:
+            logger.error(e, exc_info=True)
+            raise Exception(e)
+
+    # !gpt command function
+    async def gpt(self, prompt: str) -> str:
+        return await self.askgpt.oneTimeAsk(prompt)
+
+    # !chat command function
+    async def chat(self, prompt: str) -> str:
+        return await self.gptchatbot.ask_async(prompt)
+
+    # !help command function
+    def help(self) -> str:
+        help_info = (
+            "!gpt [content], generate response without context conversation\n"
+            + "!chat [content], chat with context conversation\n"
+            + "!pic [prompt], Image generation by Microsoft Bing\n"
+            + "!talk [content], talk using chatgpt web\n"
+            + "!goon, continue the incomplete conversation\n"
+            + "!new, start a new conversation\n"
+            + "!help, help message"
+        )
+        return help_info
diff --git a/log.py b/src/log.py
similarity index 96%
rename from log.py
rename to src/log.py
index a59ca5c..32f7a8c 100644
--- a/log.py
+++ b/src/log.py
@@ -1,30 +1,30 @@
-import logging
-
-
-def getlogger():
-    # create a custom logger if not already created
-    logger = logging.getLogger(__name__)
-    if not logger.hasHandlers():
-        logger.setLevel(logging.INFO)
-
-        # create handlers
-        info_handler = logging.StreamHandler()
-        error_handler = logging.FileHandler("bot.log", mode="a")
-        error_handler.setLevel(logging.ERROR)
-        info_handler.setLevel(logging.INFO)
-
-        # create formatters
-        error_format = logging.Formatter(
-            "%(asctime)s - %(name)s - %(funcName)s - %(levelname)s - %(message)s"
-        )
-        info_format = logging.Formatter("%(asctime)s - %(levelname)s - %(message)s")
-
-        # set formatter
-        error_handler.setFormatter(error_format)
-        info_handler.setFormatter(info_format)
-
-        # add handlers to logger
-        logger.addHandler(error_handler)
-        logger.addHandler(info_handler)
-
-    return logger
+import logging
+
+
+def getlogger():
+    # create a custom logger if not already created
+    logger = logging.getLogger(__name__)
+    if not logger.hasHandlers():
+        logger.setLevel(logging.INFO)
+
+        # create handlers
+        info_handler = logging.StreamHandler()
+        error_handler = logging.FileHandler("bot.log", mode="a")
+        error_handler.setLevel(logging.ERROR)
+        info_handler.setLevel(logging.INFO)
+
+        # create formatters
+        error_format = logging.Formatter(
+            "%(asctime)s - %(name)s - %(funcName)s - %(levelname)s - %(message)s"
+        )
+        info_format = logging.Formatter("%(asctime)s - %(levelname)s - %(message)s")
+
+        # set formatter
+        error_handler.setFormatter(error_format)
+        info_handler.setFormatter(info_format)
+
+        # add handlers to logger
+        logger.addHandler(error_handler)
+        logger.addHandler(info_handler)
+
+    return logger
diff --git a/main.py b/src/main.py
similarity index 64%
rename from main.py
rename to src/main.py
index a8f2754..ed624c7 100644
--- a/main.py
+++ b/src/main.py
@@ -1,53 +1,70 @@
-from bot import Bot
-import json
-import os
-import asyncio
-
-async def main():
-    if os.path.exists("config.json"):
-        fp = open("config.json", "r", encoding="utf-8")
-        config = json.load(fp)
-
-        mattermost_bot = Bot(
-            server_url=config.get("server_url"),
-            access_token=config.get("access_token"),
-            login_id=config.get("login_id"),
-            password=config.get("password"),
-            username=config.get("username"),
-            openai_api_key=config.get("openai_api_key"),
-            openai_api_endpoint=config.get("openai_api_endpoint"),
-            bing_api_endpoint=config.get("bing_api_endpoint"),
-            bard_token=config.get("bard_token"),
-            bing_auth_cookie=config.get("bing_auth_cookie"),
-            pandora_api_endpoint=config.get("pandora_api_endpoint"),
-            pandora_api_model=config.get("pandora_api_model"),
-            port=config.get("port"),
-            timeout=config.get("timeout"),
-        )
-
-    else:
-        mattermost_bot = Bot(
-            server_url=os.environ.get("SERVER_URL"),
-            access_token=os.environ.get("ACCESS_TOKEN"),
-            login_id=os.environ.get("LOGIN_ID"),
-            password=os.environ.get("PASSWORD"),
-            username=os.environ.get("USERNAME"),
-            openai_api_key=os.environ.get("OPENAI_API_KEY"),
-            openai_api_endpoint=os.environ.get("OPENAI_API_ENDPOINT"),
-            bing_api_endpoint=os.environ.get("BING_API_ENDPOINT"),
-            bard_token=os.environ.get("BARD_TOKEN"),
-            bing_auth_cookie=os.environ.get("BING_AUTH_COOKIE"),
-            pandora_api_endpoint=os.environ.get("PANDORA_API_ENDPOINT"),
-            pandora_api_model=os.environ.get("PANDORA_API_MODEL"),
-            port=os.environ.get("PORT"),
-            timeout=os.environ.get("TIMEOUT"),
-        )
-
-    mattermost_bot.login()
-
-    await mattermost_bot.run()
-
-
-if __name__ == "__main__":
-    asyncio.run(main())
-
+import signal
+from bot import Bot
+import json
+import os
+import asyncio
+from pathlib import Path
+from log import getlogger
+
+logger = getlogger()
+
+
+async def main():
+    config_path = Path(os.path.dirname(__file__)).parent / "config.json"
+    if os.path.isfile(config_path):
+        fp = open("config.json", "r", encoding="utf-8")
+        config = json.load(fp)
+
+        mattermost_bot = Bot(
+            server_url=config.get("server_url"),
+            access_token=config.get("access_token"),
+            login_id=config.get("login_id"),
+            password=config.get("password"),
+            username=config.get("username"),
+            openai_api_key=config.get("openai_api_key"),
+            bing_auth_cookie=config.get("bing_auth_cookie"),
+            pandora_api_endpoint=config.get("pandora_api_endpoint"),
+            pandora_api_model=config.get("pandora_api_model"),
+            port=config.get("port"),
+            scheme=config.get("scheme"),
+            timeout=config.get("timeout"),
+            gpt_engine=config.get("gpt_engine"),
+        )
+
+    else:
+        mattermost_bot = Bot(
+            server_url=os.environ.get("SERVER_URL"),
+            access_token=os.environ.get("ACCESS_TOKEN"),
+            login_id=os.environ.get("LOGIN_ID"),
+            password=os.environ.get("PASSWORD"),
+            username=os.environ.get("USERNAME"),
+            openai_api_key=os.environ.get("OPENAI_API_KEY"),
+            bing_auth_cookie=os.environ.get("BING_AUTH_COOKIE"),
+            pandora_api_endpoint=os.environ.get("PANDORA_API_ENDPOINT"),
+            pandora_api_model=os.environ.get("PANDORA_API_MODEL"),
+            port=os.environ.get("PORT"),
+            scheme=os.environ.get("SCHEME"),
+            timeout=os.environ.get("TIMEOUT"),
+            gpt_engine=os.environ.get("GPT_ENGINE"),
+        )
+
+    await mattermost_bot.login()
+
+    task = asyncio.create_task(mattermost_bot.run())
+
+    # handle signal interrupt
+    loop = asyncio.get_running_loop()
+    for signame in ("SIGINT", "SIGTERM"):
+        loop.add_signal_handler(
+            getattr(signal, signame),
+            lambda: asyncio.create_task(mattermost_bot.close(task)),
+        )
+
+    try:
+        await task
+    except asyncio.CancelledError:
+        logger.info("Bot stopped")
+
+
+if __name__ == "__main__":
+    asyncio.run(main())
diff --git a/pandora.py b/src/pandora.py
similarity index 86%
rename from pandora.py
rename to src/pandora.py
index df4cfa0..c6b5997 100644
--- a/pandora.py
+++ b/src/pandora.py
@@ -2,14 +2,16 @@
 import uuid
 import aiohttp
 import asyncio
+
+
 class Pandora:
     def __init__(self, api_endpoint: str, clientSession: aiohttp.ClientSession) -> None:
-        self.api_endpoint = api_endpoint.rstrip('/')
+        self.api_endpoint = api_endpoint.rstrip("/")
         self.session = clientSession
 
     async def __aenter__(self):
         return self
-    
+
     async def __aexit__(self, exc_type, exc_val, exc_tb):
         await self.session.close()
 
@@ -23,7 +25,9 @@ class Pandora:
         :param conversation_id: str
         :return: None
         """
-        api_endpoint = self.api_endpoint + f"/api/conversation/gen_title/{conversation_id}"
+        api_endpoint = (
+            self.api_endpoint + f"/api/conversation/gen_title/{conversation_id}"
+        )
         async with self.session.post(api_endpoint, json=data) as resp:
             return await resp.json()
 
@@ -40,10 +44,10 @@ class Pandora:
         :param data: dict
         :return: None
         """
-        data['message_id'] = str(uuid.uuid4())
+        data["message_id"] = str(uuid.uuid4())
         async with self.session.post(api_endpoint, json=data) as resp:
             return await resp.json()
-        
+
     async def goon(self, data: dict) -> None:
         """
         data = {
@@ -56,7 +60,8 @@ class Pandora:
         api_endpoint = self.api_endpoint + "/api/conversation/goon"
         async with self.session.post(api_endpoint, json=data) as resp:
             return await resp.json()
-    
+
+
 async def test():
     model = "text-davinci-002-render-sha-mobile"
     api_endpoint = "http://127.0.0.1:8008"
@@ -84,9 +89,9 @@ async def test():
                     "stream": False,
                 }
             response = await client.talk(data)
-            conversation_id = response['conversation_id']
-            parent_message_id = response['message']['id']
-            content = response['message']['content']['parts'][0]
+            conversation_id = response["conversation_id"]
+            parent_message_id = response["message"]["id"]
+            content = response["message"]["content"]["parts"][0]
             print("ChatGPT: " + content + "\n")
             if first_time:
                 first_time = False
@@ -97,5 +102,5 @@ async def test():
                 response = await client.gen_title(data, conversation_id)
 
 
-if __name__ == '__main__':
-    asyncio.run(test())
\ No newline at end of file
+if __name__ == "__main__":
+    asyncio.run(test())
diff --git a/v3.py b/v3.py
deleted file mode 100644
index e3d7f44..0000000
--- a/v3.py
+++ /dev/null
@@ -1,324 +0,0 @@
-"""
-Code derived from: https://github.com/acheong08/ChatGPT/blob/main/src/revChatGPT/V3.py
-"""
-
-import json
-import os
-from typing import AsyncGenerator
-import httpx
-import requests
-import tiktoken
-
-
-class Chatbot:
-    """
-    Official ChatGPT API
-    """
-
-    def __init__(
-        self,
-        api_key: str,
-        engine: str = os.environ.get("GPT_ENGINE") or "gpt-3.5-turbo",
-        proxy: str = None,
-        timeout: float = None,
-        max_tokens: int = None,
-        temperature: float = 0.5,
-        top_p: float = 1.0,
-        presence_penalty: float = 0.0,
-        frequency_penalty: float = 0.0,
-        reply_count: int = 1,
-        system_prompt: str = "You are ChatGPT, a large language model trained by OpenAI. Respond conversationally",
-    ) -> None:
-        """
-        Initialize Chatbot with API key (from https://platform.openai.com/account/api-keys)
-        """
-        self.engine: str = engine
-        self.api_key: str = api_key
-        self.system_prompt: str = system_prompt
-        self.max_tokens: int = max_tokens or (
-            31000 if engine == "gpt-4-32k" else 7000 if engine == "gpt-4" else 4000
-        )
-        self.truncate_limit: int = (
-            30500 if engine == "gpt-4-32k" else 6500 if engine == "gpt-4" else 3500
-        )
-        self.temperature: float = temperature
-        self.top_p: float = top_p
-        self.presence_penalty: float = presence_penalty
-        self.frequency_penalty: float = frequency_penalty
-        self.reply_count: int = reply_count
-        self.timeout: float = timeout
-        self.proxy = proxy
-        self.session = requests.Session()
-        self.session.proxies.update(
-            {
-                "http": proxy,
-                "https": proxy,
-            },
-        )
-        proxy = (
-            proxy or os.environ.get("all_proxy") or os.environ.get("ALL_PROXY") or None
-        )
-
-        if proxy:
-            if "socks5h" not in proxy:
-                self.aclient = httpx.AsyncClient(
-                    follow_redirects=True,
-                    proxies=proxy,
-                    timeout=timeout,
-                )
-        else:
-            self.aclient = httpx.AsyncClient(
-                follow_redirects=True,
-                proxies=proxy,
-                timeout=timeout,
-            )
-
-        self.conversation: dict[str, list[dict]] = {
-            "default": [
-                {
-                    "role": "system",
-                    "content": system_prompt,
-                },
-            ],
-        }
-
-    def add_to_conversation(
-        self,
-        message: str,
-        role: str,
-        convo_id: str = "default",
-    ) -> None:
-        """
-        Add a message to the conversation
-        """
-        self.conversation[convo_id].append({"role": role, "content": message})
-
-    def __truncate_conversation(self, convo_id: str = "default") -> None:
-        """
-        Truncate the conversation
-        """
-        while True:
-            if (
-                self.get_token_count(convo_id) > self.truncate_limit
-                and len(self.conversation[convo_id]) > 1
-            ):
-                # Don't remove the first message
-                self.conversation[convo_id].pop(1)
-            else:
-                break
-
-    def get_token_count(self, convo_id: str = "default") -> int:
-        """
-        Get token count
-        """
-        if self.engine not in [
-            "gpt-3.5-turbo",
-            "gpt-3.5-turbo-0301",
-            "gpt-4",
-            "gpt-4-0314",
-            "gpt-4-32k",
-            "gpt-4-32k-0314",
-        ]:
-            raise NotImplementedError("Unsupported engine {self.engine}")
-
-        tiktoken.model.MODEL_TO_ENCODING["gpt-4"] = "cl100k_base"
-
-        encoding = tiktoken.encoding_for_model(self.engine)
-
-        num_tokens = 0
-        for message in self.conversation[convo_id]:
-            # every message follows <im_start>{role/name}\n{content}<im_end>\n
-            num_tokens += 5
-            for key, value in message.items():
-                num_tokens += len(encoding.encode(value))
-                if key == "name":  # if there's a name, the role is omitted
-                    num_tokens += 5  # role is always required and always 1 token
-        num_tokens += 5  # every reply is primed with <im_start>assistant
-        return num_tokens
-
-    def get_max_tokens(self, convo_id: str) -> int:
-        """
-        Get max tokens
-        """
-        return self.max_tokens - self.get_token_count(convo_id)
-
-    def ask_stream(
-        self,
-        prompt: str,
-        role: str = "user",
-        convo_id: str = "default",
-        **kwargs,
-    ):
-        """
-        Ask a question
-        """
-        # Make conversation if it doesn't exist
-        if convo_id not in self.conversation:
-            self.reset(convo_id=convo_id, system_prompt=self.system_prompt)
-        self.add_to_conversation(prompt, "user", convo_id=convo_id)
-        self.__truncate_conversation(convo_id=convo_id)
-        # Get response
-        response = self.session.post(
-            os.environ.get("API_URL") or "https://api.openai.com/v1/chat/completions",
-            headers={"Authorization": f"Bearer {kwargs.get('api_key', self.api_key)}"},
-            json={
-                "model": self.engine,
-                "messages": self.conversation[convo_id],
-                "stream": True,
-                # kwargs
-                "temperature": kwargs.get("temperature", self.temperature),
-                "top_p": kwargs.get("top_p", self.top_p),
-                "presence_penalty": kwargs.get(
-                    "presence_penalty",
-                    self.presence_penalty,
-                ),
-                "frequency_penalty": kwargs.get(
-                    "frequency_penalty",
-                    self.frequency_penalty,
-                ),
-                "n": kwargs.get("n", self.reply_count),
-                "user": role,
-                "max_tokens": self.get_max_tokens(convo_id=convo_id),
-            },
-            timeout=kwargs.get("timeout", self.timeout),
-            stream=True,
-        )
-
-        response_role: str = None
-        full_response: str = ""
-        for line in response.iter_lines():
-            if not line:
-                continue
-            # Remove "data: "
-            line = line.decode("utf-8")[6:]
-            if line == "[DONE]":
-                break
-            resp: dict = json.loads(line)
-            choices = resp.get("choices")
-            if not choices:
-                continue
-            delta = choices[0].get("delta")
-            if not delta:
-                continue
-            if "role" in delta:
-                response_role = delta["role"]
-            if "content" in delta:
-                content = delta["content"]
-                full_response += content
-                yield content
-        self.add_to_conversation(full_response, response_role, convo_id=convo_id)
-
-    async def ask_stream_async(
-        self,
-        prompt: str,
-        role: str = "user",
-        convo_id: str = "default",
-        **kwargs,
-    ) -> AsyncGenerator[str, None]:
-        """
-        Ask a question
-        """
-        # Make conversation if it doesn't exist
-        if convo_id not in self.conversation:
-            self.reset(convo_id=convo_id, system_prompt=self.system_prompt)
-        self.add_to_conversation(prompt, "user", convo_id=convo_id)
-        self.__truncate_conversation(convo_id=convo_id)
-        # Get response
-        async with self.aclient.stream(
-            "post",
-            os.environ.get("API_URL") or "https://api.openai.com/v1/chat/completions",
-            headers={"Authorization": f"Bearer {kwargs.get('api_key', self.api_key)}"},
-            json={
-                "model": self.engine,
-                "messages": self.conversation[convo_id],
-                "stream": True,
-                # kwargs
-                "temperature": kwargs.get("temperature", self.temperature),
-                "top_p": kwargs.get("top_p", self.top_p),
-                "presence_penalty": kwargs.get(
-                    "presence_penalty",
-                    self.presence_penalty,
-                ),
-                "frequency_penalty": kwargs.get(
-                    "frequency_penalty",
-                    self.frequency_penalty,
-                ),
-                "n": kwargs.get("n", self.reply_count),
-                "user": role,
-                "max_tokens": self.get_max_tokens(convo_id=convo_id),
-            },
-            timeout=kwargs.get("timeout", self.timeout),
-        ) as response:
-            if response.status_code != 200:
-                await response.aread()
-
-            response_role: str = ""
-            full_response: str = ""
-            async for line in response.aiter_lines():
-                line = line.strip()
-                if not line:
-                    continue
-                # Remove "data: "
-                line = line[6:]
-                if line == "[DONE]":
-                    break
-                resp: dict = json.loads(line)
-                choices = resp.get("choices")
-                if not choices:
-                    continue
-                delta: dict[str, str] = choices[0].get("delta")
-                if not delta:
-                    continue
-                if "role" in delta:
-                    response_role = delta["role"]
-                if "content" in delta:
-                    content: str = delta["content"]
-                    full_response += content
-                    yield content
-        self.add_to_conversation(full_response, response_role, convo_id=convo_id)
-
-    async def ask_async(
-        self,
-        prompt: str,
-        role: str = "user",
-        convo_id: str = "default",
-        **kwargs,
-    ) -> str:
-        """
-        Non-streaming ask
-        """
-        response = self.ask_stream_async(
-            prompt=prompt,
-            role=role,
-            convo_id=convo_id,
-            **kwargs,
-        )
-        full_response: str = "".join([r async for r in response])
-        return full_response
-
-    def ask(
-        self,
-        prompt: str,
-        role: str = "user",
-        convo_id: str = "default",
-        **kwargs,
-    ) -> str:
-        """
-        Non-streaming ask
-        """
-        response = self.ask_stream(
-            prompt=prompt,
-            role=role,
-            convo_id=convo_id,
-            **kwargs,
-        )
-        full_response: str = "".join(response)
-        return full_response
-
-    def reset(self, convo_id: str = "default", system_prompt: str = None) -> None:
-        """
-        Reset the conversation
-        """
-        self.conversation[convo_id] = [
-            {"role": "system", "content": system_prompt or self.system_prompt},
-        ]

From 25fbd43a57edbd5f2e474f4c6c0bf0a66b14c02b Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sun, 6 Aug 2023 12:34:10 +0800
Subject: [PATCH 02/15] Update README.md

---
 README.md | 2 --
 1 file changed, 2 deletions(-)

diff --git a/README.md b/README.md
index be8493a..c837386 100644
--- a/README.md
+++ b/README.md
@@ -21,8 +21,6 @@ docker compose up -d
 - `!help` help message
 - `!gpt + [prompt]` generate a one time response from chatGPT
 - `!chat + [prompt]` chat using official chatGPT api with context conversation
-- `!bing + [prompt]` chat with Bing AI with context conversation
-- `!bard + [prompt]` chat with Google's Bard
 - `!pic + [prompt]` generate an image from Bing Image Creator
 
 The following commands need pandora http api: https://github.com/pengzhile/pandora/blob/master/doc/wiki_en.md#http-restful-api

From 714204529260cb5617b2ddcd8735196811e5c54d Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Mon, 18 Sep 2023 13:47:50 +0800
Subject: [PATCH 03/15] feat: refactor chat backend feat: introduce pre-commit
 hooks feat: send reply in thread fix: !gpt !chat API endpoint and API key
 validation logic

---
 .dockerignore                        |   2 +-
 .env.example                         |  10 +-
 .full-env.example                    |  19 ++
 .github/workflows/docker-release.yml |   2 +-
 .github/workflows/pylint.yml         |   2 +-
 .gitignore                           |   2 +-
 .pre-commit-config.yaml              |  16 ++
 CHANGELOG.md                         |   7 +-
 README.md                            |   4 +-
 compose.yaml                         |   2 +-
 config.json.example                  |  12 +-
 full-config.json.example             |  21 ++
 requirements.txt                     |   6 +-
 src/BingImageGen.py                  | 165 ------------
 src/askgpt.py                        |  46 ----
 src/bot.py                           | 376 ++++++++++-----------------
 src/gptbot.py                        | 296 +++++++++++++++++++++
 src/imagegen.py                      |  69 +++++
 src/main.py                          |  47 ++--
 src/pandora.py                       | 106 --------
 20 files changed, 610 insertions(+), 600 deletions(-)
 create mode 100644 .full-env.example
 create mode 100644 .pre-commit-config.yaml
 create mode 100644 full-config.json.example
 delete mode 100644 src/BingImageGen.py
 delete mode 100644 src/askgpt.py
 create mode 100644 src/gptbot.py
 create mode 100644 src/imagegen.py
 delete mode 100644 src/pandora.py

diff --git a/.dockerignore b/.dockerignore
index 7495bae..fa0b4b7 100644
--- a/.dockerignore
+++ b/.dockerignore
@@ -23,4 +23,4 @@ src/__pycache__
 .github
 settings.js
 mattermost-server
-tests
\ No newline at end of file
+tests
diff --git a/.env.example b/.env.example
index 65f8e92..77e8bed 100644
--- a/.env.example
+++ b/.env.example
@@ -1,8 +1,6 @@
 SERVER_URL="xxxxx.xxxxxx.xxxxxxxxx"
-ACCESS_TOKEN="xxxxxxxxxxxxxxxxx"
 USERNAME="@chatgpt"
-OPENAI_API_KEY="sk-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
-BING_AUTH_COOKIE="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
-PANDORA_API_ENDPOINT="http://pandora:8008"
-PANDORA_API_MODEL="text-davinci-002-render-sha-mobile"
-GPT_ENGINE="gpt-3.5-turbo"
\ No newline at end of file
+EMAIL="xxxxxx"
+PASSWORD="xxxxxxxxxxxxxx"
+OPENAI_API_KEY="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
+GPT_MODEL="gpt-3.5-turbo"
diff --git a/.full-env.example b/.full-env.example
new file mode 100644
index 0000000..53c3020
--- /dev/null
+++ b/.full-env.example
@@ -0,0 +1,19 @@
+SERVER_URL="xxxxx.xxxxxx.xxxxxxxxx"
+EMAIL="xxxxxx"
+USERNAME="@chatgpt"
+PASSWORD="xxxxxxxxxxxxxx"
+PORT=443
+SCHEME="https"
+OPENAI_API_KEY="xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx"
+GPT_API_ENDPOINT="https://api.openai.com/v1/chat/completions"
+GPT_MODEL="gpt-3.5-turbo"
+MAX_TOKENS=4000
+TOP_P=1.0
+PRESENCE_PENALTY=0.0
+FREQUENCY_PENALTY=0.0
+REPLY_COUNT=1
+SYSTEM_PROMPT="You are ChatGPT,  a large language model trained by OpenAI. Respond conversationally"
+TEMPERATURE=0.8
+IMAGE_GENERATION_ENDPOINT="http://127.0.0.1:7860/sdapi/v1/txt2img"
+IMAGE_GENERATION_BACKEND="sdwui" # openai or sdwui
+TIMEOUT=120.0
diff --git a/.github/workflows/docker-release.yml b/.github/workflows/docker-release.yml
index 7b11b5a..782212c 100644
--- a/.github/workflows/docker-release.yml
+++ b/.github/workflows/docker-release.yml
@@ -70,4 +70,4 @@ jobs:
           tags: ${{ steps.meta2.outputs.tags }}
           labels: ${{ steps.meta2.outputs.labels }}
           cache-from: type=gha
-          cache-to: type=gha,mode=max
\ No newline at end of file
+          cache-to: type=gha,mode=max
diff --git a/.github/workflows/pylint.yml b/.github/workflows/pylint.yml
index fca44f6..ff31b8c 100644
--- a/.github/workflows/pylint.yml
+++ b/.github/workflows/pylint.yml
@@ -22,4 +22,4 @@ jobs:
         pip install pylint
     - name: Analysing the code with pylint
       run: |
-        pylint $(git ls-files '*.py') --errors-only
\ No newline at end of file
+        pylint $(git ls-files '*.py') --errors-only
diff --git a/.gitignore b/.gitignore
index bd51f1a..6f29859 100644
--- a/.gitignore
+++ b/.gitignore
@@ -137,4 +137,4 @@ dmypy.json
 
 # custom
 compose-dev.yaml
-mattermost-server
\ No newline at end of file
+mattermost-server
diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
new file mode 100644
index 0000000..d811573
--- /dev/null
+++ b/.pre-commit-config.yaml
@@ -0,0 +1,16 @@
+repos:
+  - repo: https://github.com/pre-commit/pre-commit-hooks
+    rev: v4.4.0
+    hooks:
+      - id: trailing-whitespace
+      - id: end-of-file-fixer
+      - id: check-yaml
+  - repo: https://github.com/psf/black
+    rev: 23.9.1
+    hooks:
+      - id: black
+  - repo: https://github.com/astral-sh/ruff-pre-commit
+    rev: v0.0.289
+    hooks:
+      - id: ruff
+        args: [--fix, --exit-non-zero-on-fix]
diff --git a/CHANGELOG.md b/CHANGELOG.md
index 64e9ef2..d66b2ed 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,7 +1,12 @@
 # Changelog
 
+## v1.1.0
+- remove pandora
+- refactor chat and image genderation backend
+- reply in thread by default
+
 ## v1.0.4
 
 - refactor code structure and remove unused
 - remove Bing AI and Google Bard due to technical problems
-- bug fix and improvement
\ No newline at end of file
+- bug fix and improvement
diff --git a/README.md b/README.md
index c837386..f5a4cc8 100644
--- a/README.md
+++ b/README.md
@@ -1,6 +1,6 @@
 ## Introduction
 
-This is a simple Mattermost Bot that uses OpenAI's GPT API to generate responses to user inputs. The bot responds to these commands: `!gpt`, `!chat` and `!talk` and `!goon` and `!new` and  `!help` depending on the first word of the prompt.
+This is a simple Mattermost Bot that uses OpenAI's GPT API(or self-host models) to generate responses to user inputs. The bot responds to these commands: `!gpt`, `!chat` and `!new` and  `!help` depending on the first word of the prompt.
 
 ## Feature
 
@@ -26,7 +26,7 @@ docker compose up -d
 The following commands need pandora http api: https://github.com/pengzhile/pandora/blob/master/doc/wiki_en.md#http-restful-api
 - `!talk + [prompt]` chat using chatGPT web with context conversation
 - `!goon` ask chatGPT to complete the missing part from previous conversation
-- `!new` start a new converstaion 
+- `!new` start a new converstaion
 
 ## Demo
 Remove support for Bing AI, Google Bard due to technical problems.
diff --git a/compose.yaml b/compose.yaml
index d3500df..e7e713e 100644
--- a/compose.yaml
+++ b/compose.yaml
@@ -22,4 +22,4 @@ services:
   #     - mattermost_network
 
 networks:
-  mattermost_network:
\ No newline at end of file
+  mattermost_network:
diff --git a/config.json.example b/config.json.example
index b6258bc..50d59bf 100644
--- a/config.json.example
+++ b/config.json.example
@@ -1,10 +1,8 @@
 {
     "server_url": "xxxx.xxxx.xxxxx",
-    "access_token": "xxxxxxxxxxxxxxxxxxxxxx",
+    "email": "xxxxx",
     "username": "@chatgpt",
-    "openai_api_key": "sk-xxxxxxxxxxxxxxxxxxx",
-    "gpt_engine": "gpt-3.5-turbo",
-    "bing_auth_cookie": "xxxxxxxxxxxxxxxxxxxxxxxxxxxxxx",
-    "pandora_api_endpoint": "http://127.0.0.1:8008",
-    "pandora_api_model": "text-davinci-002-render-sha-mobile"
-}
\ No newline at end of file
+    "password": "xxxxxxxxxxxxxxxxx",
+    "openai_api_key": "xxxxxxxxxxxxxxxxxxxxxxxxx",
+    "gpt_model": "gpt-3.5-turbo"
+}
diff --git a/full-config.json.example b/full-config.json.example
new file mode 100644
index 0000000..5215c90
--- /dev/null
+++ b/full-config.json.example
@@ -0,0 +1,21 @@
+{
+    "server_url": "localhost",
+    "email": "bot@hibobmaster.com",
+    "username": "@bot",
+    "password": "SfBKY%K7*e&a%ZX$3g@Am&jQ",
+    "port": "8065",
+    "scheme": "http",
+    "openai_api_key": "xxxxxxxxxxxxxxxxxxxxxxxx",
+    "gpt_api_endpoint": "https://api.openai.com/v1/chat/completions",
+    "gpt_model": "gpt-3.5-turbo",
+    "max_tokens": 4000,
+    "top_p": 1.0,
+    "presence_penalty": 0.0,
+    "frequency_penalty": 0.0,
+    "reply_count": 1,
+    "temperature": 0.8,
+    "system_prompt": "You are ChatGPT,  a large language model trained by OpenAI. Respond conversationally",
+    "image_generation_endpoint": "http://localai:8080/v1/images/generations",
+    "image_generation_backend": "openai",
+    "timeout": 120.0
+}
diff --git a/requirements.txt b/requirements.txt
index 42932c4..c7f6d1f 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -1,4 +1,6 @@
-aiohttp
+aiofiles
 httpx
+Pillow
+tiktoken
+tenacity
 mattermostdriver @ git+https://github.com/hibobmaster/python-mattermost-driver
-revChatGPT>=6.8.6
\ No newline at end of file
diff --git a/src/BingImageGen.py b/src/BingImageGen.py
deleted file mode 100644
index 979346c..0000000
--- a/src/BingImageGen.py
+++ /dev/null
@@ -1,165 +0,0 @@
-"""
-Code derived from:
-https://github.com/acheong08/EdgeGPT/blob/f940cecd24a4818015a8b42a2443dd97c3c2a8f4/src/ImageGen.py
-"""
-from log import getlogger
-from uuid import uuid4
-import os
-import contextlib
-import aiohttp
-import asyncio
-import random
-import requests
-import regex
-
-logger = getlogger()
-
-BING_URL = "https://www.bing.com"
-# Generate random IP between range 13.104.0.0/14
-FORWARDED_IP = (
-    f"13.{random.randint(104, 107)}.{random.randint(0, 255)}.{random.randint(0, 255)}"
-)
-HEADERS = {
-    "accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8,application/signed-exchange;v=b3;q=0.7",
-    "accept-language": "en-US,en;q=0.9",
-    "cache-control": "max-age=0",
-    "content-type": "application/x-www-form-urlencoded",
-    "referrer": "https://www.bing.com/images/create/",
-    "origin": "https://www.bing.com",
-    "user-agent": "Mozilla/5.0 (X11; Linux x86_64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/110.0.0.0 Safari/537.36 Edg/110.0.1587.63",
-    "x-forwarded-for": FORWARDED_IP,
-}
-
-
-class ImageGenAsync:
-    """
-    Image generation by Microsoft Bing
-    Parameters:
-        auth_cookie: str
-    """
-
-    def __init__(self, auth_cookie: str, quiet: bool = True) -> None:
-        self.session = aiohttp.ClientSession(
-            headers=HEADERS,
-            cookies={"_U": auth_cookie},
-        )
-        self.quiet = quiet
-
-    async def __aenter__(self):
-        return self
-
-    async def __aexit__(self, *excinfo) -> None:
-        await self.session.close()
-
-    def __del__(self):
-        try:
-            loop = asyncio.get_running_loop()
-        except RuntimeError:
-            loop = asyncio.new_event_loop()
-            asyncio.set_event_loop(loop)
-        loop.run_until_complete(self._close())
-
-    async def _close(self):
-        await self.session.close()
-
-    async def get_images(self, prompt: str) -> list:
-        """
-        Fetches image links from Bing
-        Parameters:
-            prompt: str
-        """
-        if not self.quiet:
-            print("Sending request...")
-        url_encoded_prompt = requests.utils.quote(prompt)
-        # https://www.bing.com/images/create?q=<PROMPT>&rt=3&FORM=GENCRE
-        url = f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=4&FORM=GENCRE"
-        async with self.session.post(url, allow_redirects=False) as response:
-            content = await response.text()
-            if "this prompt has been blocked" in content.lower():
-                raise Exception(
-                    "Your prompt has been blocked by Bing. Try to change any bad words and try again.",
-                )
-            if response.status != 302:
-                # if rt4 fails, try rt3
-                url = (
-                    f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=3&FORM=GENCRE"
-                )
-                async with self.session.post(
-                    url,
-                    allow_redirects=False,
-                    timeout=200,
-                ) as response3:
-                    if response3.status != 302:
-                        print(f"ERROR: {response3.text}")
-                        raise Exception("Redirect failed")
-                    response = response3
-        # Get redirect URL
-        redirect_url = response.headers["Location"].replace("&nfy=1", "")
-        request_id = redirect_url.split("id=")[-1]
-        await self.session.get(f"{BING_URL}{redirect_url}")
-        # https://www.bing.com/images/create/async/results/{ID}?q={PROMPT}
-        polling_url = f"{BING_URL}/images/create/async/results/{request_id}?q={url_encoded_prompt}"
-        # Poll for results
-        if not self.quiet:
-            print("Waiting for results...")
-        while True:
-            if not self.quiet:
-                print(".", end="", flush=True)
-            # By default, timeout is 300s, change as needed
-            response = await self.session.get(polling_url)
-            if response.status != 200:
-                raise Exception("Could not get results")
-            content = await response.text()
-            if content and content.find("errorMessage") == -1:
-                break
-
-            await asyncio.sleep(1)
-            continue
-        # Use regex to search for src=""
-        image_links = regex.findall(r'src="([^"]+)"', content)
-        # Remove size limit
-        normal_image_links = [link.split("?w=")[0] for link in image_links]
-        # Remove duplicates
-        normal_image_links = list(set(normal_image_links))
-
-        # Bad images
-        bad_images = [
-            "https://r.bing.com/rp/in-2zU3AJUdkgFe7ZKv19yPBHVs.png",
-            "https://r.bing.com/rp/TX9QuO3WzcCJz1uaaSwQAz39Kb0.jpg",
-        ]
-        for im in normal_image_links:
-            if im in bad_images:
-                raise Exception("Bad images")
-        # No images
-        if not normal_image_links:
-            raise Exception("No images")
-        return normal_image_links
-
-    async def save_images(self, links: list, output_dir: str) -> str:
-        """
-        Saves images to output directory
-        """
-        if not self.quiet:
-            print("\nDownloading images...")
-        with contextlib.suppress(FileExistsError):
-            os.mkdir(output_dir)
-
-        # image name
-        image_name = str(uuid4())
-        # we just need one image for better display in chat room
-        if links:
-            link = links.pop()
-
-        image_path = os.path.join(output_dir, f"{image_name}.jpeg")
-        try:
-            async with self.session.get(link, raise_for_status=True) as response:
-                # save response to file
-                with open(image_path, "wb") as output_file:
-                    async for chunk in response.content.iter_chunked(8192):
-                        output_file.write(chunk)
-            return f"{output_dir}/{image_name}.jpeg"
-
-        except aiohttp.client_exceptions.InvalidURL as url_exception:
-            raise Exception(
-                "Inappropriate contents found in the generated images. Please try again or try another prompt.",
-            ) from url_exception
diff --git a/src/askgpt.py b/src/askgpt.py
deleted file mode 100644
index 0a9f0fc..0000000
--- a/src/askgpt.py
+++ /dev/null
@@ -1,46 +0,0 @@
-import aiohttp
-import asyncio
-import json
-
-from log import getlogger
-
-logger = getlogger()
-
-
-class askGPT:
-    def __init__(
-        self, session: aiohttp.ClientSession, headers: str
-    ) -> None:
-        self.session = session
-        self.api_endpoint = "https://api.openai.com/v1/chat/completions"
-        self.headers = headers
-
-    async def oneTimeAsk(self, prompt: str) -> str:
-        jsons = {
-            "model": "gpt-3.5-turbo",
-            "messages": [
-                {
-                    "role": "user",
-                    "content": prompt,
-                },
-            ],
-        }
-        max_try = 2
-        while max_try > 0:
-            try:
-                async with self.session.post(
-                    url=self.api_endpoint, json=jsons, headers=self.headers, timeout=120
-                ) as response:
-                    status_code = response.status
-                    if not status_code == 200:
-                        # print failed reason
-                        logger.warning(str(response.reason))
-                        max_try = max_try - 1
-                        # wait 2s
-                        await asyncio.sleep(2)
-                        continue
-
-                    resp = await response.read()
-                    return json.loads(resp)["choices"][0]["message"]["content"]
-            except Exception as e:
-                raise Exception(e)
diff --git a/src/bot.py b/src/bot.py
index 5b385f2..dcfcbef 100644
--- a/src/bot.py
+++ b/src/bot.py
@@ -4,47 +4,35 @@ import json
 import asyncio
 import re
 import os
-import aiohttp
-from askgpt import askGPT
-from revChatGPT.V3 import Chatbot as GPTChatBot
-from BingImageGen import ImageGenAsync
+from gptbot import Chatbot
 from log import getlogger
-from pandora import Pandora
-import uuid
+import httpx
 
 logger = getlogger()
 
-ENGINES = [
-    "gpt-3.5-turbo",
-    "gpt-3.5-turbo-16k",
-    "gpt-3.5-turbo-0301",
-    "gpt-3.5-turbo-0613",
-    "gpt-3.5-turbo-16k-0613",
-    "gpt-4",
-    "gpt-4-0314",
-    "gpt-4-32k",
-    "gpt-4-32k-0314",
-    "gpt-4-0613",
-    "gpt-4-32k-0613",
-]
-
 
 class Bot:
     def __init__(
         self,
         server_url: str,
         username: str,
-        access_token: Optional[str] = None,
-        login_id: Optional[str] = None,
-        password: Optional[str] = None,
+        email: str,
+        password: str,
+        port: Optional[int] = 443,
+        scheme: Optional[str] = "https",
         openai_api_key: Optional[str] = None,
-        pandora_api_endpoint: Optional[str] = None,
-        pandora_api_model: Optional[str] = None,
-        bing_auth_cookie: Optional[str] = None,
-        port: int = 443,
-        scheme: str = "https",
-        timeout: int = 30,
-        gpt_engine: str = "gpt-3.5-turbo",
+        gpt_api_endpoint: Optional[str] = None,
+        gpt_model: Optional[str] = None,
+        max_tokens: Optional[int] = None,
+        top_p: Optional[float] = None,
+        presence_penalty: Optional[float] = None,
+        frequency_penalty: Optional[float] = None,
+        reply_count: Optional[int] = None,
+        system_prompt: Optional[str] = None,
+        temperature: Optional[float] = None,
+        image_generation_endpoint: Optional[str] = None,
+        image_generation_backend: Optional[str] = None,
+        timeout: Optional[float] = 120.0,
     ) -> None:
         if server_url is None:
             raise ValueError("server url must be provided")
@@ -52,7 +40,8 @@ class Bot:
         if port is None:
             self.port = 443
         else:
-            if port < 0 or port > 65535:
+            port = int(port)
+            if port <= 0 or port > 65535:
                 raise ValueError("port must be between 0 and 65535")
             self.port = port
 
@@ -63,121 +52,82 @@ class Bot:
                 raise ValueError("scheme must be either http or https")
             self.scheme = scheme
 
-        if timeout is None:
-            self.timeout = 30
-        else:
-            self.timeout = timeout
-
-        if gpt_engine is None:
-            self.gpt_engine = "gpt-3.5-turbo"
-        else:
-            if gpt_engine not in ENGINES:
-                raise ValueError("gpt_engine must be one of {}".format(ENGINES))
-            self.gpt_engine = gpt_engine
-
-        # login relative info
-        if access_token is None and password is None:
-            raise ValueError("Either token or password must be provided")
-
-        if access_token is not None:
-            self.driver = AsyncDriver(
-                {
-                    "token": access_token,
-                    "url": server_url,
-                    "port": self.port,
-                    "request_timeout": self.timeout,
-                    "scheme": self.scheme,
-                }
-            )
-        else:
-            self.driver = AsyncDriver(
-                {
-                    "login_id": login_id,
-                    "password": password,
-                    "url": server_url,
-                    "port": self.port,
-                    "request_timeout": self.timeout,
-                    "scheme": self.scheme,
-                }
-            )
-
         # @chatgpt
         if username is None:
             raise ValueError("username must be provided")
         else:
             self.username = username
 
-        # aiohttp session
-        self.session = aiohttp.ClientSession()
+        self.openai_api_key: str = openai_api_key
+        self.gpt_api_endpoint = (
+            gpt_api_endpoint or "https://api.openai.com/v1/chat/completions"
+        )
+        self.gpt_model: str = gpt_model or "gpt-3.5-turbo"
+        self.max_tokens: int = max_tokens or 4000
+        self.top_p: float = top_p or 1.0
+        self.temperature: float = temperature or 0.8
+        self.presence_penalty: float = presence_penalty or 0.0
+        self.frequency_penalty: float = frequency_penalty or 0.0
+        self.reply_count: int = reply_count or 1
+        self.system_prompt: str = (
+            system_prompt
+            or "You are ChatGPT, \
+            a large language model trained by OpenAI. Respond conversationally"
+        )
+        self.image_generation_endpoint: str = image_generation_endpoint
+        self.image_generation_backend: str = image_generation_backend
+        self.timeout = timeout or 120.0
 
-        # initialize chatGPT class
-        self.openai_api_key = openai_api_key
-        if openai_api_key is not None:
-            # request header for !gpt command
-            self.headers = {
-                "Content-Type": "application/json",
-                "Authorization": f"Bearer {self.openai_api_key}",
+        #  httpx session
+        self.httpx_client = httpx.AsyncClient()
+
+        # initialize Chatbot object
+        self.chatbot = Chatbot(
+            aclient=self.httpx_client,
+            api_key=self.openai_api_key,
+            api_url=self.gpt_api_endpoint,
+            engine=self.gpt_model,
+            timeout=self.timeout,
+            max_tokens=self.max_tokens,
+            top_p=self.top_p,
+            presence_penalty=self.presence_penalty,
+            frequency_penalty=self.frequency_penalty,
+            reply_count=self.reply_count,
+            system_prompt=self.system_prompt,
+            temperature=self.temperature,
+        )
+
+        # login relative info
+        if email is None and password is None:
+            raise ValueError("user email and password must be provided")
+
+        self.driver = AsyncDriver(
+            {
+                "login_id": email,
+                "password": password,
+                "url": server_url,
+                "port": self.port,
+                "request_timeout": self.timeout,
+                "scheme": self.scheme,
             }
-
-            self.askgpt = askGPT(
-                self.session,
-                self.headers,
-            )
-
-            self.gptchatbot = GPTChatBot(
-                api_key=self.openai_api_key, engine=self.gpt_engine
-            )
-        else:
-            logger.warning(
-                "openai_api_key is not provided, !gpt and !chat command will not work"
-            )
-
-        # initialize pandora
-        self.pandora_api_endpoint = pandora_api_endpoint
-        if pandora_api_endpoint is not None:
-            self.pandora = Pandora(
-                api_endpoint=pandora_api_endpoint, clientSession=self.session
-            )
-            if pandora_api_model is None:
-                self.pandora_api_model = "text-davinci-002-render-sha-mobile"
-            else:
-                self.pandora_api_model = pandora_api_model
-        self.pandora_data = {}
-
-        # initialize image generator
-        self.bing_auth_cookie = bing_auth_cookie
-        if bing_auth_cookie is not None:
-            self.imagegen = ImageGenAsync(auth_cookie=self.bing_auth_cookie)
-        else:
-            logger.warning(
-                "bing_auth_cookie is not provided, !pic command will not work"
-            )
+        )
 
         # regular expression to match keyword
         self.gpt_prog = re.compile(r"^\s*!gpt\s*(.+)$")
         self.chat_prog = re.compile(r"^\s*!chat\s*(.+)$")
         self.pic_prog = re.compile(r"^\s*!pic\s*(.+)$")
         self.help_prog = re.compile(r"^\s*!help\s*.*$")
-        self.talk_prog = re.compile(r"^\s*!talk\s*(.+)$")
-        self.goon_prog = re.compile(r"^\s*!goon\s*.*$")
         self.new_prog = re.compile(r"^\s*!new\s*.*$")
 
     # close session
     async def close(self, task: asyncio.Task) -> None:
-        await self.session.close()
+        await self.session.aclose()
         self.driver.disconnect()
         task.cancel()
 
     async def login(self) -> None:
         await self.driver.login()
 
-    def pandora_init(self, user_id: str) -> None:
-        self.pandora_data[user_id] = {
-            "conversation_id": None,
-            "parent_message_id": str(uuid.uuid4()),
-            "first_time": True,
-        }
-
     async def run(self) -> None:
         await self.driver.init_websocket(self.websocket_handler)
 
@@ -191,37 +141,47 @@ class Bot:
                 raw_data = response["data"]["post"]
                 raw_data_dict = json.loads(raw_data)
                 user_id = raw_data_dict["user_id"]
+                root_id = (
+                    raw_data_dict["root_id"]
+                    if raw_data_dict["root_id"]
+                    else raw_data_dict["id"]
+                )
                 channel_id = raw_data_dict["channel_id"]
                 sender_name = response["data"]["sender_name"]
                 raw_message = raw_data_dict["message"]
 
-                if user_id not in self.pandora_data:
-                    self.pandora_init(user_id)
-
                 try:
                     asyncio.create_task(
                         self.message_callback(
-                            raw_message, channel_id, user_id, sender_name
+                            raw_message, channel_id, user_id, sender_name, root_id
                         )
                     )
                 except Exception as e:
-                    await self.send_message(channel_id, f"{e}")
+                    await self.send_message(channel_id, f"{e}", root_id)
 
     # message callback
     async def message_callback(
-        self, raw_message: str, channel_id: str, user_id: str, sender_name: str
+        self,
+        raw_message: str,
+        channel_id: str,
+        user_id: str,
+        sender_name: str,
+        root_id: str,
     ) -> None:
         # prevent command trigger loop
         if sender_name != self.username:
             message = raw_message
 
-            if self.openai_api_key is not None:
+            if (
+                self.openai_api_key is not None
+                or self.gpt_api_endpoint != "https://api.openai.com/v1/chat/completions"
+            ):
                 # !gpt command trigger handler
                 if self.gpt_prog.match(message):
                     prompt = self.gpt_prog.match(message).group(1)
                     try:
-                        response = await self.gpt(prompt)
-                        await self.send_message(channel_id, f"{response}")
+                        response = await self.chatbot.oneTimeAsk(prompt)
+                        await self.send_message(channel_id, f"{response}", root_id)
                     except Exception as e:
                         logger.error(e, exc_info=True)
                         raise Exception(e)
@@ -230,134 +190,60 @@ class Bot:
                 elif self.chat_prog.match(message):
                     prompt = self.chat_prog.match(message).group(1)
                     try:
-                        response = await self.chat(prompt)
-                        await self.send_message(channel_id, f"{response}")
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-            if self.pandora_api_endpoint is not None:
-                # !talk command trigger handler
-                if self.talk_prog.match(message):
-                    prompt = self.talk_prog.match(message).group(1)
-                    try:
-                        if self.pandora_data[user_id]["conversation_id"] is not None:
-                            data = {
-                                "prompt": prompt,
-                                "model": self.pandora_api_model,
-                                "parent_message_id": self.pandora_data[user_id][
-                                    "parent_message_id"
-                                ],
-                                "conversation_id": self.pandora_data[user_id][
-                                    "conversation_id"
-                                ],
-                                "stream": False,
-                            }
-                        else:
-                            data = {
-                                "prompt": prompt,
-                                "model": self.pandora_api_model,
-                                "parent_message_id": self.pandora_data[user_id][
-                                    "parent_message_id"
-                                ],
-                                "stream": False,
-                            }
-                        response = await self.pandora.talk(data)
-                        self.pandora_data[user_id]["conversation_id"] = response[
-                            "conversation_id"
-                        ]
-                        self.pandora_data[user_id]["parent_message_id"] = response[
-                            "message"
-                        ]["id"]
-                        content = response["message"]["content"]["parts"][0]
-                        if self.pandora_data[user_id]["first_time"]:
-                            self.pandora_data[user_id]["first_time"] = False
-                            data = {
-                                "model": self.pandora_api_model,
-                                "message_id": self.pandora_data[user_id][
-                                    "parent_message_id"
-                                ],
-                            }
-                            await self.pandora.gen_title(
-                                data, self.pandora_data[user_id]["conversation_id"]
-                            )
-
-                        await self.send_message(channel_id, f"{content}")
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-                # !goon command trigger handler
-                if (
-                    self.goon_prog.match(message)
-                    and self.pandora_data[user_id]["conversation_id"] is not None
-                ):
-                    try:
-                        data = {
-                            "model": self.pandora_api_model,
-                            "parent_message_id": self.pandora_data[user_id][
-                                "parent_message_id"
-                            ],
-                            "conversation_id": self.pandora_data[user_id][
-                                "conversation_id"
-                            ],
-                            "stream": False,
-                        }
-                        response = await self.pandora.goon(data)
-                        self.pandora_data[user_id]["conversation_id"] = response[
-                            "conversation_id"
-                        ]
-                        self.pandora_data[user_id]["parent_message_id"] = response[
-                            "message"
-                        ]["id"]
-                        content = response["message"]["content"]["parts"][0]
-                        await self.send_message(channel_id, f"{content}")
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
-
-                # !new command trigger handler
-                if self.new_prog.match(message):
-                    self.pandora_init(user_id)
-                    try:
-                        await self.send_message(
-                            channel_id,
-                            "New conversation created, " +
-                            "please use !talk to start chatting!",
+                        response = await self.chatbot.ask_async(
+                            prompt=prompt, convo_id=user_id
                         )
-                    except Exception:
-                        pass
-
-            if self.bing_auth_cookie is not None:
-                # !pic command trigger handler
-                if self.pic_prog.match(message):
-                    prompt = self.pic_prog.match(message).group(1)
-                    # generate image
-                    try:
-                        links = await self.imagegen.get_images(prompt)
-                        image_path = await self.imagegen.save_images(links, "images")
+                        await self.send_message(channel_id, f"{response}", root_id)
                     except Exception as e:
                         logger.error(e, exc_info=True)
                         raise Exception(e)
 
-                    # send image
-                    try:
-                        await self.send_file(channel_id, prompt, image_path)
-                    except Exception as e:
-                        logger.error(e, exc_info=True)
-                        raise Exception(e)
+            # !new command trigger handler
+            if self.new_prog.match(message):
+                self.chatbot.reset(convo_id=user_id)
+                try:
+                    await self.send_message(
+                        channel_id,
+                        "New conversation created, "
+                        + "please use !chat to start chatting!",
+                    )
+                except Exception as e:
+                    logger.error(e, exc_info=True)
+                    raise Exception(e)
+
+            # !pic command trigger handler
+            if self.pic_prog.match(message):
+                prompt = self.pic_prog.match(message).group(1)
+                # generate image
+                try:
+                    links = await self.imagegen.get_images(prompt)
+                    image_path = await self.imagegen.save_images(links, "images")
+                except Exception as e:
+                    logger.error(e, exc_info=True)
+                    raise Exception(e)
+
+                # send image
+                try:
+                    await self.send_file(channel_id, prompt, image_path)
+                except Exception as e:
+                    logger.error(e, exc_info=True)
+                    raise Exception(e)
 
             # !help command trigger handler
             if self.help_prog.match(message):
                 try:
-                    await self.send_message(channel_id, self.help())
+                    await self.send_message(channel_id, self.help(), root_id)
                 except Exception as e:
                     logger.error(e, exc_info=True)
 
     # send message to room
-    async def send_message(self, channel_id: str, message: str) -> None:
+    async def send_message(self, channel_id: str, message: str, root_id: str) -> None:
         await self.driver.posts.create_post(
-            options={"channel_id": channel_id, "message": message}
+            options={
+                "channel_id": channel_id,
+                "message": message,
+                "root_id": root_id,
+            }
         )
 
     # send file to room
diff --git a/src/gptbot.py b/src/gptbot.py
new file mode 100644
index 0000000..454e0a1
--- /dev/null
+++ b/src/gptbot.py
@@ -0,0 +1,296 @@
+"""
+Code derived from https://github.com/acheong08/ChatGPT/blob/main/src/revChatGPT/V3.py
+A simple wrapper for the official ChatGPT API
+"""
+import json
+from typing import AsyncGenerator
+from tenacity import retry, wait_random_exponential, stop_after_attempt
+import httpx
+import tiktoken
+
+
+ENGINES = [
+    "gpt-3.5-turbo",
+    "gpt-3.5-turbo-16k",
+    "gpt-3.5-turbo-0613",
+    "gpt-3.5-turbo-16k-0613",
+    "gpt-4",
+    "gpt-4-32k",
+    "gpt-4-0613",
+    "gpt-4-32k-0613",
+]
+
+
+class Chatbot:
+    """
+    Official ChatGPT API
+    """
+
+    def __init__(
+        self,
+        aclient: httpx.AsyncClient,
+        api_key: str,
+        api_url: str = None,
+        engine: str = None,
+        timeout: float = None,
+        max_tokens: int = None,
+        temperature: float = 0.8,
+        top_p: float = 1.0,
+        presence_penalty: float = 0.0,
+        frequency_penalty: float = 0.0,
+        reply_count: int = 1,
+        truncate_limit: int = None,
+        system_prompt: str = None,
+    ) -> None:
+        """
+        Initialize Chatbot with API key (from https://platform.openai.com/account/api-keys)
+        """
+        self.engine: str = engine or "gpt-3.5-turbo"
+        self.api_key: str = api_key
+        self.api_url: str = api_url or "https://api.openai.com/v1/chat/completions"
+        self.system_prompt: str = (
+            system_prompt
+            or "You are ChatGPT, \
+            a large language model trained by OpenAI. Respond conversationally"
+        )
+        self.max_tokens: int = max_tokens or (
+            31000
+            if "gpt-4-32k" in engine
+            else 7000
+            if "gpt-4" in engine
+            else 15000
+            if "gpt-3.5-turbo-16k" in engine
+            else 4000
+        )
+        self.truncate_limit: int = truncate_limit or (
+            30500
+            if "gpt-4-32k" in engine
+            else 6500
+            if "gpt-4" in engine
+            else 14500
+            if "gpt-3.5-turbo-16k" in engine
+            else 3500
+        )
+        self.temperature: float = temperature
+        self.top_p: float = top_p
+        self.presence_penalty: float = presence_penalty
+        self.frequency_penalty: float = frequency_penalty
+        self.reply_count: int = reply_count
+        self.timeout: float = timeout
+
+        self.aclient = aclient
+
+        self.conversation: dict[str, list[dict]] = {
+            "default": [
+                {
+                    "role": "system",
+                    "content": system_prompt,
+                },
+            ],
+        }
+
+        if self.get_token_count("default") > self.max_tokens:
+            raise Exception("System prompt is too long")
+
+    def add_to_conversation(
+        self,
+        message: str,
+        role: str,
+        convo_id: str = "default",
+    ) -> None:
+        """
+        Add a message to the conversation
+        """
+        self.conversation[convo_id].append({"role": role, "content": message})
+
+    def __truncate_conversation(self, convo_id: str = "default") -> None:
+        """
+        Truncate the conversation
+        """
+        while True:
+            if (
+                self.get_token_count(convo_id) > self.truncate_limit
+                and len(self.conversation[convo_id]) > 1
+            ):
+                # Don't remove the first message
+                self.conversation[convo_id].pop(1)
+            else:
+                break
+
+    # https://github.com/openai/openai-cookbook/blob/main/examples/How_to_count_tokens_with_tiktoken.ipynb
+    def get_token_count(self, convo_id: str = "default") -> int:
+        """
+        Get token count
+        """
+        if self.engine not in ENGINES:
+            raise NotImplementedError(
+                f"Engine {self.engine} is not supported. Select from {ENGINES}",
+            )
+        tiktoken.model.MODEL_TO_ENCODING["gpt-4"] = "cl100k_base"
+
+        encoding = tiktoken.encoding_for_model(self.engine)
+
+        num_tokens = 0
+        for message in self.conversation[convo_id]:
+            # every message follows <im_start>{role/name}\n{content}<im_end>\n
+            num_tokens += 5
+            for key, value in message.items():
+                if value:
+                    num_tokens += len(encoding.encode(value))
+                if key == "name":  # if there's a name, the role is omitted
+                    num_tokens += 5  # role is always required and always 1 token
+        num_tokens += 5  # every reply is primed with <im_start>assistant
+        return num_tokens
+
+    def get_max_tokens(self, convo_id: str) -> int:
+        """
+        Get max tokens
+        """
+        return self.max_tokens - self.get_token_count(convo_id)
+
+    async def ask_stream_async(
+        self,
+        prompt: str,
+        role: str = "user",
+        convo_id: str = "default",
+        model: str = None,
+        pass_history: bool = True,
+        **kwargs,
+    ) -> AsyncGenerator[str, None]:
+        """
+        Ask a question
+        """
+        # Make conversation if it doesn't exist
+        if convo_id not in self.conversation:
+            self.reset(convo_id=convo_id, system_prompt=self.system_prompt)
+        self.add_to_conversation(prompt, "user", convo_id=convo_id)
+        self.__truncate_conversation(convo_id=convo_id)
+        # Get response
+        async with self.aclient.stream(
+            "post",
+            self.api_url,
+            headers={"Authorization": f"Bearer {kwargs.get('api_key', self.api_key)}"},
+            json={
+                "model": model or self.engine,
+                "messages": self.conversation[convo_id] if pass_history else [prompt],
+                "stream": True,
+                # kwargs
+                "temperature": kwargs.get("temperature", self.temperature),
+                "top_p": kwargs.get("top_p", self.top_p),
+                "presence_penalty": kwargs.get(
+                    "presence_penalty",
+                    self.presence_penalty,
+                ),
+                "frequency_penalty": kwargs.get(
+                    "frequency_penalty",
+                    self.frequency_penalty,
+                ),
+                "n": kwargs.get("n", self.reply_count),
+                "user": role,
+                "max_tokens": min(
+                    self.get_max_tokens(convo_id=convo_id),
+                    kwargs.get("max_tokens", self.max_tokens),
+                ),
+            },
+            timeout=kwargs.get("timeout", self.timeout),
+        ) as response:
+            if response.status_code != 200:
+                await response.aread()
+                raise Exception(
+                    f"{response.status_code} {response.reason_phrase} {response.text}",
+                )
+
+            response_role: str = ""
+            full_response: str = ""
+            async for line in response.aiter_lines():
+                line = line.strip()
+                if not line:
+                    continue
+                # Remove "data: "
+                line = line[6:]
+                if line == "[DONE]":
+                    break
+                resp: dict = json.loads(line)
+                if "error" in resp:
+                    raise Exception(f"{resp['error']}")
+                choices = resp.get("choices")
+                if not choices:
+                    continue
+                delta: dict[str, str] = choices[0].get("delta")
+                if not delta:
+                    continue
+                if "role" in delta:
+                    response_role = delta["role"]
+                if "content" in delta:
+                    content: str = delta["content"]
+                    full_response += content
+                    yield content
+        self.add_to_conversation(full_response, response_role, convo_id=convo_id)
+
+    async def ask_async(
+        self,
+        prompt: str,
+        role: str = "user",
+        convo_id: str = "default",
+        model: str = None,
+        pass_history: bool = True,
+        **kwargs,
+    ) -> str:
+        """
+        Non-streaming ask
+        """
+        response = self.ask_stream_async(
+            prompt=prompt,
+            role=role,
+            convo_id=convo_id,
+            model=model,
+            pass_history=pass_history,
+            **kwargs,
+        )
+        full_response: str = "".join([r async for r in response])
+        return full_response
+
+    def reset(self, convo_id: str = "default", system_prompt: str = None) -> None:
+        """
+        Reset the conversation
+        """
+        self.conversation[convo_id] = [
+            {"role": "system", "content": system_prompt or self.system_prompt},
+        ]
+
+    @retry(wait=wait_random_exponential(min=2, max=5), stop=stop_after_attempt(3))
+    async def oneTimeAsk(
+        self,
+        prompt: str,
+        role: str = "user",
+        model: str = None,
+        **kwargs,
+    ) -> str:
+        response = await self.aclient.post(
+            url=self.api_url,
+            json={
+                "model": model or self.engine,
+                "messages": [
+                    {
+                        "role": role,
+                        "content": prompt,
+                    }
+                ],
+                # kwargs
+                "temperature": kwargs.get("temperature", self.temperature),
+                "top_p": kwargs.get("top_p", self.top_p),
+                "presence_penalty": kwargs.get(
+                    "presence_penalty",
+                    self.presence_penalty,
+                ),
+                "frequency_penalty": kwargs.get(
+                    "frequency_penalty",
+                    self.frequency_penalty,
+                ),
+                "user": role,
+            },
+            headers={"Authorization": f"Bearer {kwargs.get('api_key', self.api_key)}"},
+            timeout=kwargs.get("timeout", self.timeout),
+        )
+        resp = response.json()
+        return resp["choices"][0]["message"]["content"]
diff --git a/src/imagegen.py b/src/imagegen.py
new file mode 100644
index 0000000..fb54f14
--- /dev/null
+++ b/src/imagegen.py
@@ -0,0 +1,69 @@
+import httpx
+from pathlib import Path
+import uuid
+import base64
+import io
+from PIL import Image
+
+
+async def get_images(
+    aclient: httpx.AsyncClient, url: str, prompt: str, backend_type: str, **kwargs
+) -> list[str]:
+    timeout = kwargs.get("timeout", 120.0)
+    if backend_type == "openai":
+        resp = await aclient.post(
+            url,
+            headers={
+                "Content-Type": "application/json",
+                "Authorization": "Bearer " + kwargs.get("api_key"),
+            },
+            json={
+                "prompt": prompt,
+                "n": kwargs.get("n", 1),
+                "size": kwargs.get("size", "256x256"),
+                "response_format": "b64_json",
+            },
+            timeout=timeout,
+        )
+        if resp.status_code == 200:
+            b64_datas = []
+            for data in resp.json()["data"]:
+                b64_datas.append(data["b64_json"])
+            return b64_datas
+        else:
+            raise Exception(
+                f"{resp.status_code} {resp.reason_phrase} {resp.text}",
+            )
+    elif backend_type == "sdwui":
+        resp = await aclient.post(
+            url,
+            headers={
+                "Content-Type": "application/json",
+            },
+            json={
+                "prompt": prompt,
+                "sampler_name": kwargs.get("sampler_name", "Euler a"),
+                "batch_size": kwargs.get("n", 1),
+                "steps": kwargs.get("steps", 20),
+                "width": 256 if "256" in kwargs.get("size") else 512,
+                "height": 256 if "256" in kwargs.get("size") else 512,
+            },
+            timeout=timeout,
+        )
+        if resp.status_code == 200:
+            b64_datas = resp.json()["images"]
+            return b64_datas
+        else:
+            raise Exception(
+                f"{resp.status_code} {resp.reason_phrase} {resp.text}",
+            )
+
+
+def save_images(b64_datas: list[str], path: Path, **kwargs) -> list[str]:
+    images = []
+    for b64_data in b64_datas:
+        image_path = path / (str(uuid.uuid4()) + ".jpeg")
+        img = Image.open(io.BytesIO(base64.decodebytes(bytes(b64_data, "utf-8"))))
+        img.save(image_path)
+        images.append(image_path)
+    return images
diff --git a/src/main.py b/src/main.py
index ed624c7..0934842 100644
--- a/src/main.py
+++ b/src/main.py
@@ -2,6 +2,7 @@ import signal
 from bot import Bot
 import json
 import os
+import sys
 import asyncio
 from pathlib import Path
 from log import getlogger
@@ -13,39 +14,55 @@ async def main():
     config_path = Path(os.path.dirname(__file__)).parent / "config.json"
     if os.path.isfile(config_path):
         fp = open("config.json", "r", encoding="utf-8")
-        config = json.load(fp)
+        try:
+            config = json.load(fp)
+        except Exception as e:
+            logger.error(e, exc_info=True)
+            sys.exit(1)
 
         mattermost_bot = Bot(
             server_url=config.get("server_url"),
-            access_token=config.get("access_token"),
-            login_id=config.get("login_id"),
+            email=config.get("email"),
             password=config.get("password"),
             username=config.get("username"),
-            openai_api_key=config.get("openai_api_key"),
-            bing_auth_cookie=config.get("bing_auth_cookie"),
-            pandora_api_endpoint=config.get("pandora_api_endpoint"),
-            pandora_api_model=config.get("pandora_api_model"),
             port=config.get("port"),
             scheme=config.get("scheme"),
+            openai_api_key=config.get("openai_api_key"),
+            gpt_api_endpoint=config.get("gpt_api_endpoint"),
+            gpt_model=config.get("gpt_model"),
+            max_tokens=config.get("max_tokens"),
+            top_p=config.get("top_p"),
+            presence_penalty=config.get("presence_penalty"),
+            frequency_penalty=config.get("frequency_penalty"),
+            reply_count=config.get("reply_count"),
+            system_prompt=config.get("system_prompt"),
+            temperature=config.get("temperature"),
+            image_generation_endpoint=config.get("image_generation_endpoint"),
+            image_generation_backend=config.get("image_generation_backend"),
             timeout=config.get("timeout"),
-            gpt_engine=config.get("gpt_engine"),
         )
 
     else:
         mattermost_bot = Bot(
             server_url=os.environ.get("SERVER_URL"),
-            access_token=os.environ.get("ACCESS_TOKEN"),
-            login_id=os.environ.get("LOGIN_ID"),
+            email=os.environ.get("EMAIL"),
             password=os.environ.get("PASSWORD"),
             username=os.environ.get("USERNAME"),
-            openai_api_key=os.environ.get("OPENAI_API_KEY"),
-            bing_auth_cookie=os.environ.get("BING_AUTH_COOKIE"),
-            pandora_api_endpoint=os.environ.get("PANDORA_API_ENDPOINT"),
-            pandora_api_model=os.environ.get("PANDORA_API_MODEL"),
             port=os.environ.get("PORT"),
             scheme=os.environ.get("SCHEME"),
+            openai_api_key=os.environ.get("OPENAI_API_KEY"),
+            gpt_api_endpoint=os.environ.get("GPT_API_ENDPOINT"),
+            gpt_model=os.environ.get("GPT_MODEL"),
+            max_tokens=os.environ.get("MAX_TOKENS"),
+            top_p=os.environ.get("TOP_P"),
+            presence_penalty=os.environ.get("PRESENCE_PENALTY"),
+            frequency_penalty=os.environ.get("FREQUENCY_PENALTY"),
+            reply_count=os.environ.get("REPLY_COUNT"),
+            system_prompt=os.environ.get("SYSTEM_PROMPT"),
+            temperature=os.environ.get("TEMPERATURE"),
+            image_generation_endpoint=os.environ.get("IMAGE_GENERATION_ENDPOINT"),
+            image_generation_backend=os.environ.get("IMAGE_GENERATION_BACKEND"),
             timeout=os.environ.get("TIMEOUT"),
-            gpt_engine=os.environ.get("GPT_ENGINE"),
         )
 
     await mattermost_bot.login()
diff --git a/src/pandora.py b/src/pandora.py
deleted file mode 100644
index c6b5997..0000000
--- a/src/pandora.py
+++ /dev/null
@@ -1,106 +0,0 @@
-# https://github.com/pengzhile/pandora/blob/master/doc/HTTP-API.md
-import uuid
-import aiohttp
-import asyncio
-
-
-class Pandora:
-    def __init__(self, api_endpoint: str, clientSession: aiohttp.ClientSession) -> None:
-        self.api_endpoint = api_endpoint.rstrip("/")
-        self.session = clientSession
-
-    async def __aenter__(self):
-        return self
-
-    async def __aexit__(self, exc_type, exc_val, exc_tb):
-        await self.session.close()
-
-    async def gen_title(self, data: dict, conversation_id: str) -> None:
-        """
-        data = {
-            "model": "",
-            "message_id": "",
-        }
-        :param data: dict
-        :param conversation_id: str
-        :return: None
-        """
-        api_endpoint = (
-            self.api_endpoint + f"/api/conversation/gen_title/{conversation_id}"
-        )
-        async with self.session.post(api_endpoint, json=data) as resp:
-            return await resp.json()
-
-    async def talk(self, data: dict) -> None:
-        api_endpoint = self.api_endpoint + "/api/conversation/talk"
-        """
-        data = {
-            "prompt": "",
-            "model": "",
-            "parent_message_id": "",
-            "conversation_id": "", # ignore at the first time
-            "stream": True,
-        }
-        :param data: dict
-        :return: None
-        """
-        data["message_id"] = str(uuid.uuid4())
-        async with self.session.post(api_endpoint, json=data) as resp:
-            return await resp.json()
-
-    async def goon(self, data: dict) -> None:
-        """
-        data = {
-            "model": "",
-            "parent_message_id": "",
-            "conversation_id": "",
-            "stream": True,
-        }
-        """
-        api_endpoint = self.api_endpoint + "/api/conversation/goon"
-        async with self.session.post(api_endpoint, json=data) as resp:
-            return await resp.json()
-
-
-async def test():
-    model = "text-davinci-002-render-sha-mobile"
-    api_endpoint = "http://127.0.0.1:8008"
-    async with aiohttp.ClientSession() as session:
-        client = Pandora(api_endpoint, session)
-    conversation_id = None
-    parent_message_id = str(uuid.uuid4())
-    first_time = True
-    async with client:
-        while True:
-            prompt = input("BobMaster: ")
-            if conversation_id:
-                data = {
-                    "prompt": prompt,
-                    "model": model,
-                    "parent_message_id": parent_message_id,
-                    "conversation_id": conversation_id,
-                    "stream": False,
-                }
-            else:
-                data = {
-                    "prompt": prompt,
-                    "model": model,
-                    "parent_message_id": parent_message_id,
-                    "stream": False,
-                }
-            response = await client.talk(data)
-            conversation_id = response["conversation_id"]
-            parent_message_id = response["message"]["id"]
-            content = response["message"]["content"]["parts"][0]
-            print("ChatGPT: " + content + "\n")
-            if first_time:
-                first_time = False
-                data = {
-                    "model": model,
-                    "message_id": parent_message_id,
-                }
-                response = await client.gen_title(data, conversation_id)
-
-
-if __name__ == "__main__":
-    asyncio.run(test())

From a2027590c8dc85dbd5961450bfc5c94ed33ec3c3 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Mon, 18 Sep 2023 13:50:21 +0800
Subject: [PATCH 04/15] remove pylint check

---
 .github/workflows/pylint.yml | 25 -------------------------
 1 file changed, 25 deletions(-)
 delete mode 100644 .github/workflows/pylint.yml

diff --git a/.github/workflows/pylint.yml b/.github/workflows/pylint.yml
deleted file mode 100644
index ff31b8c..0000000
--- a/.github/workflows/pylint.yml
+++ /dev/null
@@ -1,25 +0,0 @@
-name: Pylint
-
-on: [push, pull_request]
-
-jobs:
-  build:
-    runs-on: ubuntu-latest
-    strategy:
-      matrix:
-        python-version: ["3.10", "3.11"]
-    steps:
-    - uses: actions/checkout@v3
-    - name: Set up Python ${{ matrix.python-version }}
-      uses: actions/setup-python@v4
-      with:
-        python-version: ${{ matrix.python-version }}
-        cache: 'pip'
-    - name: Install dependencies
-      run: |
-        pip install -U pip setuptools wheel
-        pip install -r requirements.txt
-        pip install pylint
-    - name: Analysing the code with pylint
-      run: |
-        pylint $(git ls-files '*.py') --errors-only

From d017323c256021a150b71042e3218c594aabe017 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Mon, 18 Sep 2023 13:56:07 +0800
Subject: [PATCH 05/15] Fix

---
 CHANGELOG.md |  1 +
 src/bot.py   | 15 +++------------
 2 files changed, 4 insertions(+), 12 deletions(-)

diff --git a/CHANGELOG.md b/CHANGELOG.md
index d66b2ed..cf0eaef 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -4,6 +4,7 @@
 - remove pandora
 - refactor chat and image genderation backend
 - reply in thread by default
+- introduce pre-commit hooks
 
 ## v1.0.4
 
diff --git a/src/bot.py b/src/bot.py
index dcfcbef..19e4dee 100644
--- a/src/bot.py
+++ b/src/bot.py
@@ -121,7 +121,7 @@ class Bot:
 
     # close session
     async def close(self, task: asyncio.Task) -> None:
-        await self.session.aclose()
+        await self.httpx_client.aclose()
         self.driver.disconnect()
         task.cancel()
 
@@ -206,6 +206,7 @@ class Bot:
                         channel_id,
                         "New conversation created, "
                         + "please use !chat to start chatting!",
+                        root_id,
                     )
                 except Exception as e:
                     logger.error(e, exc_info=True)
@@ -274,22 +275,12 @@ class Bot:
             logger.error(e, exc_info=True)
             raise Exception(e)
 
-    # !gpt command function
-    async def gpt(self, prompt: str) -> str:
-        return await self.askgpt.oneTimeAsk(prompt)
-
-    # !chat command function
-    async def chat(self, prompt: str) -> str:
-        return await self.gptchatbot.ask_async(prompt)
-
     # !help command function
     def help(self) -> str:
         help_info = (
             "!gpt [content], generate response without context conversation\n"
             + "!chat [content], chat with context conversation\n"
-            + "!pic [prompt], Image generation by Microsoft Bing\n"
-            + "!talk [content], talk using chatgpt web\n"
-            + "!goon, continue the incomplete conversation\n"
+            + "!pic [prompt], Image generation with DALL·E or LocalAI or stable-diffusion-webui\n"  # noqa: E501
             + "!new, start a new conversation\n"
             + "!help, help message"
         )

From 06ccd8c61c0caf9d0c9314d11c7a725dea96a647 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Mon, 18 Sep 2023 13:58:53 +0800
Subject: [PATCH 06/15] Update ruff-pre-commit

---
 .pre-commit-config.yaml | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
index d811573..cc25d41 100644
--- a/.pre-commit-config.yaml
+++ b/.pre-commit-config.yaml
@@ -10,7 +10,7 @@ repos:
     hooks:
       - id: black
   - repo: https://github.com/astral-sh/ruff-pre-commit
-    rev: v0.0.289
+    rev: v0.0.290
     hooks:
       - id: ruff
         args: [--fix, --exit-non-zero-on-fix]

From 0b20e0ac1ac503ed7c4504d09d97bd8f58c97f4e Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Mon, 18 Sep 2023 15:19:27 +0800
Subject: [PATCH 07/15] feat: refactor image genderation backend

---
 README.md        | 14 ++++-------
 requirements.txt |  1 -
 src/bot.py       | 63 ++++++++++++++++++++++++++++++++++--------------
 src/imagegen.py  |  2 +-
 4 files changed, 51 insertions(+), 29 deletions(-)

diff --git a/README.md b/README.md
index f5a4cc8..3b83913 100644
--- a/README.md
+++ b/README.md
@@ -4,8 +4,8 @@ This is a simple Mattermost Bot that uses OpenAI's GPT API(or self-host models)
 
 ## Feature
 
-1. Support Openai ChatGPT
-3. ChatGPT web ([pandora](https://github.com/pengzhile/pandora))
+1. Support official openai api and self host models([LocalAI](https://localai.io/model-compatibility/))
+2. Image Generation with [DALL·E](https://platform.openai.com/docs/api-reference/images/create) or [LocalAI](https://localai.io/features/image-generation/) or [stable-diffusion-webui](https://github.com/AUTOMATIC1111/stable-diffusion-webui/wiki/API)
 ## Installation and Setup
 
 See https://github.com/hibobmaster/mattermost_bot/wiki
@@ -21,18 +21,14 @@ docker compose up -d
 - `!help` help message
 - `!gpt + [prompt]` generate a one time response from chatGPT
 - `!chat + [prompt]` chat using official chatGPT api with context conversation
-- `!pic + [prompt]` generate an image from Bing Image Creator
+- `!pic + [prompt]` Image generation with DALL·E or LocalAI or stable-diffusion-webui
 
-The following commands need pandora http api: https://github.com/pengzhile/pandora/blob/master/doc/wiki_en.md#http-restful-api
-- `!talk + [prompt]` chat using chatGPT web with context conversation
-- `!goon` ask chatGPT to complete the missing part from previous conversation
 - `!new` start a new converstaion
 
 ## Demo
 Remove support for Bing AI, Google Bard due to technical problems.
-![demo1](https://i.imgur.com/XRAQB4B.jpg)
-![demo2](https://i.imgur.com/if72kyH.jpg)
-![demo3](https://i.imgur.com/GHczfkv.jpg)
+![gpt command](https://imgur.com/vdT83Ln.jpg)
+![image generation](https://i.imgur.com/GHczfkv.jpg)
 
 ## Thanks
 <a href="https://jb.gg/OpenSourceSupport" target="_blank">
diff --git a/requirements.txt b/requirements.txt
index c7f6d1f..6314c78 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -1,4 +1,3 @@
-aiofiles
 httpx
 Pillow
 tiktoken
diff --git a/src/bot.py b/src/bot.py
index 19e4dee..988fed1 100644
--- a/src/bot.py
+++ b/src/bot.py
@@ -4,9 +4,11 @@ import json
 import asyncio
 import re
 import os
+from pathlib import Path
 from gptbot import Chatbot
 from log import getlogger
 import httpx
+import imagegen
 
 logger = getlogger()
 
@@ -78,6 +80,11 @@ class Bot:
         self.image_generation_backend: str = image_generation_backend
         self.timeout = timeout or 120.0
 
+        self.base_path = Path(os.path.dirname(__file__)).parent
+
+        if not os.path.exists(self.base_path / "images"):
+            os.mkdir(self.base_path / "images")
+
         #  httpx session
         self.httpx_client = httpx.AsyncClient()
 
@@ -213,22 +220,38 @@ class Bot:
                     raise Exception(e)
 
             # !pic command trigger handler
-            if self.pic_prog.match(message):
-                prompt = self.pic_prog.match(message).group(1)
-                # generate image
-                try:
-                    links = await self.imagegen.get_images(prompt)
-                    image_path = await self.imagegen.save_images(links, "images")
-                except Exception as e:
-                    logger.error(e, exc_info=True)
-                    raise Exception(e)
-
-                # send image
-                try:
-                    await self.send_file(channel_id, prompt, image_path)
-                except Exception as e:
-                    logger.error(e, exc_info=True)
-                    raise Exception(e)
+            if self.image_generation_endpoint and self.image_generation_backend:
+                if self.pic_prog.match(message):
+                    prompt = self.pic_prog.match(message).group(1)
+                    # generate image
+                    try:
+                        # generate image
+                        b64_datas = await imagegen.get_images(
+                            self.httpx_client,
+                            self.image_generation_endpoint,
+                            prompt,
+                            self.image_generation_backend,
+                            timeount=self.timeout,
+                            api_key=self.openai_api_key,
+                            n=1,
+                            size="256x256",
+                        )
+                        image_path_list = await asyncio.to_thread(
+                            imagegen.save_images,
+                            b64_datas,
+                            self.base_path / "images",
+                        )
+                        # send image
+                        for image_path in image_path_list:
+                            await self.send_file(
+                                channel_id,
+                                f"{prompt}",
+                                image_path,
+                                root_id,
+                            )
+                    except Exception as e:
+                        logger.error(e, exc_info=True)
+                        raise Exception(e)
 
             # !help command trigger handler
             if self.help_prog.match(message):
@@ -248,7 +271,9 @@ class Bot:
         )
 
     # send file to room
-    async def send_file(self, channel_id: str, message: str, filepath: str) -> None:
+    async def send_file(
+        self, channel_id: str, message: str, filepath: str, root_id: str
+    ) -> None:
         filename = os.path.split(filepath)[-1]
         try:
             file_id = await self.driver.files.upload_file(
@@ -256,7 +281,8 @@ class Bot:
                 files={
                     "files": (filename, open(filepath, "rb")),
                 },
-            )["file_infos"][0]["id"]
+            )
+            file_id = file_id["file_infos"][0]["id"]
         except Exception as e:
             logger.error(e, exc_info=True)
             raise Exception(e)
@@ -267,6 +293,7 @@ class Bot:
                     "channel_id": channel_id,
                     "message": message,
                     "file_ids": [file_id],
+                    "root_id": root_id,
                 }
             )
             # remove image after posting
diff --git a/src/imagegen.py b/src/imagegen.py
index fb54f14..2214eac 100644
--- a/src/imagegen.py
+++ b/src/imagegen.py
@@ -15,7 +15,7 @@ async def get_images(
             url,
             headers={
                 "Content-Type": "application/json",
-                "Authorization": "Bearer " + kwargs.get("api_key"),
+                "Authorization": f"Bearer {kwargs.get('api_key')}",
             },
             json={
                 "prompt": prompt,

From 3bef3e1a51cbca1bd6c2a674c09d8f7b666846f9 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Wed, 20 Sep 2023 09:49:07 +0800
Subject: [PATCH 08/15] feat: support sending typing state

---
 src/bot.py | 27 ++++++++++++++++++++++++++-
 1 file changed, 26 insertions(+), 1 deletion(-)

diff --git a/src/bot.py b/src/bot.py
index 988fed1..e801715 100644
--- a/src/bot.py
+++ b/src/bot.py
@@ -80,6 +80,8 @@ class Bot:
         self.image_generation_backend: str = image_generation_backend
         self.timeout = timeout or 120.0
 
+        self.bot_id = None
+
         self.base_path = Path(os.path.dirname(__file__)).parent
 
         if not os.path.exists(self.base_path / "images"):
@@ -134,6 +136,9 @@ class Bot:
 
     async def login(self) -> None:
         await self.driver.login()
+        # get user id
+        resp = await self.driver.users.get_user(user_id="me")
+        self.bot_id = resp["id"]
 
     async def run(self) -> None:
         await self.driver.init_websocket(self.websocket_handler)
@@ -187,6 +192,13 @@ class Bot:
                 if self.gpt_prog.match(message):
                     prompt = self.gpt_prog.match(message).group(1)
                     try:
+                        # sending typing state
+                        await self.driver.users.publish_user_typing(
+                            self.bot_id,
+                            options={
+                                "channel_id": channel_id,
+                            },
+                        )
                         response = await self.chatbot.oneTimeAsk(prompt)
                         await self.send_message(channel_id, f"{response}", root_id)
                     except Exception as e:
@@ -197,6 +209,13 @@ class Bot:
                 elif self.chat_prog.match(message):
                     prompt = self.chat_prog.match(message).group(1)
                     try:
+                        # sending typing state
+                        await self.driver.users.publish_user_typing(
+                            self.bot_id,
+                            options={
+                                "channel_id": channel_id,
+                            },
+                        )
                         response = await self.chatbot.ask_async(
                             prompt=prompt, convo_id=user_id
                         )
@@ -225,7 +244,13 @@ class Bot:
                     prompt = self.pic_prog.match(message).group(1)
                     # generate image
                     try:
-                        # generate image
+                        # sending typing state
+                        await self.driver.users.publish_user_typing(
+                            self.bot_id,
+                            options={
+                                "channel_id": channel_id,
+                            },
+                        )
                         b64_datas = await imagegen.get_images(
                             self.httpx_client,
                             self.image_generation_endpoint,

From 57c054b5d2bf2e06ab0f662b4b21c2e3da10757e Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Wed, 20 Sep 2023 09:54:28 +0800
Subject: [PATCH 09/15] v1.2.0

---
 CHANGELOG.md | 3 +++
 1 file changed, 3 insertions(+)

diff --git a/CHANGELOG.md b/CHANGELOG.md
index cf0eaef..b2f9d93 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,5 +1,8 @@
 # Changelog
 
+## v1.2.0
+- support sending typing state
+
 ## v1.1.0
 - remove pandora
 - refactor chat and image genderation backend

From 4dd62d39405faf2366d92a667710aafde0e19be2 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Wed, 20 Sep 2023 10:08:13 +0800
Subject: [PATCH 10/15] fix: improve dockerignore

---
 .dockerignore | 3 +++
 1 file changed, 3 insertions(+)

diff --git a/.dockerignore b/.dockerignore
index fa0b4b7..8069a90 100644
--- a/.dockerignore
+++ b/.dockerignore
@@ -24,3 +24,6 @@ src/__pycache__
 settings.js
 mattermost-server
 tests
+full-config.json.example
+config.json.example
+.full-env.example

From e0aff199053816a0e6f620d6624601a3d498d142 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sat, 23 Dec 2023 22:30:11 +0800
Subject: [PATCH 11/15] Fix localai v2.0+ image generation Support specific
 output image format(jpeg, png) and size

---
 .full-env.example        |  4 ++-
 full-config.json.example |  4 ++-
 requirements.txt         |  1 +
 src/bot.py               | 49 ++++++++++++++++++++++++++------
 src/imagegen.py          | 60 ++++++++++++++++++++++++++++++++--------
 src/main.py              | 20 ++++++++------
 6 files changed, 107 insertions(+), 31 deletions(-)

diff --git a/.full-env.example b/.full-env.example
index 53c3020..0176496 100644
--- a/.full-env.example
+++ b/.full-env.example
@@ -15,5 +15,7 @@ REPLY_COUNT=1
 SYSTEM_PROMPT="You are ChatGPT,  a large language model trained by OpenAI. Respond conversationally"
 TEMPERATURE=0.8
 IMAGE_GENERATION_ENDPOINT="http://127.0.0.1:7860/sdapi/v1/txt2img"
-IMAGE_GENERATION_BACKEND="sdwui" # openai or sdwui
+IMAGE_GENERATION_BACKEND="sdwui" # openai or sdwui or localai
+IMAGE_GENERATION_SIZE="512x512"
+IMAGE_FORMAT="jpeg"
 TIMEOUT=120.0
diff --git a/full-config.json.example b/full-config.json.example
index 5215c90..d964c3a 100644
--- a/full-config.json.example
+++ b/full-config.json.example
@@ -16,6 +16,8 @@
     "temperature": 0.8,
     "system_prompt": "You are ChatGPT,  a large language model trained by OpenAI. Respond conversationally",
     "image_generation_endpoint": "http://localai:8080/v1/images/generations",
-    "image_generation_backend": "openai",
+    "image_generation_backend": "localai",
+    "image_generation_size": "512x512",
+    "image_format": "webp",
     "timeout": 120.0
 }
diff --git a/requirements.txt b/requirements.txt
index 6314c78..1a4b9d9 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -2,4 +2,5 @@ httpx
 Pillow
 tiktoken
 tenacity
+aiofiles
 mattermostdriver @ git+https://github.com/hibobmaster/python-mattermost-driver
diff --git a/src/bot.py b/src/bot.py
index e801715..5c1ca59 100644
--- a/src/bot.py
+++ b/src/bot.py
@@ -1,3 +1,5 @@
+import sys
+import aiofiles.os
 from mattermostdriver import AsyncDriver
 from typing import Optional
 import json
@@ -34,6 +36,8 @@ class Bot:
         temperature: Optional[float] = None,
         image_generation_endpoint: Optional[str] = None,
         image_generation_backend: Optional[str] = None,
+        image_generation_size: Optional[str] = None,
+        image_format: Optional[str] = None,
         timeout: Optional[float] = 120.0,
     ) -> None:
         if server_url is None:
@@ -54,6 +58,19 @@ class Bot:
                 raise ValueError("scheme must be either http or https")
             self.scheme = scheme
 
+        if image_generation_endpoint and image_generation_backend not in [
+            "openai",
+            "sdwui",
+            "localai",
+            None,
+        ]:
+            logger.error("image_generation_backend must be openai or sdwui or localai")
+            sys.exit(1)
+
+        if image_format not in ["jpeg", "png", None]:
+            logger.error("image_format should be jpeg or png, leave blank for jpeg")
+            sys.exit(1)
+
         # @chatgpt
         if username is None:
             raise ValueError("username must be provided")
@@ -78,6 +95,21 @@ class Bot:
         )
         self.image_generation_endpoint: str = image_generation_endpoint
         self.image_generation_backend: str = image_generation_backend
+
+        if image_format:
+            self.image_format: str = image_format
+        else:
+            self.image_format = "jpeg"
+
+        if image_generation_size is None:
+            self.image_generation_size = "512x512"
+            self.image_generation_width = 512
+            self.image_generation_height = 512
+        else:
+            self.image_generation_size = image_generation_size
+            self.image_generation_width = self.image_generation_size.split("x")[0]
+            self.image_generation_height = self.image_generation_size.split("x")[1]
+
         self.timeout = timeout or 120.0
 
         self.bot_id = None
@@ -251,20 +283,19 @@ class Bot:
                                 "channel_id": channel_id,
                             },
                         )
-                        b64_datas = await imagegen.get_images(
+                        image_path_list = await imagegen.get_images(
                             self.httpx_client,
                             self.image_generation_endpoint,
                             prompt,
                             self.image_generation_backend,
                             timeount=self.timeout,
                             api_key=self.openai_api_key,
+                            output_path=self.base_path / "images",
                             n=1,
-                            size="256x256",
-                        )
-                        image_path_list = await asyncio.to_thread(
-                            imagegen.save_images,
-                            b64_datas,
-                            self.base_path / "images",
+                            size=self.image_generation_size,
+                            width=self.image_generation_width,
+                            height=self.image_generation_height,
+                            image_format=self.image_format,
                         )
                         # send image
                         for image_path in image_path_list:
@@ -274,6 +305,7 @@ class Bot:
                                 image_path,
                                 root_id,
                             )
+                            await aiofiles.os.remove(image_path)
                     except Exception as e:
                         logger.error(e, exc_info=True)
                         raise Exception(e)
@@ -321,8 +353,7 @@ class Bot:
                     "root_id": root_id,
                 }
             )
-            # remove image after posting
-            os.remove(filepath)
+
         except Exception as e:
             logger.error(e, exc_info=True)
             raise Exception(e)
diff --git a/src/imagegen.py b/src/imagegen.py
index 2214eac..8f059d9 100644
--- a/src/imagegen.py
+++ b/src/imagegen.py
@@ -7,9 +7,14 @@ from PIL import Image
 
 
 async def get_images(
-    aclient: httpx.AsyncClient, url: str, prompt: str, backend_type: str, **kwargs
+    aclient: httpx.AsyncClient,
+    url: str,
+    prompt: str,
+    backend_type: str,
+    output_path: str,
+    **kwargs,
 ) -> list[str]:
-    timeout = kwargs.get("timeout", 120.0)
+    timeout = kwargs.get("timeout", 180.0)
     if backend_type == "openai":
         resp = await aclient.post(
             url,
@@ -20,7 +25,7 @@ async def get_images(
             json={
                 "prompt": prompt,
                 "n": kwargs.get("n", 1),
-                "size": kwargs.get("size", "256x256"),
+                "size": kwargs.get("size", "512x512"),
                 "response_format": "b64_json",
             },
             timeout=timeout,
@@ -29,7 +34,7 @@ async def get_images(
             b64_datas = []
             for data in resp.json()["data"]:
                 b64_datas.append(data["b64_json"])
-            return b64_datas
+            return save_images_b64(b64_datas, output_path, **kwargs)
         else:
             raise Exception(
                 f"{resp.status_code} {resp.reason_phrase} {resp.text}",
@@ -45,25 +50,56 @@ async def get_images(
                 "sampler_name": kwargs.get("sampler_name", "Euler a"),
                 "batch_size": kwargs.get("n", 1),
                 "steps": kwargs.get("steps", 20),
-                "width": 256 if "256" in kwargs.get("size") else 512,
-                "height": 256 if "256" in kwargs.get("size") else 512,
+                "width": kwargs.get("width", 512),
+                "height": kwargs.get("height", 512),
             },
             timeout=timeout,
         )
         if resp.status_code == 200:
             b64_datas = resp.json()["images"]
-            return b64_datas
+            return save_images_b64(b64_datas, output_path, **kwargs)
         else:
             raise Exception(
                 f"{resp.status_code} {resp.reason_phrase} {resp.text}",
             )
+    elif backend_type == "localai":
+        resp = await aclient.post(
+            url,
+            headers={
+                "Content-Type": "application/json",
+                "Authorization": f"Bearer {kwargs.get('api_key')}",
+            },
+            json={
+                "prompt": prompt,
+                "size": kwargs.get("size", "512x512"),
+            },
+            timeout=timeout,
+        )
+        if resp.status_code == 200:
+            image_url = resp.json()["data"][0]["url"]
+            return await save_image_url(image_url, aclient, output_path, **kwargs)
 
 
-def save_images(b64_datas: list[str], path: Path, **kwargs) -> list[str]:
-    images = []
+def save_images_b64(b64_datas: list[str], path: Path, **kwargs) -> list[str]:
+    images_path_list = []
     for b64_data in b64_datas:
-        image_path = path / (str(uuid.uuid4()) + ".jpeg")
+        image_path = path / (
+            str(uuid.uuid4()) + "." + kwargs.get("image_format", "jpeg")
+        )
         img = Image.open(io.BytesIO(base64.decodebytes(bytes(b64_data, "utf-8"))))
         img.save(image_path)
-        images.append(image_path)
-    return images
+        images_path_list.append(image_path)
+    return images_path_list
+
+
+async def save_image_url(
+    url: str, aclient: httpx.AsyncClient, path: Path, **kwargs
+) -> list[str]:
+    images_path_list = []
+    r = await aclient.get(url)
+    image_path = path / (str(uuid.uuid4()) + "." + kwargs.get("image_format", "jpeg"))
+    if r.status_code == 200:
+        img = Image.open(io.BytesIO(r.content))
+        img.save(image_path)
+        images_path_list.append(image_path)
+    return images_path_list
diff --git a/src/main.py b/src/main.py
index 0934842..37d3cb8 100644
--- a/src/main.py
+++ b/src/main.py
@@ -39,6 +39,8 @@ async def main():
             temperature=config.get("temperature"),
             image_generation_endpoint=config.get("image_generation_endpoint"),
             image_generation_backend=config.get("image_generation_backend"),
+            image_generation_size=config.get("image_generation_size"),
+            image_format=config.get("image_format"),
             timeout=config.get("timeout"),
         )
 
@@ -48,21 +50,23 @@ async def main():
             email=os.environ.get("EMAIL"),
             password=os.environ.get("PASSWORD"),
             username=os.environ.get("USERNAME"),
-            port=os.environ.get("PORT"),
+            port=int(os.environ.get("PORT", 443)),
             scheme=os.environ.get("SCHEME"),
             openai_api_key=os.environ.get("OPENAI_API_KEY"),
             gpt_api_endpoint=os.environ.get("GPT_API_ENDPOINT"),
             gpt_model=os.environ.get("GPT_MODEL"),
-            max_tokens=os.environ.get("MAX_TOKENS"),
-            top_p=os.environ.get("TOP_P"),
-            presence_penalty=os.environ.get("PRESENCE_PENALTY"),
-            frequency_penalty=os.environ.get("FREQUENCY_PENALTY"),
-            reply_count=os.environ.get("REPLY_COUNT"),
+            max_tokens=int(os.environ.get("MAX_TOKENS", 4000)),
+            top_p=float(os.environ.get("TOP_P", 1.0)),
+            presence_penalty=float(os.environ.get("PRESENCE_PENALTY", 0.0)),
+            frequency_penalty=float(os.environ.get("FREQUENCY_PENALTY", 0.0)),
+            reply_count=int(os.environ.get("REPLY_COUNT", 1)),
             system_prompt=os.environ.get("SYSTEM_PROMPT"),
-            temperature=os.environ.get("TEMPERATURE"),
+            temperature=float(os.environ.get("TEMPERATURE", 0.8)),
             image_generation_endpoint=os.environ.get("IMAGE_GENERATION_ENDPOINT"),
             image_generation_backend=os.environ.get("IMAGE_GENERATION_BACKEND"),
-            timeout=os.environ.get("TIMEOUT"),
+            image_generation_size=os.environ.get("IMAGE_GENERATION_SIZE"),
+            image_format=os.environ.get("IMAGE_FORMAT"),
+            timeout=float(os.environ.get("TIMEOUT", 120.0)),
         )
 
     await mattermost_bot.login()

From 7ddb2434bd600ec57c34f0bdfc49dbd3df171726 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sat, 23 Dec 2023 22:35:06 +0800
Subject: [PATCH 12/15] Fallback to gpt-3.5-turbo when caculate tokens using
 custom model

---
 src/gptbot.py | 8 ++++----
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/src/gptbot.py b/src/gptbot.py
index 454e0a1..31e9c72 100644
--- a/src/gptbot.py
+++ b/src/gptbot.py
@@ -122,13 +122,13 @@ class Chatbot:
         """
         Get token count
         """
+        _engine = self.engine
         if self.engine not in ENGINES:
-            raise NotImplementedError(
-                f"Engine {self.engine} is not supported. Select from {ENGINES}",
-            )
+            # use gpt-3.5-turbo to caculate token
+            _engine = "gpt-3.5-turbo"
         tiktoken.model.MODEL_TO_ENCODING["gpt-4"] = "cl100k_base"
 
-        encoding = tiktoken.encoding_for_model(self.engine)
+        encoding = tiktoken.encoding_for_model(_engine)
 
         num_tokens = 0
         for message in self.conversation[convo_id]:

From dfa0c5b3613de01be3c618d66704d58f18503830 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sat, 23 Dec 2023 22:36:30 +0800
Subject: [PATCH 13/15] Bump pre-commit hook version

---
 .pre-commit-config.yaml | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)

diff --git a/.pre-commit-config.yaml b/.pre-commit-config.yaml
index cc25d41..e88a84f 100644
--- a/.pre-commit-config.yaml
+++ b/.pre-commit-config.yaml
@@ -1,16 +1,16 @@
 repos:
   - repo: https://github.com/pre-commit/pre-commit-hooks
-    rev: v4.4.0
+    rev: v4.5.0
     hooks:
       - id: trailing-whitespace
       - id: end-of-file-fixer
       - id: check-yaml
   - repo: https://github.com/psf/black
-    rev: 23.9.1
+    rev: 23.12.0
     hooks:
       - id: black
   - repo: https://github.com/astral-sh/ruff-pre-commit
-    rev: v0.0.290
+    rev: v0.1.7
     hooks:
       - id: ruff
         args: [--fix, --exit-non-zero-on-fix]

From af5d54f1ce0e94cd518c6ed3ee2cacb091260f35 Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sat, 23 Dec 2023 22:37:36 +0800
Subject: [PATCH 14/15] v1.3.0

---
 CHANGELOG.md             | 3 +++
 full-config.json.example | 4 ++--
 2 files changed, 5 insertions(+), 2 deletions(-)

diff --git a/CHANGELOG.md b/CHANGELOG.md
index b2f9d93..ffb3bb3 100644
--- a/CHANGELOG.md
+++ b/CHANGELOG.md
@@ -1,4 +1,7 @@
 # Changelog
+## v1.3.0
+- Fix localai v2.0+ image generation
+- Support specific output image format(jpeg, png) and size
 
 ## v1.2.0
 - support sending typing state
diff --git a/full-config.json.example b/full-config.json.example
index d964c3a..cdbff48 100644
--- a/full-config.json.example
+++ b/full-config.json.example
@@ -3,7 +3,7 @@
     "email": "bot@hibobmaster.com",
     "username": "@bot",
     "password": "SfBKY%K7*e&a%ZX$3g@Am&jQ",
-    "port": "8065",
+    "port": 8065,
     "scheme": "http",
     "openai_api_key": "xxxxxxxxxxxxxxxxxxxxxxxx",
     "gpt_api_endpoint": "https://api.openai.com/v1/chat/completions",
@@ -18,6 +18,6 @@
     "image_generation_endpoint": "http://localai:8080/v1/images/generations",
     "image_generation_backend": "localai",
     "image_generation_size": "512x512",
-    "image_format": "webp",
+    "image_format": "jpeg",
     "timeout": 120.0
 }

From 4cfa4fcf3ba0ff8b0eb5e00434fda99315f0f90b Mon Sep 17 00:00:00 2001
From: hibobmaster <32976627+hibobmaster@users.noreply.github.com>
Date: Sun, 24 Dec 2023 20:39:50 +0800
Subject: [PATCH 15/15] Update demo pic

---
 README.md | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/README.md b/README.md
index 3b83913..169ceb6 100644
--- a/README.md
+++ b/README.md
@@ -28,7 +28,7 @@ docker compose up -d
 ## Demo
 Remove support for Bing AI, Google Bard due to technical problems.
 ![gpt command](https://imgur.com/vdT83Ln.jpg)
-![image generation](https://i.imgur.com/GHczfkv.jpg)
+![image generation](https://i.imgur.com/DQ3i3wW.jpg)
 
 ## Thanks
 <a href="https://jb.gg/OpenSourceSupport" target="_blank">